code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
from urllib.parse import quote import pytest from datasets.utils.hub import hf_hub_url @pytest.mark.parametrize('repo_id' , ['canonical_dataset_name', 'org-name/dataset-name'] ) @pytest.mark.parametrize('path' , ['filename.csv', 'filename with blanks.csv'] ) @pytest.mark.parametrize('revision' , [None, 'v2'] ) def UpperCamelCase ( snake_case__ : Optional[Any] , snake_case__ : Optional[Any] , snake_case__ : List[Any] ) -> int: UpperCamelCase : Optional[int] = hf_hub_url(repo_id=snake_case__ , path=snake_case__ , revision=snake_case__ ) assert url == F"""https://huggingface.co/datasets/{repo_id}/resolve/{revision or "main"}/{quote(snake_case__ )}"""
40
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class a (_lowerCAmelCase ): """simple docstring""" def __snake_case ( self : str ) -> str: __snake_case : Tuple = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCamelCase , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "neck_hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "num_attention_heads" ) ) class a : """simple docstring""" def __init__( self : Optional[int] , lowerCamelCase : List[str] , lowerCamelCase : Tuple=13 , lowerCamelCase : str=32 , lowerCamelCase : Dict=2 , lowerCamelCase : List[str]=3 , lowerCamelCase : Any=640 , lowerCamelCase : Optional[Any]=4 , lowerCamelCase : Tuple="silu" , lowerCamelCase : int=3 , lowerCamelCase : Dict=32 , lowerCamelCase : str=0.1 , lowerCamelCase : Optional[int]=0.1 , lowerCamelCase : Optional[Any]=0.1 , lowerCamelCase : Dict=0.02 , lowerCamelCase : Union[str, Any]=True , lowerCamelCase : Optional[int]=True , lowerCamelCase : Union[str, Any]=10 , lowerCamelCase : int=None , ) -> str: __snake_case : Optional[Any] = parent __snake_case : Optional[Any] = batch_size __snake_case : Any = image_size __snake_case : List[Any] = patch_size __snake_case : Any = num_channels __snake_case : Union[str, Any] = last_hidden_size __snake_case : Any = num_attention_heads __snake_case : Any = hidden_act __snake_case : Tuple = conv_kernel_size __snake_case : Any = output_stride __snake_case : Any = hidden_dropout_prob __snake_case : List[Any] = attention_probs_dropout_prob __snake_case : Optional[Any] = classifier_dropout_prob __snake_case : Union[str, Any] = use_labels __snake_case : Optional[int] = is_training __snake_case : Dict = num_labels __snake_case : Any = initializer_range __snake_case : Optional[int] = scope def __snake_case ( self : str ) -> Union[str, Any]: __snake_case : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : List[Any] = None __snake_case : Optional[int] = None if self.use_labels: __snake_case : Optional[int] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : Any = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __snake_case : Optional[Any] = self.get_config() return config, pixel_values, labels, pixel_labels def __snake_case ( self : Any ) -> Union[str, Any]: return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __snake_case ( self : Any , lowerCamelCase : Tuple , lowerCamelCase : int , lowerCamelCase : Tuple , lowerCamelCase : Optional[int] ) -> Dict: __snake_case : List[Any] = MobileViTModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[str] = model(lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[Any] , lowerCamelCase : List[str] , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] , lowerCamelCase : Tuple ) -> List[str]: __snake_case : str = self.num_labels __snake_case : List[Any] = MobileViTForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : Dict , lowerCamelCase : Any , lowerCamelCase : Any , lowerCamelCase : Optional[Any] , lowerCamelCase : Dict ) -> Dict: __snake_case : Union[str, Any] = self.num_labels __snake_case : Optional[int] = MobileViTForSemanticSegmentation(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Tuple = model(lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[int] ) -> List[Any]: __snake_case : Optional[Any] = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case , __snake_case : Union[str, Any] = config_and_inputs __snake_case : Dict = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) __UpperCAmelCase : Optional[Any] = ( { "feature-extraction": MobileViTModel, "image-classification": MobileViTForImageClassification, "image-segmentation": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) __UpperCAmelCase : List[str] = False __UpperCAmelCase : int = False __UpperCAmelCase : Optional[int] = False __UpperCAmelCase : Optional[int] = False def __snake_case ( self : Optional[int] ) -> Dict: __snake_case : Tuple = MobileViTModelTester(self ) __snake_case : Any = MobileViTConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase ) def __snake_case ( self : Optional[int] ) -> Dict: self.config_tester.run_common_tests() @unittest.skip(reason="MobileViT does not use inputs_embeds" ) def __snake_case ( self : Dict ) -> Any: pass @unittest.skip(reason="MobileViT does not support input and output embeddings" ) def __snake_case ( self : Dict ) -> List[Any]: pass @unittest.skip(reason="MobileViT does not output attentions" ) def __snake_case ( self : int ) -> Dict: pass def __snake_case ( self : int ) -> Union[str, Any]: __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Optional[Any] = model_class(lowerCamelCase ) __snake_case : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : int = [*signature.parameters.keys()] __snake_case : List[str] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __snake_case ( self : int ) -> Tuple: pass def __snake_case ( self : Any ) -> Tuple: __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def __snake_case ( self : Any ) -> str: def check_hidden_states_output(lowerCamelCase : Union[str, Any] , lowerCamelCase : Dict , lowerCamelCase : Any ): __snake_case : int = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : int = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : Union[str, Any] = outputs.hidden_states __snake_case : int = 5 self.assertEqual(len(lowerCamelCase ) , lowerCamelCase ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. __snake_case : List[Any] = 2 for i in range(len(lowerCamelCase ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) __snake_case , __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : str = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : List[Any] = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Any ) -> Any: __snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) def __snake_case ( self : List[str] ) -> List[str]: __snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowerCamelCase ) @slow def __snake_case ( self : List[str] ) -> Any: for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : List[str] = MobileViTModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def lowerCAmelCase_ ( ): __snake_case : Optional[int] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a (unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self : str ) -> Dict: return MobileViTImageProcessor.from_pretrained("apple/mobilevit-xx-small" ) if is_vision_available() else None @slow def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : Tuple = MobileViTForImageClassification.from_pretrained("apple/mobilevit-xx-small" ).to(lowerCamelCase ) __snake_case : Optional[Any] = self.default_image_processor __snake_case : Union[str, Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Dict = model(**lowerCamelCase ) # verify the logits __snake_case : Union[str, Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([-1.93_64, -1.23_27, -0.46_53] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : str ) -> Optional[int]: __snake_case : Optional[int] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : str = model.to(lowerCamelCase ) __snake_case : int = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Optional[int] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : List[str] = model(**lowerCamelCase ) __snake_case : Union[str, Any] = outputs.logits # verify the logits __snake_case : Tuple = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , lowerCamelCase ) __snake_case : Union[str, Any] = torch.tensor( [ [[6.97_13, 6.97_86, 7.24_22], [7.28_93, 7.28_25, 7.44_46], [7.65_80, 7.87_97, 7.94_20]], [[-10.68_69, -10.32_50, -10.34_71], [-10.42_28, -9.98_68, -9.71_32], [-11.04_05, -11.02_21, -10.73_18]], [[-3.30_89, -2.85_39, -2.67_40], [-3.27_06, -2.56_21, -2.51_08], [-3.25_34, -2.66_15, -2.66_51]], ] , device=lowerCamelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : Union[str, Any] ) -> Optional[int]: __snake_case : Optional[Any] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Tuple = model.to(lowerCamelCase ) __snake_case : Dict = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : List[Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Any = model(**lowerCamelCase ) __snake_case : Dict = outputs.logits.detach().cpu() __snake_case : Any = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase , target_sizes=[(50, 60)] ) __snake_case : int = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , lowerCamelCase ) __snake_case : List[str] = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase ) __snake_case : Optional[int] = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , lowerCamelCase )
81
0
'''simple docstring''' from collections.abc import Generator def _A ( ): """simple docstring""" __lowercase , __lowercase = 0, 1 while True: __lowercase , __lowercase = b, a + b yield b def _A ( A__ = 1000 ): """simple docstring""" __lowercase = 1 __lowercase = fibonacci_generator() while len(str(next(A__ ) ) ) < n: answer += 1 return answer + 1 if __name__ == "__main__": print(solution(int(str(input()).strip())))
41
import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor _snake_case : int = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[int] , *lowerCamelCase : List[Any] , **lowerCamelCase : int ) -> None: warnings.warn( "The class ImageGPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use ImageGPTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
'''simple docstring''' import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartTokenizer, MBartTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin A_ = get_tests_dir("fixtures/test_sentencepiece.model") if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right A_ = 250_004 A_ = 250_020 @require_sentencepiece @require_tokenizers class UpperCAmelCase ( UpperCAmelCase__ , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE_ = MBartTokenizer SCREAMING_SNAKE_CASE_ = MBartTokenizerFast SCREAMING_SNAKE_CASE_ = True SCREAMING_SNAKE_CASE_ = True def UpperCamelCase( self ) -> List[Any]: '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing lowerCamelCase_ = MBartTokenizer(SCREAMING_SNAKE_CASE_ , keep_accents=SCREAMING_SNAKE_CASE_ ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase( self ) -> List[Any]: '''simple docstring''' lowerCamelCase_ = MBartTokenizer(SCREAMING_SNAKE_CASE_ , keep_accents=SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = tokenizer.tokenize('This is a test' ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(SCREAMING_SNAKE_CASE_ ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) lowerCamelCase_ = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( SCREAMING_SNAKE_CASE_ , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) lowerCamelCase_ = tokenizer.convert_tokens_to_ids(SCREAMING_SNAKE_CASE_ ) self.assertListEqual( SCREAMING_SNAKE_CASE_ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) lowerCamelCase_ = tokenizer.convert_ids_to_tokens(SCREAMING_SNAKE_CASE_ ) self.assertListEqual( SCREAMING_SNAKE_CASE_ , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) def UpperCamelCase( self ) -> int: '''simple docstring''' if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return lowerCamelCase_ = (self.rust_tokenizer_class, 'hf-internal-testing/tiny-random-mbart', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowerCamelCase_ = self.rust_tokenizer_class.from_pretrained(SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = self.tokenizer_class.from_pretrained(SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = tempfile.mkdtemp() lowerCamelCase_ = tokenizer_r.save_pretrained(SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = tokenizer_p.save_pretrained(SCREAMING_SNAKE_CASE_ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('tokenizer.json' in f for f in tokenizer_r_files ) ) lowerCamelCase_ = tuple(f for f in tokenizer_r_files if 'tokenizer.json' not in f ) self.assertSequenceEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Checks everything loads correctly in the same way lowerCamelCase_ = tokenizer_r.from_pretrained(SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = tokenizer_p.from_pretrained(SCREAMING_SNAKE_CASE_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(SCREAMING_SNAKE_CASE_ ) # Save tokenizer rust, legacy_format=True lowerCamelCase_ = tempfile.mkdtemp() lowerCamelCase_ = tokenizer_r.save_pretrained(SCREAMING_SNAKE_CASE_ , legacy_format=SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = tokenizer_p.save_pretrained(SCREAMING_SNAKE_CASE_ ) # Checks it save with the same files self.assertSequenceEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Checks everything loads correctly in the same way lowerCamelCase_ = tokenizer_r.from_pretrained(SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = tokenizer_p.from_pretrained(SCREAMING_SNAKE_CASE_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) shutil.rmtree(SCREAMING_SNAKE_CASE_ ) # Save tokenizer rust, legacy_format=False lowerCamelCase_ = tempfile.mkdtemp() lowerCamelCase_ = tokenizer_r.save_pretrained(SCREAMING_SNAKE_CASE_ , legacy_format=SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = tokenizer_p.save_pretrained(SCREAMING_SNAKE_CASE_ ) # Checks it saved the tokenizer.json file self.assertTrue(any('tokenizer.json' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way lowerCamelCase_ = tokenizer_r.from_pretrained(SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = tokenizer_p.from_pretrained(SCREAMING_SNAKE_CASE_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) shutil.rmtree(SCREAMING_SNAKE_CASE_ ) @require_torch @require_sentencepiece @require_tokenizers class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE_ = 'facebook/mbart-large-en-ro' SCREAMING_SNAKE_CASE_ = [ ' UN Chief Says There Is No Military Solution in Syria', ' Secretary-General Ban Ki-moon says his response to Russia\'s stepped up military support for Syria is that "there is no military solution" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.', ] SCREAMING_SNAKE_CASE_ = [ 'Şeful ONU declară că nu există o soluţie militară în Siria', 'Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei' ' pentru Siria este că "nu există o soluţie militară" la conflictul de aproape cinci ani şi că noi arme nu vor' ' face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.', ] SCREAMING_SNAKE_CASE_ = [82_74, 12_78_73, 2_59_16, 7, 86_22, 20_71, 4_38, 6_74_85, 53, 18_78_95, 23, 5_17_12, 2, EN_CODE] @classmethod def UpperCamelCase( cls ) -> Optional[Any]: '''simple docstring''' lowerCamelCase_ = MBartTokenizer.from_pretrained( cls.checkpoint_name , src_lang='en_XX' , tgt_lang='ro_RO' ) lowerCamelCase_ = 1 return cls def UpperCamelCase( self ) -> List[Any]: '''simple docstring''' self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['ar_AR'] , 250001 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['en_EN'] , 250004 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['ro_RO'] , 250020 ) def UpperCamelCase( self ) -> Dict: '''simple docstring''' lowerCamelCase_ = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , SCREAMING_SNAKE_CASE_ ) def UpperCamelCase( self ) -> Optional[int]: '''simple docstring''' self.assertIn(SCREAMING_SNAKE_CASE_ , self.tokenizer.all_special_ids ) lowerCamelCase_ = [RO_CODE, 884, 9019, 96, 9, 916, 86792, 36, 18743, 15596, 5, 2] lowerCamelCase_ = self.tokenizer.decode(SCREAMING_SNAKE_CASE_ , skip_special_tokens=SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) self.assertNotIn(self.tokenizer.eos_token , SCREAMING_SNAKE_CASE_ ) def UpperCamelCase( self ) -> List[Any]: '''simple docstring''' lowerCamelCase_ = ['this is gunna be a long sentence ' * 20] assert isinstance(src_text[0] , SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = 10 lowerCamelCase_ = self.tokenizer(SCREAMING_SNAKE_CASE_ , max_length=SCREAMING_SNAKE_CASE_ , truncation=SCREAMING_SNAKE_CASE_ ).input_ids[0] self.assertEqual(ids[-2] , 2 ) self.assertEqual(ids[-1] , SCREAMING_SNAKE_CASE_ ) self.assertEqual(len(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ ) def UpperCamelCase( self ) -> Tuple: '''simple docstring''' self.assertListEqual(self.tokenizer.convert_tokens_to_ids(['<mask>', 'ar_AR'] ) , [250026, 250001] ) def UpperCamelCase( self ) -> List[str]: '''simple docstring''' lowerCamelCase_ = tempfile.mkdtemp() lowerCamelCase_ = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = MBartTokenizer.from_pretrained(SCREAMING_SNAKE_CASE_ ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , SCREAMING_SNAKE_CASE_ ) @require_torch def UpperCamelCase( self ) -> Dict: '''simple docstring''' lowerCamelCase_ = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=SCREAMING_SNAKE_CASE_ , return_tensors='pt' ) lowerCamelCase_ = shift_tokens_right(batch['labels'] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][-2:].tolist() == [2, EN_CODE] assert batch.decoder_input_ids[1][0].tolist() == RO_CODE assert batch.decoder_input_ids[1][-1] == 2 assert batch.labels[1][-2:].tolist() == [2, RO_CODE] @require_torch def UpperCamelCase( self ) -> Dict: '''simple docstring''' lowerCamelCase_ = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=SCREAMING_SNAKE_CASE_ , truncation=SCREAMING_SNAKE_CASE_ , max_length=len(self.expected_src_tokens ) , return_tensors='pt' , ) lowerCamelCase_ = shift_tokens_right(batch['labels'] , self.tokenizer.pad_token_id ) self.assertIsInstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) self.assertEqual((2, 14) , batch.input_ids.shape ) self.assertEqual((2, 14) , batch.attention_mask.shape ) lowerCamelCase_ = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , SCREAMING_SNAKE_CASE_ ) self.assertEqual(2 , batch.decoder_input_ids[0, -1] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, EN_CODE] ) def UpperCamelCase( self ) -> Optional[Any]: '''simple docstring''' lowerCamelCase_ = self.tokenizer(self.src_text , padding=SCREAMING_SNAKE_CASE_ , truncation=SCREAMING_SNAKE_CASE_ , max_length=3 , return_tensors='pt' ) lowerCamelCase_ = self.tokenizer( text_target=self.tgt_text , padding=SCREAMING_SNAKE_CASE_ , truncation=SCREAMING_SNAKE_CASE_ , max_length=10 , return_tensors='pt' ) lowerCamelCase_ = targets['input_ids'] lowerCamelCase_ = shift_tokens_right(SCREAMING_SNAKE_CASE_ , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def UpperCamelCase( self ) -> Any: '''simple docstring''' lowerCamelCase_ = self.tokenizer._build_translation_inputs( 'A test' , return_tensors='pt' , src_lang='en_XX' , tgt_lang='ar_AR' ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE_ ) , { # A, test, EOS, en_XX 'input_ids': [[62, 3034, 2, 250004]], 'attention_mask': [[1, 1, 1, 1]], # ar_AR 'forced_bos_token_id': 250001, } , )
42
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _snake_case : Optional[Any] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[str] = ["pixel_values"] def __init__( self : List[Any] , lowerCamelCase : bool = True , lowerCamelCase : Union[int, float] = 1 / 255 , lowerCamelCase : bool = True , lowerCamelCase : int = 8 , **lowerCamelCase : Tuple , ) -> None: super().__init__(**lowerCamelCase ) __snake_case : Dict = do_rescale __snake_case : Dict = rescale_factor __snake_case : Optional[Any] = do_pad __snake_case : Tuple = pad_size def __snake_case ( self : Dict , lowerCamelCase : np.ndarray , lowerCamelCase : float , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase : Optional[int] ) -> np.ndarray: return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def __snake_case ( self : Optional[Any] , lowerCamelCase : np.ndarray , lowerCamelCase : int , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None ) -> Tuple: __snake_case , __snake_case : List[str] = get_image_size(lowerCamelCase ) __snake_case : Optional[Any] = (old_height // size + 1) * size - old_height __snake_case : List[Any] = (old_width // size + 1) * size - old_width return pad(lowerCamelCase , ((0, pad_height), (0, pad_width)) , mode="symmetric" , data_format=lowerCamelCase ) def __snake_case ( self : Tuple , lowerCamelCase : ImageInput , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[float] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[int] = None , lowerCamelCase : Optional[Union[str, TensorType]] = None , lowerCamelCase : Union[str, ChannelDimension] = ChannelDimension.FIRST , **lowerCamelCase : Union[str, Any] , ) -> List[str]: __snake_case : int = do_rescale if do_rescale is not None else self.do_rescale __snake_case : Dict = rescale_factor if rescale_factor is not None else self.rescale_factor __snake_case : str = do_pad if do_pad is not None else self.do_pad __snake_case : Any = pad_size if pad_size is not None else self.pad_size __snake_case : int = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) # All transformations expect numpy arrays. __snake_case : str = [to_numpy_array(lowerCamelCase ) for image in images] if do_rescale: __snake_case : Optional[int] = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_pad: __snake_case : Optional[Any] = [self.pad(lowerCamelCase , size=lowerCamelCase ) for image in images] __snake_case : int = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __snake_case : Union[str, Any] = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
81
0
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto.configuration_auto import CONFIG_MAPPING lowerCAmelCase = logging.get_logger(__name__) class _a ( UpperCamelCase__ ): _lowercase : Dict = '''upernet''' def __init__( self: Optional[Any] , UpperCamelCase_: Dict=None , UpperCamelCase_: List[str]=512 , UpperCamelCase_: Dict=0.02 , UpperCamelCase_: str=[1, 2, 3, 6] , UpperCamelCase_: Union[str, Any]=True , UpperCamelCase_: Optional[Any]=0.4 , UpperCamelCase_: Optional[Any]=384 , UpperCamelCase_: List[Any]=256 , UpperCamelCase_: Any=1 , UpperCamelCase_: str=False , UpperCamelCase_: Optional[int]=255 , **UpperCamelCase_: Any , ) -> Optional[Any]: """simple docstring""" super().__init__(**UpperCamelCase_ ) if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''' ) lowercase__ = CONFIG_MAPPING['''resnet'''](out_features=['''stage1''', '''stage2''', '''stage3''', '''stage4'''] ) elif isinstance(UpperCamelCase_ , UpperCamelCase_ ): lowercase__ = backbone_config.get('''model_type''' ) lowercase__ = CONFIG_MAPPING[backbone_model_type] lowercase__ = config_class.from_dict(UpperCamelCase_ ) lowercase__ = backbone_config lowercase__ = hidden_size lowercase__ = initializer_range lowercase__ = pool_scales lowercase__ = use_auxiliary_head lowercase__ = auxiliary_loss_weight lowercase__ = auxiliary_in_channels lowercase__ = auxiliary_channels lowercase__ = auxiliary_num_convs lowercase__ = auxiliary_concat_input lowercase__ = loss_ignore_index def lowerCamelCase_ ( self: List[str] ) -> Optional[int]: """simple docstring""" lowercase__ = copy.deepcopy(self.__dict__ ) lowercase__ = self.backbone_config.to_dict() lowercase__ = self.__class__.model_type return output
43
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case : Any = { "configuration_informer": [ "INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "InformerConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "InformerForPrediction", "InformerModel", "InformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys _snake_case : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
0
'''simple docstring''' def A_ ( _lowerCAmelCase : List[Any] , _lowerCAmelCase : Any ): """simple docstring""" return (pointa[0] - pointa[0]) ** 2 + (pointa[1] - pointa[1]) ** 2 def A_ ( _lowerCAmelCase : Any , _lowerCAmelCase : str=0 ): """simple docstring""" return sorted(_lowerCAmelCase , key=lambda _lowerCAmelCase : x[column] ) def A_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Tuple=float("inf" ) ): """simple docstring""" for i in range(points_counts - 1 ): for j in range(i + 1 , _lowerCAmelCase ): _lowerCamelCase : Optional[Any] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: _lowerCamelCase : Optional[int] = current_dis return min_dis def A_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Any , _lowerCAmelCase : str=float("inf" ) ): """simple docstring""" for i in range(min(6 , points_counts - 1 ) , _lowerCAmelCase ): for j in range(max(0 , i - 6 ) , _lowerCAmelCase ): _lowerCamelCase : Any = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: _lowerCamelCase : Optional[int] = current_dis return min_dis def A_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : str ): """simple docstring""" if points_counts <= 3: return dis_between_closest_pair(_lowerCAmelCase , _lowerCAmelCase ) # recursion _lowerCamelCase : int = points_counts // 2 _lowerCamelCase : Dict = closest_pair_of_points_sqr( _lowerCAmelCase , points_sorted_on_y[:mid] , _lowerCAmelCase ) _lowerCamelCase : Optional[Any] = closest_pair_of_points_sqr( _lowerCAmelCase , points_sorted_on_y[mid:] , points_counts - mid ) _lowerCamelCase : Tuple = min(_lowerCAmelCase , _lowerCAmelCase ) _lowerCamelCase : Optional[Any] = [] for point in points_sorted_on_x: if abs(point[0] - points_sorted_on_x[mid][0] ) < closest_pair_dis: cross_strip.append(_lowerCAmelCase ) _lowerCamelCase : Union[str, Any] = dis_between_closest_in_strip( _lowerCAmelCase , len(_lowerCAmelCase ) , _lowerCAmelCase ) return min(_lowerCAmelCase , _lowerCAmelCase ) def A_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] ): """simple docstring""" _lowerCamelCase : Tuple = column_based_sort(_lowerCAmelCase , column=0 ) _lowerCamelCase : int = column_based_sort(_lowerCAmelCase , column=1 ) return ( closest_pair_of_points_sqr( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) ) ** 0.5 if __name__ == "__main__": UpperCAmelCase_ : Tuple = [(2, 3), (12, 30), (40, 50), (5, 1), (12, 10), (3, 4)] print('Distance:', closest_pair_of_points(points, len(points)))
44
from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging _snake_case : Any = logging.get_logger(__name__) # pylint: disable=invalid-name class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[Any] , lowerCamelCase : CLIPSegForImageSegmentation , lowerCamelCase : CLIPSegProcessor , lowerCamelCase : AutoencoderKL , lowerCamelCase : CLIPTextModel , lowerCamelCase : CLIPTokenizer , lowerCamelCase : UNetaDConditionModel , lowerCamelCase : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , lowerCamelCase : StableDiffusionSafetyChecker , lowerCamelCase : CLIPImageProcessor , ) -> Tuple: super().__init__() if hasattr(scheduler.config , "steps_offset" ) and scheduler.config.steps_offset != 1: __snake_case : Tuple = ( F'The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`' F' should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure ' "to update the config accordingly as leaving `steps_offset` might led to incorrect results" " in future versions. If you have downloaded this checkpoint from the Hugging Face Hub," " it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`" " file" ) deprecate("steps_offset!=1" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : Any = dict(scheduler.config ) __snake_case : List[Any] = 1 __snake_case : Tuple = FrozenDict(lowerCamelCase ) if hasattr(scheduler.config , "skip_prk_steps" ) and scheduler.config.skip_prk_steps is False: __snake_case : List[str] = ( F'The configuration file of this scheduler: {scheduler} has not set the configuration' " `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make" " sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to" " incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face" " Hub, it would be very nice if you could open a Pull request for the" " `scheduler/scheduler_config.json` file" ) deprecate("skip_prk_steps not set" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : List[str] = dict(scheduler.config ) __snake_case : List[str] = True __snake_case : Any = FrozenDict(lowerCamelCase ) if safety_checker is None: logger.warning( F'You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure' " that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered" " results in services or applications open to the public. Both the diffusers team and Hugging Face" " strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling" " it only for use-cases that involve analyzing network behavior or auditing its results. For more" " information, please have a look at https://github.com/huggingface/diffusers/pull/254 ." ) self.register_modules( segmentation_model=lowerCamelCase , segmentation_processor=lowerCamelCase , vae=lowerCamelCase , text_encoder=lowerCamelCase , tokenizer=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase , safety_checker=lowerCamelCase , feature_extractor=lowerCamelCase , ) def __snake_case ( self : Dict , lowerCamelCase : Optional[Union[str, int]] = "auto" ) -> Tuple: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __snake_case : Any = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowerCamelCase ) def __snake_case ( self : List[Any] ) -> Any: self.enable_attention_slicing(lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> str: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) __snake_case : Optional[int] = torch.device("cuda" ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(lowerCamelCase , lowerCamelCase ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __snake_case ( self : int ) -> Any: if self.device != torch.device("meta" ) or not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowerCamelCase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self : List[Any] , lowerCamelCase : Union[str, List[str]] , lowerCamelCase : Union[torch.FloatTensor, PIL.Image.Image] , lowerCamelCase : str , lowerCamelCase : int = 512 , lowerCamelCase : int = 512 , lowerCamelCase : int = 50 , lowerCamelCase : float = 7.5 , lowerCamelCase : Optional[Union[str, List[str]]] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : float = 0.0 , lowerCamelCase : Optional[torch.Generator] = None , lowerCamelCase : Optional[torch.FloatTensor] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , lowerCamelCase : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , lowerCamelCase : int = 1 , **lowerCamelCase : Dict , ) -> List[str]: __snake_case : Tuple = self.segmentation_processor( text=[text] , images=[image] , padding="max_length" , return_tensors="pt" ).to(self.device ) __snake_case : str = self.segmentation_model(**lowerCamelCase ) __snake_case : Dict = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() __snake_case : List[Any] = self.numpy_to_pil(lowerCamelCase )[0].resize(image.size ) # Run inpainting pipeline with the generated mask __snake_case : Tuple = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=lowerCamelCase , image=lowerCamelCase , mask_image=lowerCamelCase , height=lowerCamelCase , width=lowerCamelCase , num_inference_steps=lowerCamelCase , guidance_scale=lowerCamelCase , negative_prompt=lowerCamelCase , num_images_per_prompt=lowerCamelCase , eta=lowerCamelCase , generator=lowerCamelCase , latents=lowerCamelCase , output_type=lowerCamelCase , return_dict=lowerCamelCase , callback=lowerCamelCase , callback_steps=lowerCamelCase , )
81
0
UpperCamelCase = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/" def A ( lowercase__ : bytes ) -> bytes: # Make sure the supplied data is a bytes-like object if not isinstance(lowercase__ , lowercase__ ): UpperCamelCase__ :Dict = f"""a bytes-like object is required, not '{data.__class__.__name__}'""" raise TypeError(lowercase__ ) UpperCamelCase__ :Any = """""".join(bin(lowercase__ )[2:].zfill(8 ) for byte in data ) UpperCamelCase__ :Optional[Any] = len(lowercase__ ) % 6 != 0 if padding_needed: # The padding that will be added later UpperCamelCase__ :int = b"""=""" * ((6 - len(lowercase__ ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(lowercase__ ) % 6) else: UpperCamelCase__ :List[Any] = b"""""" # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(lowercase__ ) , 6 ) ).encode() + padding ) def A ( lowercase__ : str ) -> bytes: # Make sure encoded_data is either a string or a bytes-like object if not isinstance(lowercase__ , lowercase__ ) and not isinstance(lowercase__ , lowercase__ ): UpperCamelCase__ :Dict = ( """argument should be a bytes-like object or ASCII string, """ f"""not '{encoded_data.__class__.__name__}'""" ) raise TypeError(lowercase__ ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(lowercase__ , lowercase__ ): try: UpperCamelCase__ :List[str] = encoded_data.decode("""utf-8""" ) except UnicodeDecodeError: raise ValueError("""base64 encoded data should only contain ASCII characters""" ) UpperCamelCase__ :int = encoded_data.count("""=""" ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(lowercase__ ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one UpperCamelCase__ :int = encoded_data[:-padding] UpperCamelCase__ :Optional[int] = """""".join( bin(B64_CHARSET.index(lowercase__ ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: UpperCamelCase__ :List[str] = """""".join( bin(B64_CHARSET.index(lowercase__ ) )[2:].zfill(6 ) for char in encoded_data ) UpperCamelCase__ :Optional[int] = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(lowercase__ ) , 8 ) ] return bytes(lowercase__ ) if __name__ == "__main__": import doctest doctest.testmod()
45
class a : """simple docstring""" def __init__( self : Tuple , lowerCamelCase : list ) -> None: __snake_case : str = set_counts __snake_case : Union[str, Any] = max(lowerCamelCase ) __snake_case : List[Any] = len(lowerCamelCase ) __snake_case : Tuple = [1] * num_sets __snake_case : Dict = list(range(lowerCamelCase ) ) def __snake_case ( self : str , lowerCamelCase : int , lowerCamelCase : int ) -> bool: __snake_case : List[Any] = self.get_parent(lowerCamelCase ) __snake_case : Tuple = self.get_parent(lowerCamelCase ) if src_parent == dst_parent: return False if self.ranks[dst_parent] >= self.ranks[src_parent]: self.set_counts[dst_parent] += self.set_counts[src_parent] __snake_case : List[str] = 0 __snake_case : List[Any] = dst_parent if self.ranks[dst_parent] == self.ranks[src_parent]: self.ranks[dst_parent] += 1 __snake_case : Dict = self.set_counts[dst_parent] else: self.set_counts[src_parent] += self.set_counts[dst_parent] __snake_case : Union[str, Any] = 0 __snake_case : Optional[int] = src_parent __snake_case : Tuple = self.set_counts[src_parent] __snake_case : str = max(self.max_set , lowerCamelCase ) return True def __snake_case ( self : int , lowerCamelCase : int ) -> int: if self.parents[disj_set] == disj_set: return disj_set __snake_case : Optional[int] = self.get_parent(self.parents[disj_set] ) return self.parents[disj_set]
81
0
"""simple docstring""" import unittest from transformers import AutoConfig, AutoTokenizer, BertConfig, TensorType, is_flax_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, require_flax, slow if is_flax_available(): import jax from transformers.models.auto.modeling_flax_auto import FlaxAutoModel from transformers.models.bert.modeling_flax_bert import FlaxBertModel from transformers.models.roberta.modeling_flax_roberta import FlaxRobertaModel @require_flax class A_ ( unittest.TestCase ): @slow def _lowercase ( self: List[str] ): '''simple docstring''' for model_name in ["bert-base-cased", "bert-large-uncased"]: with self.subTest(__lowerCAmelCase ): _lowerCamelCase : List[Any] = AutoConfig.from_pretrained(__lowerCAmelCase ) self.assertIsNotNone(__lowerCAmelCase ) self.assertIsInstance(__lowerCAmelCase ,__lowerCAmelCase ) _lowerCamelCase : Optional[int] = FlaxAutoModel.from_pretrained(__lowerCAmelCase ) self.assertIsNotNone(__lowerCAmelCase ) self.assertIsInstance(__lowerCAmelCase ,__lowerCAmelCase ) @slow def _lowercase ( self: int ): '''simple docstring''' for model_name in ["roberta-base", "roberta-large"]: with self.subTest(__lowerCAmelCase ): _lowerCamelCase : int = AutoConfig.from_pretrained(__lowerCAmelCase ) self.assertIsNotNone(__lowerCAmelCase ) self.assertIsInstance(__lowerCAmelCase ,__lowerCAmelCase ) _lowerCamelCase : Union[str, Any] = FlaxAutoModel.from_pretrained(__lowerCAmelCase ) self.assertIsNotNone(__lowerCAmelCase ) self.assertIsInstance(__lowerCAmelCase ,__lowerCAmelCase ) @slow def _lowercase ( self: int ): '''simple docstring''' for model_name in ["bert-base-cased", "bert-large-uncased"]: _lowerCamelCase : List[str] = AutoTokenizer.from_pretrained(__lowerCAmelCase ) _lowerCamelCase : str = FlaxBertModel.from_pretrained(__lowerCAmelCase ) _lowerCamelCase : Union[str, Any] = tokenizer("Do you support jax jitted function?" ,return_tensors=TensorType.JAX ) @jax.jit def eval(**__lowerCAmelCase: Union[str, Any] ): return model(**__lowerCAmelCase ) eval(**__lowerCAmelCase ).block_until_ready() @slow def _lowercase ( self: Union[str, Any] ): '''simple docstring''' for model_name in ["roberta-base", "roberta-large"]: _lowerCamelCase : List[Any] = AutoTokenizer.from_pretrained(__lowerCAmelCase ) _lowerCamelCase : List[Any] = FlaxRobertaModel.from_pretrained(__lowerCAmelCase ) _lowerCamelCase : Any = tokenizer("Do you support jax jitted function?" ,return_tensors=TensorType.JAX ) @jax.jit def eval(**__lowerCAmelCase: int ): return model(**__lowerCAmelCase ) eval(**__lowerCAmelCase ).block_until_ready() def _lowercase ( self: Any ): '''simple docstring''' with self.assertRaisesRegex( __lowerCAmelCase ,"bert-base is not a local folder and is not a valid model identifier" ): _lowerCamelCase : Union[str, Any] = FlaxAutoModel.from_pretrained("bert-base" ) def _lowercase ( self: int ): '''simple docstring''' with self.assertRaisesRegex( __lowerCAmelCase ,r"aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)" ): _lowerCamelCase : Union[str, Any] = FlaxAutoModel.from_pretrained(__lowerCAmelCase ,revision="aaaaaa" ) def _lowercase ( self: int ): '''simple docstring''' with self.assertRaisesRegex( __lowerCAmelCase ,"hf-internal-testing/config-no-model does not appear to have a file named flax_model.msgpack" ,): _lowerCamelCase : Union[str, Any] = FlaxAutoModel.from_pretrained("hf-internal-testing/config-no-model" ) def _lowercase ( self: Optional[int] ): '''simple docstring''' with self.assertRaisesRegex(__lowerCAmelCase ,"Use `from_pt=True` to load this model" ): _lowerCamelCase : Any = FlaxAutoModel.from_pretrained("hf-internal-testing/tiny-bert-pt-only" )
46
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class a (unittest.TestCase ): """simple docstring""" def __init__( self : int , lowerCamelCase : Any , lowerCamelCase : int=7 , lowerCamelCase : str=3 , lowerCamelCase : Optional[int]=30 , lowerCamelCase : Dict=400 , lowerCamelCase : str=True , lowerCamelCase : str=None , lowerCamelCase : Any=True , lowerCamelCase : Union[str, Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[str]=True , lowerCamelCase : Optional[int]=1 / 255 , lowerCamelCase : Any=True , ) -> str: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __snake_case : Optional[Any] = size if size is not None else {"shortest_edge": 18, "longest_edge": 1333} __snake_case : Optional[Any] = parent __snake_case : List[Any] = batch_size __snake_case : Optional[int] = num_channels __snake_case : str = min_resolution __snake_case : int = max_resolution __snake_case : int = do_resize __snake_case : Tuple = size __snake_case : Any = do_normalize __snake_case : int = image_mean __snake_case : Tuple = image_std __snake_case : Dict = do_rescale __snake_case : Optional[Any] = rescale_factor __snake_case : str = do_pad def __snake_case ( self : Any ) -> int: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def __snake_case ( self : Dict , lowerCamelCase : List[Any] , lowerCamelCase : Optional[Any]=False ) -> List[str]: if not batched: __snake_case : Dict = image_inputs[0] if isinstance(lowerCamelCase , Image.Image ): __snake_case , __snake_case : Dict = image.size else: __snake_case , __snake_case : List[str] = image.shape[1], image.shape[2] if w < h: __snake_case : Optional[int] = int(self.size["shortest_edge"] * h / w ) __snake_case : int = self.size["shortest_edge"] elif w > h: __snake_case : List[str] = self.size["shortest_edge"] __snake_case : Optional[Any] = int(self.size["shortest_edge"] * w / h ) else: __snake_case : List[Any] = self.size["shortest_edge"] __snake_case : Any = self.size["shortest_edge"] else: __snake_case : int = [] for image in image_inputs: __snake_case , __snake_case : List[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[0] )[0] __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class a (_lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ConditionalDetrImageProcessor if is_vision_available() else None def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : Optional[Any] = ConditionalDetrImageProcessingTester(self ) @property def __snake_case ( self : Any ) -> str: return self.image_processor_tester.prepare_image_processor_dict() def __snake_case ( self : Optional[Any] ) -> Optional[int]: __snake_case : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase , "image_mean" ) ) self.assertTrue(hasattr(lowerCamelCase , "image_std" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_normalize" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase , "size" ) ) def __snake_case ( self : Any ) -> Dict: __snake_case : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 1333} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) __snake_case : str = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=lowerCamelCase ) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> Dict: pass def __snake_case ( self : Tuple ) -> str: # Initialize image_processing __snake_case : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , Image.Image ) # Test not batched input __snake_case : int = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) __snake_case : Dict = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> str: # Initialize image_processing __snake_case : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __snake_case : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , numpify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , np.ndarray ) # Test not batched input __snake_case : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : List[Any] = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> List[str]: # Initialize image_processing __snake_case : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , torchify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , torch.Tensor ) # Test not batched input __snake_case : Optional[Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : List[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : int = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def __snake_case ( self : Any ) -> Optional[int]: # prepare image and target __snake_case : List[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : List[Any] = {"image_id": 39769, "annotations": target} # encode them __snake_case : List[str] = ConditionalDetrImageProcessor.from_pretrained("microsoft/conditional-detr-resnet-50" ) __snake_case : List[str] = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : Tuple = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Optional[int] = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : List[Any] = torch.tensor([58_87.96_00, 1_12_50.20_61, 48_93_53.84_38, 83_71_22.75_00, 14_79_67.51_56, 16_57_32.34_38] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : Optional[int] = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([0.55_03, 0.27_65, 0.06_04, 0.22_15] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Optional[Any] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : List[Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : Dict = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify orig_size __snake_case : int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Tuple = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) ) @slow def __snake_case ( self : str ) -> Tuple: # prepare image, target and masks_path __snake_case : Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : str = {"file_name": "000000039769.png", "image_id": 39769, "segments_info": target} __snake_case : Optional[int] = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them __snake_case : int = ConditionalDetrImageProcessor(format="coco_panoptic" ) __snake_case : str = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , masks_path=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : List[str] = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Dict = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : Any = torch.tensor([14_79_79.68_75, 16_55_27.04_69, 48_46_38.59_38, 1_12_92.93_75, 58_79.65_62, 76_34.11_47] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : str = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : Optional[Any] = torch.tensor([0.26_25, 0.54_37, 0.46_88, 0.86_25] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Tuple = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : Dict = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : int = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify masks __snake_case : List[Any] = 822873 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , lowerCamelCase ) # verify orig_size __snake_case : List[str] = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Any = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) )
81
0
from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLModel, ) class _UpperCamelCase: def __init__( self : Any , SCREAMING_SNAKE_CASE__ : List[Any] , ): '''simple docstring''' __a : Optional[int] = parent __a : int = 1_3 __a : Union[str, Any] = 7 __a : Dict = 3_0 __a : Tuple = self.seq_length + self.mem_len __a : Optional[int] = 1_5 __a : Tuple = True __a : Optional[Any] = True __a : str = 9_9 __a : int = [1_0, 5_0, 8_0] __a : Union[str, Any] = 3_2 __a : Dict = 3_2 __a : Tuple = 4 __a : Union[str, Any] = 8 __a : Optional[Any] = 1_2_8 __a : str = 2 __a : str = 2 __a : Dict = None __a : List[str] = 1 __a : Union[str, Any] = 0 __a : Optional[Any] = 3 __a : List[str] = self.vocab_size - 1 __a : int = 0.01 def __lowerCAmelCase ( self : Optional[int] ): '''simple docstring''' __a : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __a : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __a : Tuple = None if self.use_labels: __a : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __a : Optional[int] = TransfoXLConfig( vocab_size=self.vocab_size , mem_len=self.mem_len , clamp_len=self.clamp_len , cutoffs=self.cutoffs , d_model=self.hidden_size , d_embed=self.d_embed , n_head=self.num_attention_heads , d_head=self.d_head , d_inner=self.d_inner , div_val=self.div_val , n_layer=self.num_hidden_layers , eos_token_id=self.eos_token_id , pad_token_id=self.vocab_size - 1 , init_range=self.init_range , num_labels=self.num_labels , ) return (config, input_ids_a, input_ids_a, lm_labels) def __lowerCAmelCase ( self : Optional[int] ): '''simple docstring''' random.seed(self.seed ) tf.random.set_seed(self.seed ) def __lowerCAmelCase ( self : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Any ): '''simple docstring''' __a : Optional[int] = TFTransfoXLModel(SCREAMING_SNAKE_CASE__ ) __a , __a : Dict = model(SCREAMING_SNAKE_CASE__ ).to_tuple() __a : str = {'input_ids': input_ids_a, 'mems': mems_a} __a , __a : Optional[Any] = model(SCREAMING_SNAKE_CASE__ ).to_tuple() self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def __lowerCAmelCase ( self : int , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : str ): '''simple docstring''' __a : str = TFTransfoXLLMHeadModel(SCREAMING_SNAKE_CASE__ ) __a , __a : Dict = model(SCREAMING_SNAKE_CASE__ ).to_tuple() __a : List[str] = {'input_ids': input_ids_a, 'labels': lm_labels} __a , __a : List[str] = model(SCREAMING_SNAKE_CASE__ ).to_tuple() __a , __a : int = model([input_ids_a, mems_a] ).to_tuple() __a : Dict = {'input_ids': input_ids_a, 'mems': mems_a, 'labels': lm_labels} __a , __a : Tuple = model(SCREAMING_SNAKE_CASE__ ).to_tuple() self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def __lowerCAmelCase ( self : Union[str, Any] , SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : str ): '''simple docstring''' __a : Any = TFTransfoXLForSequenceClassification(SCREAMING_SNAKE_CASE__ ) __a : Tuple = model(SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __lowerCAmelCase ( self : List[str] ): '''simple docstring''' __a : str = self.prepare_config_and_inputs() ((__a) , (__a) , (__a) , (__a)) : Union[str, Any] = config_and_inputs __a : str = {'input_ids': input_ids_a} return config, inputs_dict @require_tf class _UpperCamelCase( __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE : List[str] = ( (TFTransfoXLModel, TFTransfoXLLMHeadModel, TFTransfoXLForSequenceClassification) if is_tf_available() else () ) __SCREAMING_SNAKE_CASE : int = () if is_tf_available() else () __SCREAMING_SNAKE_CASE : Dict = ( { '''feature-extraction''': TFTransfoXLModel, '''text-classification''': TFTransfoXLForSequenceClassification, '''text-generation''': TFTransfoXLLMHeadModel, '''zero-shot''': TFTransfoXLForSequenceClassification, } if is_tf_available() else {} ) # TODO: add this test when TFTransfoXLLMHead has a linear output layer implemented __SCREAMING_SNAKE_CASE : List[Any] = False __SCREAMING_SNAKE_CASE : str = False __SCREAMING_SNAKE_CASE : Optional[Any] = False __SCREAMING_SNAKE_CASE : Any = False def __lowerCAmelCase ( self : List[str] , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : Tuple ): '''simple docstring''' if pipeline_test_casse_name == "TextGenerationPipelineTests": # Get `ValueError: AttributeError: 'NoneType' object has no attribute 'new_ones'` or `AssertionError`. # `TransfoXLConfig` was never used in pipeline tests: cannot create a simple # tokenizer. return True return False def __lowerCAmelCase ( self : Optional[int] ): '''simple docstring''' __a : List[str] = TFTransfoXLModelTester(self ) __a : List[str] = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE__ , d_embed=3_7 ) def __lowerCAmelCase ( self : Optional[int] ): '''simple docstring''' self.config_tester.run_common_tests() def __lowerCAmelCase ( self : Any ): '''simple docstring''' self.model_tester.set_seed() __a : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_model(*SCREAMING_SNAKE_CASE__ ) def __lowerCAmelCase ( self : Tuple ): '''simple docstring''' self.model_tester.set_seed() __a : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_lm_head(*SCREAMING_SNAKE_CASE__ ) def __lowerCAmelCase ( self : Tuple ): '''simple docstring''' __a : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_for_sequence_classification(*SCREAMING_SNAKE_CASE__ ) def __lowerCAmelCase ( self : Tuple ): '''simple docstring''' __a , __a : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() __a : Any = [TFTransfoXLForSequenceClassification] for model_class in self.all_model_classes: __a : int = model_class(SCREAMING_SNAKE_CASE__ ) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer ) if model_class in list_other_models_with_output_ebd: __a : Tuple = model.get_output_embeddings() assert isinstance(SCREAMING_SNAKE_CASE__ , tf.keras.layers.Layer ) __a : Optional[int] = model.get_bias() assert name is None else: __a : Optional[Any] = model.get_output_embeddings() assert x is None __a : Union[str, Any] = model.get_bias() assert name is None def __lowerCAmelCase ( self : Dict ): '''simple docstring''' pass @slow def __lowerCAmelCase ( self : Dict ): '''simple docstring''' for model_name in TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a : Optional[int] = TFTransfoXLModel.from_pretrained(SCREAMING_SNAKE_CASE__ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE__ ) @unittest.skip(reason='This model doesn\'t play well with fit() due to not returning a single loss.' ) def __lowerCAmelCase ( self : List[str] ): '''simple docstring''' pass @require_tf class _UpperCamelCase( unittest.TestCase ): @unittest.skip('Skip test until #12651 is resolved.' ) @slow def __lowerCAmelCase ( self : int ): '''simple docstring''' __a : str = TFTransfoXLLMHeadModel.from_pretrained('transfo-xl-wt103' ) # fmt: off __a : Optional[int] = tf.convert_to_tensor([[3_3,1_2_9_7,2,1,1_0_0_9,4,1_1_0_9,1_1_7_3_9,4_7_6_2,3_5_8,5,2_5,2_4_5,2_2,1_7_0_6,1_7,2_0_0_9_8,5,3_2_1_5,2_1,3_7,1_1_1_0,3,1_3,1_0_4_1,4,2_4,6_0_3,4_9_0,2,7_1_4_7_7,2_0_0_9_8,1_0_4_4_4_7,2,2_0_9_6_1,1,2_6_0_4,4,1,3_2_9,3,6_2_2_4,8_3_1,1_6_0_0_2,2,8,6_0_3,7_8_9_6_7,2_9_5_4_6,2_3,8_0_3,2_0,2_5,4_1_6,5,8,2_3_2,4,2_7_7,6,1_8_5_5,4_6_0_1,3,2_9_5_4_6,5_4,8,3_6_0_9,5,5_7_2_1_1,4_9,4,1,2_7_7,1_8,8,1_7_5_5,1_5_6_9_1,3,3_4_1,2_5,4_1_6,6_9_3,4_2_5_7_3,7_1,1_7,4_0_1,9_4,3_1,1_7_9_1_9,2,2_9_5_4_6,7_8_7_3,1_8,1,4_3_5,2_3,1_1_0_1_1,7_5_5,5,5_1_6_7,3,7_9_8_3,9_8,8_4,2,2_9_5_4_6,3_2_6_7,8,3_6_0_9,4,1,4_8_6_5,1_0_7_5,2,6_0_8_7,7_1,6,3_4_6,8,5_8_5_4,3,2_9_5_4_6,8_2_4,1_4_0_0,1_8_6_8,2,1_9,1_6_0,2,3_1_1,8,5_4_9_6,2,2_0_9_2_0,1_7,2_5,1_5_0_9_7,3,2_4,2_4,0]] , dtype=tf.intaa ) # noqa: E231 # fmt: on # In 1991 , the remains of Russian Tsar Nicholas II and his family # ( except for Alexei and Maria ) are discovered . # The voice of Nicholas's young son , Tsarevich Alexei Nikolaevich , narrates the # remainder of the story . 1883 Western Siberia , # a young Grigori Rasputin is asked by his father and a group of men to perform magic . # Rasputin has a vision and denounces one of the men as a horse thief . Although his # father initially slaps him for making such an accusation , Rasputin watches as the # man is chased outside and beaten . Twenty years later , Rasputin sees a vision of # the Virgin Mary , prompting him to become a priest . Rasputin quickly becomes famous , # with people , even a bishop , begging for his blessing . <eod> </s> <eos> # fmt: off __a : List[Any] = [3_3,1_2_9_7,2,1,1_0_0_9,4,1_1_0_9,1_1_7_3_9,4_7_6_2,3_5_8,5,2_5,2_4_5,2_2,1_7_0_6,1_7,2_0_0_9_8,5,3_2_1_5,2_1,3_7,1_1_1_0,3,1_3,1_0_4_1,4,2_4,6_0_3,4_9_0,2,7_1_4_7_7,2_0_0_9_8,1_0_4_4_4_7,2,2_0_9_6_1,1,2_6_0_4,4,1,3_2_9,3,6_2_2_4,8_3_1,1_6_0_0_2,2,8,6_0_3,7_8_9_6_7,2_9_5_4_6,2_3,8_0_3,2_0,2_5,4_1_6,5,8,2_3_2,4,2_7_7,6,1_8_5_5,4_6_0_1,3,2_9_5_4_6,5_4,8,3_6_0_9,5,5_7_2_1_1,4_9,4,1,2_7_7,1_8,8,1_7_5_5,1_5_6_9_1,3,3_4_1,2_5,4_1_6,6_9_3,4_2_5_7_3,7_1,1_7,4_0_1,9_4,3_1,1_7_9_1_9,2,2_9_5_4_6,7_8_7_3,1_8,1,4_3_5,2_3,1_1_0_1_1,7_5_5,5,5_1_6_7,3,7_9_8_3,9_8,8_4,2,2_9_5_4_6,3_2_6_7,8,3_6_0_9,4,1,4_8_6_5,1_0_7_5,2,6_0_8_7,7_1,6,3_4_6,8,5_8_5_4,3,2_9_5_4_6,8_2_4,1_4_0_0,1_8_6_8,2,1_9,1_6_0,2,3_1_1,8,5_4_9_6,2,2_0_9_2_0,1_7,2_5,1_5_0_9_7,3,2_4,2_4,0,3_3,1,1_8_5_7,2,1,1_0_0_9,4,1_1_0_9,1_1_7_3_9,4_7_6_2,3_5_8,5,2_5,2_4_5,2_8,1_1_1_0,3,1_3,1_0_4_1,4,2_4,6_0_3,4_9_0,2,7_1_4_7_7,2_0_0_9_8,1_0_4_4_4_7,2,2_0_9_6_1,1,2_6_0_4,4,1,3_2_9,3,0] # noqa: E231 # fmt: on # In 1991, the remains of Russian Tsar Nicholas II and his family ( # except for Alexei and Maria ) are discovered. The voice of young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story. # 1883 Western Siberia, a young Grigori Rasputin is asked by his father # and a group of men to perform magic. Rasputin has a vision and # denounces one of the men as a horse thief. Although his father initially # slaps him for making such an accusation, Rasputin watches as the man # is chased outside and beaten. Twenty years later, Rasputin sees a vision # of the Virgin Mary, prompting him to become a priest. # Rasputin quickly becomes famous, with people, even a bishop, begging for # his blessing. <unk> <unk> <eos> In the 1990s, the remains of Russian Tsar # Nicholas II and his family were discovered. The voice of <unk> young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story.<eos> __a : List[str] = model.generate(SCREAMING_SNAKE_CASE__ , max_length=2_0_0 , do_sample=SCREAMING_SNAKE_CASE__ ) self.assertListEqual(output_ids[0].numpy().tolist() , SCREAMING_SNAKE_CASE__ )
47
from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from ... import AutoBackbone from ...modeling_outputs import SemanticSegmenterOutput from ...modeling_utils import PreTrainedModel from ...utils import add_start_docstrings, add_start_docstrings_to_model_forward, replace_return_docstrings from ...utils.backbone_utils import BackboneMixin from .configuration_upernet import UperNetConfig _snake_case : Optional[Any] = [ "openmmlab/upernet-convnext-tiny", # See all UperNet models at https://huggingface.co/models?filter=upernet ] # General docstring _snake_case : Dict = "UperNetConfig" class a (nn.Module ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : Union[int, Tuple[int, int]] , lowerCamelCase : Union[int, Tuple[int, int], str] = 0 , lowerCamelCase : bool = False , lowerCamelCase : Union[int, Tuple[int, int]] = 1 , ) -> None: super().__init__() __snake_case : Union[str, Any] = nn.Convad( in_channels=lowerCamelCase , out_channels=lowerCamelCase , kernel_size=lowerCamelCase , padding=lowerCamelCase , bias=lowerCamelCase , dilation=lowerCamelCase , ) __snake_case : Dict = nn.BatchNormad(lowerCamelCase ) __snake_case : List[Any] = nn.ReLU() def __snake_case ( self : List[Any] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : Dict = self.conv(lowerCamelCase ) __snake_case : int = self.batch_norm(lowerCamelCase ) __snake_case : Optional[Any] = self.activation(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : str , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : int ) -> None: super().__init__() __snake_case : Tuple = [ nn.AdaptiveAvgPoolad(lowerCamelCase ), UperNetConvModule(lowerCamelCase , lowerCamelCase , kernel_size=1 ), ] for i, layer in enumerate(self.layers ): self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : Dict , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : List[str] = input for layer in self.layers: __snake_case : Tuple = layer(lowerCamelCase ) return hidden_state class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Tuple[int, ...] , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : bool ) -> None: super().__init__() __snake_case : Dict = pool_scales __snake_case : List[str] = align_corners __snake_case : List[Any] = in_channels __snake_case : str = channels __snake_case : Optional[Any] = [] for i, pool_scale in enumerate(lowerCamelCase ): __snake_case : Union[str, Any] = UperNetPyramidPoolingBlock(pool_scale=lowerCamelCase , in_channels=lowerCamelCase , channels=lowerCamelCase ) self.blocks.append(lowerCamelCase ) self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> List[torch.Tensor]: __snake_case : Tuple = [] for ppm in self.blocks: __snake_case : Any = ppm(lowerCamelCase ) __snake_case : List[Any] = nn.functional.interpolate( lowerCamelCase , size=x.size()[2:] , mode="bilinear" , align_corners=self.align_corners ) ppm_outs.append(lowerCamelCase ) return ppm_outs class a (nn.Module ): """simple docstring""" def __init__( self : int , lowerCamelCase : str , lowerCamelCase : str ) -> List[str]: super().__init__() __snake_case : Dict = config __snake_case : List[str] = config.pool_scales # e.g. (1, 2, 3, 6) __snake_case : Tuple = in_channels __snake_case : str = config.hidden_size __snake_case : List[str] = False __snake_case : Any = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) # PSP Module __snake_case : Tuple = UperNetPyramidPoolingModule( self.pool_scales , self.in_channels[-1] , self.channels , align_corners=self.align_corners , ) __snake_case : List[str] = UperNetConvModule( self.in_channels[-1] + len(self.pool_scales ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) # FPN Module __snake_case : List[Any] = nn.ModuleList() __snake_case : Dict = nn.ModuleList() for in_channels in self.in_channels[:-1]: # skip the top layer __snake_case : Union[str, Any] = UperNetConvModule(lowerCamelCase , self.channels , kernel_size=1 ) __snake_case : Optional[int] = UperNetConvModule(self.channels , self.channels , kernel_size=3 , padding=1 ) self.lateral_convs.append(lowerCamelCase ) self.fpn_convs.append(lowerCamelCase ) __snake_case : int = UperNetConvModule( len(self.in_channels ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) def __snake_case ( self : List[str] ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> str: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : List[Any] , lowerCamelCase : Tuple ) -> Optional[int]: __snake_case : str = inputs[-1] __snake_case : int = [x] psp_outs.extend(self.psp_modules(lowerCamelCase ) ) __snake_case : Tuple = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Union[str, Any] = self.bottleneck(lowerCamelCase ) return output def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # build laterals __snake_case : Any = [lateral_conv(encoder_hidden_states[i] ) for i, lateral_conv in enumerate(self.lateral_convs )] laterals.append(self.psp_forward(lowerCamelCase ) ) # build top-down path __snake_case : Dict = len(lowerCamelCase ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Union[str, Any] = laterals[i - 1].shape[2:] __snake_case : Optional[int] = laterals[i - 1] + nn.functional.interpolate( laterals[i] , size=lowerCamelCase , mode="bilinear" , align_corners=self.align_corners ) # build outputs __snake_case : str = [self.fpn_convs[i](laterals[i] ) for i in range(used_backbone_levels - 1 )] # append psp feature fpn_outs.append(laterals[-1] ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Tuple = nn.functional.interpolate( fpn_outs[i] , size=fpn_outs[0].shape[2:] , mode="bilinear" , align_corners=self.align_corners ) __snake_case : str = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Optional[Any] = self.fpn_bottleneck(lowerCamelCase ) __snake_case : Tuple = self.classifier(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Optional[int] , lowerCamelCase : int = 2 , lowerCamelCase : int = 3 , lowerCamelCase : Union[int, Tuple[int, int]] = 1 ) -> None: super().__init__() __snake_case : List[Any] = config __snake_case : List[str] = config.auxiliary_in_channels __snake_case : List[Any] = config.auxiliary_channels __snake_case : Tuple = config.auxiliary_num_convs __snake_case : int = config.auxiliary_concat_input __snake_case : Optional[int] = in_index __snake_case : Tuple = (kernel_size // 2) * dilation __snake_case : Optional[int] = [] convs.append( UperNetConvModule( self.in_channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) for i in range(self.num_convs - 1 ): convs.append( UperNetConvModule( self.channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) if self.num_convs == 0: __snake_case : Union[str, Any] = nn.Identity() else: __snake_case : Any = nn.Sequential(*lowerCamelCase ) if self.concat_input: __snake_case : int = UperNetConvModule( self.in_channels + self.channels , self.channels , kernel_size=lowerCamelCase , padding=kernel_size // 2 ) __snake_case : Optional[Any] = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) def __snake_case ( self : Dict ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : Tuple ) -> Optional[int]: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : Optional[int] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # just take the relevant feature maps __snake_case : List[str] = encoder_hidden_states[self.in_index] __snake_case : Optional[Any] = self.convs(lowerCamelCase ) if self.concat_input: __snake_case : Optional[Any] = self.conv_cat(torch.cat([hidden_states, output] , dim=1 ) ) __snake_case : Union[str, Any] = self.classifier(lowerCamelCase ) return output class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = UperNetConfig __UpperCAmelCase : int = "pixel_values" __UpperCAmelCase : str = True def __snake_case ( self : Union[str, Any] , lowerCamelCase : List[Any] ) -> List[Any]: if isinstance(lowerCamelCase , lowerCamelCase ): module.backbone.init_weights() module.decode_head.init_weights() module.auxiliary_head.init_weights() def __snake_case ( self : Optional[Any] ) -> List[str]: self.backbone.init_weights() self.decode_head.init_weights() self.auxiliary_head.init_weights() def __snake_case ( self : List[Any] , lowerCamelCase : Any , lowerCamelCase : Optional[Any]=False ) -> Dict: if isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : Union[str, Any] = value _snake_case : Dict = R"\n Parameters:\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n config ([`UperNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" _snake_case : Tuple = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using\n [`AutoImageProcessor`]. See [`SegformerImageProcessor.__call__`] for details.\n output_attentions (`bool`, *optional*):\n Whether or not to return the attentions tensors of all attention layers in case the backbone has them. See\n `attentions` under returned tensors for more detail.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers of the backbone. See `hidden_states` under\n returned tensors for more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n" @add_start_docstrings( "UperNet framework leveraging any vision backbone e.g. for ADE20k, CityScapes." , _lowerCAmelCase , ) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Dict , lowerCamelCase : int ) -> Optional[int]: super().__init__(lowerCamelCase ) __snake_case : Any = AutoBackbone.from_config(config.backbone_config ) # Semantic segmentation head(s) __snake_case : Union[str, Any] = UperNetHead(lowerCamelCase , in_channels=self.backbone.channels ) __snake_case : Any = UperNetFCNHead(lowerCamelCase ) if config.use_auxiliary_head else None # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UPERNET_INPUTS_DOCSTRING.format("batch_size, sequence_length" ) ) @replace_return_docstrings(output_type=lowerCamelCase , config_class=_CONFIG_FOR_DOC ) def __snake_case ( self : Union[str, Any] , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , ) -> Union[tuple, SemanticSegmenterOutput]: __snake_case : Any = return_dict if return_dict is not None else self.config.use_return_dict __snake_case : int = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __snake_case : str = output_attentions if output_attentions is not None else self.config.output_attentions __snake_case : Tuple = self.backbone.forward_with_filtered_kwargs( lowerCamelCase , output_hidden_states=lowerCamelCase , output_attentions=lowerCamelCase ) __snake_case : List[Any] = outputs.feature_maps __snake_case : List[Any] = self.decode_head(lowerCamelCase ) __snake_case : List[str] = nn.functional.interpolate(lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : Optional[int] = None if self.auxiliary_head is not None: __snake_case : Dict = self.auxiliary_head(lowerCamelCase ) __snake_case : Dict = nn.functional.interpolate( lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : int = None if labels is not None: if self.config.num_labels == 1: raise ValueError("The number of labels should be greater than one" ) else: # compute weighted loss __snake_case : Any = CrossEntropyLoss(ignore_index=self.config.loss_ignore_index ) __snake_case : Union[str, Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = main_loss + self.config.auxiliary_loss_weight * auxiliary_loss if not return_dict: if output_hidden_states: __snake_case : Any = (logits,) + outputs[1:] else: __snake_case : str = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return SemanticSegmenterOutput( loss=lowerCamelCase , logits=lowerCamelCase , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
81
0
'''simple docstring''' import os import pytest import yaml from datasets.features.features import Features, Value from datasets.info import DatasetInfo, DatasetInfosDict @pytest.mark.parametrize( "files" , [ ["full:README.md", "dataset_infos.json"], ["empty:README.md", "dataset_infos.json"], ["dataset_infos.json"], ["full:README.md"], ] , ) def A ( UpperCamelCase_ : Tuple , UpperCamelCase_ : Optional[Any] ) -> List[Any]: '''simple docstring''' lowerCAmelCase__ = tmp_path_factory.mktemp("dset_infos_dir" ) if "full:README.md" in files: with open(dataset_infos_dir / "README.md" , "w" ) as f: f.write("---\ndataset_info:\n dataset_size: 42\n---" ) if "empty:README.md" in files: with open(dataset_infos_dir / "README.md" , "w" ) as f: f.write("" ) # we want to support dataset_infos.json for backward compatibility if "dataset_infos.json" in files: with open(dataset_infos_dir / "dataset_infos.json" , "w" ) as f: f.write("{\"default\": {\"dataset_size\": 42}}" ) lowerCAmelCase__ = DatasetInfosDict.from_directory(UpperCamelCase_ ) assert dataset_infos assert dataset_infos["default"].dataset_size == 42 @pytest.mark.parametrize( "dataset_info" , [ DatasetInfo(), DatasetInfo( description="foo" , features=Features({"a": Value("int32" )} ) , builder_name="builder" , config_name="config" , version="1.0.0" , splits=[{"name": "train"}] , download_size=42 , ), ] , ) def A ( UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : DatasetInfo ) -> Optional[int]: '''simple docstring''' lowerCAmelCase__ = str(UpperCamelCase_ ) dataset_info.write_to_directory(UpperCamelCase_ ) lowerCAmelCase__ = DatasetInfo.from_directory(UpperCamelCase_ ) assert dataset_info == reloaded assert os.path.exists(os.path.join(UpperCamelCase_ , "dataset_info.json" ) ) def A ( ) -> List[Any]: '''simple docstring''' lowerCAmelCase__ = DatasetInfo( description="foo" , citation="bar" , homepage="https://foo.bar" , license="CC0" , features=Features({"a": Value("int32" )} ) , post_processed={} , supervised_keys=() , task_templates=[] , builder_name="builder" , config_name="config" , version="1.0.0" , splits=[{"name": "train", "num_examples": 42}] , download_checksums={} , download_size=13_37 , post_processing_size=4_42 , dataset_size=12_34 , size_in_bytes=13_37 + 4_42 + 12_34 , ) lowerCAmelCase__ = dataset_info._to_yaml_dict() assert sorted(UpperCamelCase_ ) == sorted(DatasetInfo._INCLUDED_INFO_IN_YAML ) for key in DatasetInfo._INCLUDED_INFO_IN_YAML: assert key in dataset_info_yaml_dict assert isinstance(dataset_info_yaml_dict[key] , (list, dict, int, str) ) lowerCAmelCase__ = yaml.safe_dump(UpperCamelCase_ ) lowerCAmelCase__ = yaml.safe_load(UpperCamelCase_ ) assert dataset_info_yaml_dict == reloaded def A ( ) -> str: '''simple docstring''' lowerCAmelCase__ = DatasetInfo() lowerCAmelCase__ = dataset_info._to_yaml_dict() assert dataset_info_yaml_dict == {} @pytest.mark.parametrize( "dataset_infos_dict" , [ DatasetInfosDict(), DatasetInfosDict({"default": DatasetInfo()} ), DatasetInfosDict({"my_config_name": DatasetInfo()} ), DatasetInfosDict( { "default": DatasetInfo( description="foo" , features=Features({"a": Value("int32" )} ) , builder_name="builder" , config_name="config" , version="1.0.0" , splits=[{"name": "train"}] , download_size=42 , ) } ), DatasetInfosDict( { "v1": DatasetInfo(dataset_size=42 ), "v2": DatasetInfo(dataset_size=13_37 ), } ), ] , ) def A ( UpperCamelCase_ : Dict , UpperCamelCase_ : DatasetInfosDict ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase__ = str(UpperCamelCase_ ) dataset_infos_dict.write_to_directory(UpperCamelCase_ ) lowerCAmelCase__ = DatasetInfosDict.from_directory(UpperCamelCase_ ) # the config_name of the dataset_infos_dict take over the attribute for config_name, dataset_info in dataset_infos_dict.items(): lowerCAmelCase__ = config_name # the yaml representation doesn't include fields like description or citation # so we just test that we can recover what we can from the yaml lowerCAmelCase__ = DatasetInfo._from_yaml_dict(dataset_info._to_yaml_dict() ) assert dataset_infos_dict == reloaded if dataset_infos_dict: assert os.path.exists(os.path.join(UpperCamelCase_ , "README.md" ) )
48
from __future__ import annotations def lowerCAmelCase_ ( __lowerCamelCase ): if not nums: return 0 __snake_case : Any = nums[0] __snake_case : str = 0 for num in nums[1:]: __snake_case , __snake_case : List[str] = ( max_excluding + num, max(__lowerCamelCase , __lowerCamelCase ), ) return max(__lowerCamelCase , __lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
81
0
"""simple docstring""" # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _lowercase : List[Any] = { 'configuration_vivit': ['VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'VivitConfig'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : Dict = ['VivitImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : List[str] = [ 'VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'VivitModel', 'VivitPreTrainedModel', 'VivitForVideoClassification', ] if TYPE_CHECKING: from .configuration_vivit import VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, VivitConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_vivit import VivitImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vivit import ( VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST, VivitForVideoClassification, VivitModel, VivitPreTrainedModel, ) else: import sys _lowercase : Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
49
from __future__ import annotations from typing import Any def lowerCAmelCase_ ( __lowerCamelCase ): create_state_space_tree(__lowerCamelCase , [] , 0 ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): if index == len(__lowerCamelCase ): print(__lowerCamelCase ) return create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.pop() if __name__ == "__main__": _snake_case : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["A", "B", "C"]) generate_all_subsequences(seq)
81
0
'''simple docstring''' import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class UpperCamelCase__ : '''simple docstring''' def __init__( self ,_lowerCAmelCase ,_lowerCAmelCase=13 ,_lowerCAmelCase=32 ,_lowerCAmelCase=2 ,_lowerCAmelCase=3 ,_lowerCAmelCase=16 ,_lowerCAmelCase=[1, 2, 1] ,_lowerCAmelCase=[2, 2, 4] ,_lowerCAmelCase=2 ,_lowerCAmelCase=2.0 ,_lowerCAmelCase=True ,_lowerCAmelCase=0.0 ,_lowerCAmelCase=0.0 ,_lowerCAmelCase=0.1 ,_lowerCAmelCase="gelu" ,_lowerCAmelCase=False ,_lowerCAmelCase=True ,_lowerCAmelCase=0.02 ,_lowerCAmelCase=1E-5 ,_lowerCAmelCase=True ,_lowerCAmelCase=None ,_lowerCAmelCase=True ,_lowerCAmelCase=10 ,_lowerCAmelCase=8 ,): lowerCamelCase__ = parent lowerCamelCase__ = batch_size lowerCamelCase__ = image_size lowerCamelCase__ = patch_size lowerCamelCase__ = num_channels lowerCamelCase__ = embed_dim lowerCamelCase__ = depths lowerCamelCase__ = num_heads lowerCamelCase__ = window_size lowerCamelCase__ = mlp_ratio lowerCamelCase__ = qkv_bias lowerCamelCase__ = hidden_dropout_prob lowerCamelCase__ = attention_probs_dropout_prob lowerCamelCase__ = drop_path_rate lowerCamelCase__ = hidden_act lowerCamelCase__ = use_absolute_embeddings lowerCamelCase__ = patch_norm lowerCamelCase__ = layer_norm_eps lowerCamelCase__ = initializer_range lowerCamelCase__ = is_training lowerCamelCase__ = scope lowerCamelCase__ = use_labels lowerCamelCase__ = type_sequence_label_size lowerCamelCase__ = encoder_stride def UpperCamelCase_ ( self ): lowerCamelCase__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ = None if self.use_labels: lowerCamelCase__ = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) lowerCamelCase__ = self.get_config() return config, pixel_values, labels def UpperCamelCase_ ( self ): return SwinvaConfig( image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,embed_dim=self.embed_dim ,depths=self.depths ,num_heads=self.num_heads ,window_size=self.window_size ,mlp_ratio=self.mlp_ratio ,qkv_bias=self.qkv_bias ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,drop_path_rate=self.drop_path_rate ,hidden_act=self.hidden_act ,use_absolute_embeddings=self.use_absolute_embeddings ,path_norm=self.patch_norm ,layer_norm_eps=self.layer_norm_eps ,initializer_range=self.initializer_range ,encoder_stride=self.encoder_stride ,) def UpperCamelCase_ ( self ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ): lowerCamelCase__ = SwinvaModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() lowerCamelCase__ = model(_lowerCAmelCase ) lowerCamelCase__ = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) lowerCamelCase__ = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, expected_seq_len, expected_dim) ) def UpperCamelCase_ ( self ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ): lowerCamelCase__ = SwinvaForMaskedImageModeling(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() lowerCamelCase__ = model(_lowerCAmelCase ) self.parent.assertEqual( result.logits.shape ,(self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images lowerCamelCase__ = 1 lowerCamelCase__ = SwinvaForMaskedImageModeling(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() lowerCamelCase__ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCamelCase__ = model(_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, 1, self.image_size, self.image_size) ) def UpperCamelCase_ ( self ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ): lowerCamelCase__ = self.type_sequence_label_size lowerCamelCase__ = SwinvaForImageClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() lowerCamelCase__ = model(_lowerCAmelCase ,labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) def UpperCamelCase_ ( self ): lowerCamelCase__ = self.prepare_config_and_inputs() lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = config_and_inputs lowerCamelCase__ = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class UpperCamelCase__ (a ,a ,unittest.TestCase ): '''simple docstring''' _UpperCamelCase = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) _UpperCamelCase = ( {'feature-extraction': SwinvaModel, 'image-classification': SwinvaForImageClassification} if is_torch_available() else {} ) _UpperCamelCase = False _UpperCamelCase = False _UpperCamelCase = False _UpperCamelCase = False def UpperCamelCase_ ( self ): lowerCamelCase__ = SwinvaModelTester(self ) lowerCamelCase__ = ConfigTester(self ,config_class=_lowerCAmelCase ,embed_dim=37 ) def UpperCamelCase_ ( self ): self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase_ ( self ): lowerCamelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) @unittest.skip(reason="""Got `CUDA error: misaligned address` with PyTorch 2.0.0.""" ) def UpperCamelCase_ ( self ): pass @unittest.skip(reason="""Swinv2 does not use inputs_embeds""" ) def UpperCamelCase_ ( self ): pass def UpperCamelCase_ ( self ): lowerCamelCase__ , lowerCamelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ = model_class(_lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() ,(nn.Module) ) lowerCamelCase__ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCAmelCase ,nn.Linear ) ) def UpperCamelCase_ ( self ): lowerCamelCase__ , lowerCamelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ = model_class(_lowerCAmelCase ) lowerCamelCase__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ = [*signature.parameters.keys()] lowerCamelCase__ = ["""pixel_values"""] self.assertListEqual(arg_names[:1] ,_lowerCAmelCase ) def UpperCamelCase_ ( self ): lowerCamelCase__ , lowerCamelCase__ = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ = True for model_class in self.all_model_classes: lowerCamelCase__ = True lowerCamelCase__ = False lowerCamelCase__ = True lowerCamelCase__ = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): lowerCamelCase__ = model(**self._prepare_for_class(_lowerCAmelCase ,_lowerCAmelCase ) ) lowerCamelCase__ = outputs.attentions lowerCamelCase__ = len(self.model_tester.depths ) self.assertEqual(len(_lowerCAmelCase ) ,_lowerCAmelCase ) # check that output_attentions also work using config del inputs_dict["output_attentions"] lowerCamelCase__ = True lowerCamelCase__ = config.window_size**2 lowerCamelCase__ = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): lowerCamelCase__ = model(**self._prepare_for_class(_lowerCAmelCase ,_lowerCAmelCase ) ) lowerCamelCase__ = outputs.attentions self.assertEqual(len(_lowerCAmelCase ) ,_lowerCAmelCase ) self.assertListEqual( list(attentions[0].shape[-3:] ) ,[self.model_tester.num_heads[0], window_size_squared, window_size_squared] ,) lowerCamelCase__ = len(_lowerCAmelCase ) # Check attention is always last and order is fine lowerCamelCase__ = True lowerCamelCase__ = True lowerCamelCase__ = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): lowerCamelCase__ = model(**self._prepare_for_class(_lowerCAmelCase ,_lowerCAmelCase ) ) if hasattr(self.model_tester ,"""num_hidden_states_types""" ): lowerCamelCase__ = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states lowerCamelCase__ = 2 self.assertEqual(out_len + added_hidden_states ,len(_lowerCAmelCase ) ) lowerCamelCase__ = outputs.attentions self.assertEqual(len(_lowerCAmelCase ) ,_lowerCAmelCase ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) ,[self.model_tester.num_heads[0], window_size_squared, window_size_squared] ,) def UpperCamelCase_ ( self ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ): lowerCamelCase__ = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): lowerCamelCase__ = model(**self._prepare_for_class(_lowerCAmelCase ,_lowerCAmelCase ) ) lowerCamelCase__ = outputs.hidden_states lowerCamelCase__ = getattr( self.model_tester ,"""expected_num_hidden_layers""" ,len(self.model_tester.depths ) + 1 ) self.assertEqual(len(_lowerCAmelCase ) ,_lowerCAmelCase ) # Swinv2 has a different seq_length lowerCamelCase__ = ( config.patch_size if isinstance(config.patch_size ,collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowerCamelCase__ = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) ,[num_patches, self.model_tester.embed_dim] ,) lowerCamelCase__ = outputs.reshaped_hidden_states self.assertEqual(len(_lowerCAmelCase ) ,_lowerCAmelCase ) lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = reshaped_hidden_states[0].shape lowerCamelCase__ = ( reshaped_hidden_states[0].view(_lowerCAmelCase ,_lowerCAmelCase ,height * width ).permute(0 ,2 ,1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) ,[num_patches, self.model_tester.embed_dim] ,) def UpperCamelCase_ ( self ): lowerCamelCase__ , lowerCamelCase__ = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ = ( self.model_tester.image_size if isinstance(self.model_tester.image_size ,collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: lowerCamelCase__ = True self.check_hidden_states_output(_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCamelCase__ = True self.check_hidden_states_output(_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ) def UpperCamelCase_ ( self ): lowerCamelCase__ , lowerCamelCase__ = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ = 3 lowerCamelCase__ = ( self.model_tester.image_size if isinstance(self.model_tester.image_size ,collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) lowerCamelCase__ = ( config.patch_size if isinstance(config.patch_size ,collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowerCamelCase__ = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) lowerCamelCase__ = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: lowerCamelCase__ = True self.check_hidden_states_output(_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ,(padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCamelCase__ = True self.check_hidden_states_output(_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ,(padded_height, padded_width) ) def UpperCamelCase_ ( self ): lowerCamelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_lowerCAmelCase ) def UpperCamelCase_ ( self ): lowerCamelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) @slow def UpperCamelCase_ ( self ): for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ = SwinvaModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def UpperCamelCase_ ( self ): lowerCamelCase__ , lowerCamelCase__ = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ = _config_zero_init(_lowerCAmelCase ) for model_class in self.all_model_classes: lowerCamelCase__ = model_class(config=_lowerCAmelCase ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() ,[0.0, 1.0] ,msg=F'''Parameter {name} of model {model_class} seems not properly initialized''' ,) @require_vision @require_torch class UpperCamelCase__ (unittest.TestCase ): '''simple docstring''' @cached_property def UpperCamelCase_ ( self ): return ( AutoImageProcessor.from_pretrained("""microsoft/swinv2-tiny-patch4-window8-256""" ) if is_vision_available() else None ) @slow def UpperCamelCase_ ( self ): lowerCamelCase__ = SwinvaForImageClassification.from_pretrained("""microsoft/swinv2-tiny-patch4-window8-256""" ).to( _lowerCAmelCase ) lowerCamelCase__ = self.default_image_processor lowerCamelCase__ = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) lowerCamelCase__ = image_processor(images=_lowerCAmelCase ,return_tensors="""pt""" ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): lowerCamelCase__ = model(**_lowerCAmelCase ) # verify the logits lowerCamelCase__ = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape ,_lowerCAmelCase ) lowerCamelCase__ = torch.tensor([-0.3947, -0.4306, 0.0026] ).to(_lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,_lowerCAmelCase ,atol=1E-4 ) )
50
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case , __snake_case : List[Any] = image.size __snake_case , __snake_case : Tuple = (x - x % 3_2 for x in (w, h)) # resize to integer multiple of 32 __snake_case : str = image.resize((w, h) , resample=PIL_INTERPOLATION["lanczos"] ) __snake_case : int = np.array(__lowerCamelCase ).astype(np.floataa ) / 2_5_5.0 __snake_case : Union[str, Any] = image[None].transpose(0 , 3 , 1 , 2 ) __snake_case : Union[str, Any] = torch.from_numpy(__lowerCamelCase ) return 2.0 * image - 1.0 class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : VQModel , lowerCamelCase : UNetaDModel , lowerCamelCase : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ) -> Union[str, Any]: super().__init__() self.register_modules(vqvae=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase ) @torch.no_grad() def __call__( self : List[str] , lowerCamelCase : Union[torch.Tensor, PIL.Image.Image] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : Optional[int] = 100 , lowerCamelCase : Optional[float] = 0.0 , lowerCamelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , ) -> Union[Tuple, ImagePipelineOutput]: if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : Any = 1 elif isinstance(lowerCamelCase , torch.Tensor ): __snake_case : Any = image.shape[0] else: raise ValueError(F'`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(lowerCamelCase )}' ) if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : List[Any] = preprocess(lowerCamelCase ) __snake_case , __snake_case : int = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image __snake_case : str = (batch_size, self.unet.config.in_channels // 2, height, width) __snake_case : str = next(self.unet.parameters() ).dtype __snake_case : Tuple = randn_tensor(lowerCamelCase , generator=lowerCamelCase , device=self.device , dtype=lowerCamelCase ) __snake_case : List[Any] = image.to(device=self.device , dtype=lowerCamelCase ) # set timesteps and move to the correct device self.scheduler.set_timesteps(lowerCamelCase , device=self.device ) __snake_case : str = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler __snake_case : Dict = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] __snake_case : Union[str, Any] = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __snake_case : int = {} if accepts_eta: __snake_case : List[str] = eta for t in self.progress_bar(lowerCamelCase ): # concat latents and low resolution image in the channel dimension. __snake_case : Union[str, Any] = torch.cat([latents, image] , dim=1 ) __snake_case : Optional[Any] = self.scheduler.scale_model_input(lowerCamelCase , lowerCamelCase ) # predict the noise residual __snake_case : int = self.unet(lowerCamelCase , lowerCamelCase ).sample # compute the previous noisy sample x_t -> x_t-1 __snake_case : Union[str, Any] = self.scheduler.step(lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase ).prev_sample # decode the image latents with the VQVAE __snake_case : List[Any] = self.vqvae.decode(lowerCamelCase ).sample __snake_case : Dict = torch.clamp(lowerCamelCase , -1.0 , 1.0 ) __snake_case : Any = image / 2 + 0.5 __snake_case : Tuple = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __snake_case : Tuple = self.numpy_to_pil(lowerCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCamelCase )
81
0
'''simple docstring''' from __future__ import annotations import csv import requests from bsa import BeautifulSoup def __snake_case ( SCREAMING_SNAKE_CASE_ : str = "" ) -> dict[str, float]: """simple docstring""" UpperCAmelCase = url or '''https://www.imdb.com/chart/top/?ref_=nv_mv_250''' UpperCAmelCase = BeautifulSoup(requests.get(SCREAMING_SNAKE_CASE_ ).text , '''html.parser''' ) UpperCAmelCase = soup.find_all('''td''' , attrs='''titleColumn''' ) UpperCAmelCase = soup.find_all('''td''' , class_='''ratingColumn imdbRating''' ) return { title.a.text: float(rating.strong.text ) for title, rating in zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) } def __snake_case ( SCREAMING_SNAKE_CASE_ : str = "IMDb_Top_250_Movies.csv" ) -> None: """simple docstring""" UpperCAmelCase = get_imdb_top_aaa_movies() with open(SCREAMING_SNAKE_CASE_ , '''w''' , newline='''''' ) as out_file: UpperCAmelCase = csv.writer(SCREAMING_SNAKE_CASE_ ) writer.writerow(['''Movie title''', '''IMDb rating'''] ) for title, rating in movies.items(): writer.writerow([title, rating] ) if __name__ == "__main__": write_movies()
51
import gc import unittest import torch from parameterized import parameterized from diffusers import AutoencoderKL from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = AutoencoderKL __UpperCAmelCase : Optional[Any] = "sample" __UpperCAmelCase : Optional[int] = 1e-2 @property def __snake_case ( self : Dict ) -> Optional[Any]: __snake_case : Optional[Any] = 4 __snake_case : Tuple = 3 __snake_case : List[str] = (32, 32) __snake_case : str = floats_tensor((batch_size, num_channels) + sizes ).to(lowerCamelCase ) return {"sample": image} @property def __snake_case ( self : Union[str, Any] ) -> Tuple: return (3, 32, 32) @property def __snake_case ( self : int ) -> int: return (3, 32, 32) def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Optional[Any] = { "block_out_channels": [32, 64], "in_channels": 3, "out_channels": 3, "down_block_types": ["DownEncoderBlock2D", "DownEncoderBlock2D"], "up_block_types": ["UpDecoderBlock2D", "UpDecoderBlock2D"], "latent_channels": 4, } __snake_case : Any = self.dummy_input return init_dict, inputs_dict def __snake_case ( self : str ) -> Dict: pass def __snake_case ( self : Tuple ) -> List[str]: pass @unittest.skipIf(torch_device == "mps" , "Gradient checkpointing skipped on MPS" ) def __snake_case ( self : Any ) -> Optional[Any]: # enable deterministic behavior for gradient checkpointing __snake_case , __snake_case : int = self.prepare_init_args_and_inputs_for_common() __snake_case : str = self.model_class(**lowerCamelCase ) model.to(lowerCamelCase ) assert not model.is_gradient_checkpointing and model.training __snake_case : str = model(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model.zero_grad() __snake_case : Any = torch.randn_like(lowerCamelCase ) __snake_case : str = (out - labels).mean() loss.backward() # re-instantiate the model now enabling gradient checkpointing __snake_case : Optional[int] = self.model_class(**lowerCamelCase ) # clone model model_a.load_state_dict(model.state_dict() ) model_a.to(lowerCamelCase ) model_a.enable_gradient_checkpointing() assert model_a.is_gradient_checkpointing and model_a.training __snake_case : int = model_a(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model_a.zero_grad() __snake_case : Union[str, Any] = (out_a - labels).mean() loss_a.backward() # compare the output and parameters gradients self.assertTrue((loss - loss_a).abs() < 1E-5 ) __snake_case : Optional[int] = dict(model.named_parameters() ) __snake_case : List[Any] = dict(model_a.named_parameters() ) for name, param in named_params.items(): self.assertTrue(torch_all_close(param.grad.data , named_params_a[name].grad.data , atol=5E-5 ) ) def __snake_case ( self : List[Any] ) -> Optional[int]: __snake_case , __snake_case : Optional[Any] = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" , output_loading_info=lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(lowerCamelCase ) __snake_case : Optional[Any] = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def __snake_case ( self : Optional[Any] ) -> Union[str, Any]: __snake_case : Tuple = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" ) __snake_case : Dict = model.to(lowerCamelCase ) model.eval() if torch_device == "mps": __snake_case : int = torch.manual_seed(0 ) else: __snake_case : str = torch.Generator(device=lowerCamelCase ).manual_seed(0 ) __snake_case : List[str] = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) __snake_case : Union[str, Any] = image.to(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , sample_posterior=lowerCamelCase , generator=lowerCamelCase ).sample __snake_case : List[Any] = output[0, -1, -3:, -3:].flatten().cpu() # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. if torch_device == "mps": __snake_case : Union[str, Any] = torch.tensor( [ -4.0078E-01, -3.8323E-04, -1.2681E-01, -1.1462E-01, 2.0095E-01, 1.0893E-01, -8.8247E-02, -3.0361E-01, -9.8644E-03, ] ) elif torch_device == "cpu": __snake_case : Tuple = torch.tensor( [-0.13_52, 0.08_78, 0.04_19, -0.08_18, -0.10_69, 0.06_88, -0.14_58, -0.44_46, -0.00_26] ) else: __snake_case : List[str] = torch.tensor( [-0.24_21, 0.46_42, 0.25_07, -0.04_38, 0.06_82, 0.31_60, -0.20_18, -0.07_27, 0.24_85] ) self.assertTrue(torch_all_close(lowerCamelCase , lowerCamelCase , rtol=1E-2 ) ) @slow class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : int , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] ) -> List[str]: return F'gaussian_noise_s={seed}_shape={"_".join([str(lowerCamelCase ) for s in shape] )}.npy' def __snake_case ( self : List[Any] ) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __snake_case ( self : Tuple , lowerCamelCase : List[Any]=0 , lowerCamelCase : Tuple=(4, 3, 512, 512) , lowerCamelCase : Optional[int]=False ) -> str: __snake_case : List[Any] = torch.floataa if fpaa else torch.floataa __snake_case : Tuple = torch.from_numpy(load_hf_numpy(self.get_file_format(lowerCamelCase , lowerCamelCase ) ) ).to(lowerCamelCase ).to(lowerCamelCase ) return image def __snake_case ( self : Optional[Any] , lowerCamelCase : int="CompVis/stable-diffusion-v1-4" , lowerCamelCase : int=False ) -> int: __snake_case : str = "fp16" if fpaa else None __snake_case : int = torch.floataa if fpaa else torch.floataa __snake_case : int = AutoencoderKL.from_pretrained( lowerCamelCase , subfolder="vae" , torch_dtype=lowerCamelCase , revision=lowerCamelCase , ) model.to(lowerCamelCase ).eval() return model def __snake_case ( self : str , lowerCamelCase : int=0 ) -> Optional[Any]: if torch_device == "mps": return torch.manual_seed(lowerCamelCase ) return torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) @parameterized.expand( [ # fmt: off [33, [-0.16_03, 0.98_78, -0.04_95, -0.07_90, -0.27_09, 0.83_75, -0.20_60, -0.08_24], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_76, 0.11_68, 0.13_32, -0.48_40, -0.25_08, -0.07_91, -0.04_93, -0.40_89], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[str] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] ) -> List[Any]: __snake_case : Optional[Any] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) __snake_case : Tuple = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : List[Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : int = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [33, [-0.05_13, 0.02_89, 1.37_99, 0.21_66, -0.25_73, -0.08_71, 0.51_03, -0.09_99]], [47, [-0.41_28, -0.13_20, -0.37_04, 0.19_65, -0.41_16, -0.23_32, -0.33_40, 0.22_47]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : Any , lowerCamelCase : List[str] , lowerCamelCase : List[str] ) -> Tuple: __snake_case : Any = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Any = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.16_09, 0.98_66, -0.04_87, -0.07_77, -0.27_16, 0.83_68, -0.20_55, -0.08_14], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_77, 0.11_47, 0.13_33, -0.48_41, -0.25_06, -0.08_05, -0.04_91, -0.40_85], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Any , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) with torch.no_grad(): __snake_case : int = model(lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Union[str, Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : List[str] = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [13, [-0.20_51, -0.18_03, -0.23_11, -0.21_14, -0.32_92, -0.35_74, -0.29_53, -0.33_23]], [37, [-0.26_32, -0.26_25, -0.21_99, -0.27_41, -0.45_39, -0.49_90, -0.37_20, -0.49_25]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : List[str] , lowerCamelCase : Tuple , lowerCamelCase : Any ) -> Optional[Any]: __snake_case : List[str] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : str = sample[-1, -2:, :2, -2:].flatten().cpu() __snake_case : Optional[int] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-3 ) @parameterized.expand( [ # fmt: off [27, [-0.03_69, 0.02_07, -0.07_76, -0.06_82, -0.17_47, -0.19_30, -0.14_65, -0.20_39]], [16, [-0.16_28, -0.21_34, -0.27_47, -0.26_42, -0.37_74, -0.44_04, -0.36_87, -0.42_77]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : str , lowerCamelCase : Optional[int] , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[str] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : Union[str, Any] = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Optional[Any] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=5E-3 ) @parameterized.expand([(13,), (16,), (27,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> Tuple: __snake_case : Dict = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : Any = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Any = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-1 ) @parameterized.expand([(13,), (16,), (37,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : List[Any] , lowerCamelCase : Any ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : Union[str, Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : List[Any] = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Dict = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.30_01, 0.09_18, -2.69_84, -3.97_20, -3.20_99, -5.03_53, 1.73_38, -0.20_65, 3.42_67]], [47, [-1.50_30, -4.38_71, -6.03_55, -9.11_57, -1.66_61, -2.78_53, 2.16_07, -5.08_23, 2.56_33]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Dict ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : int = self.get_sd_image(lowerCamelCase ) __snake_case : int = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model.encode(lowerCamelCase ).latent_dist __snake_case : Dict = dist.sample(generator=lowerCamelCase ) assert list(sample.shape ) == [image.shape[0], 4] + [i // 8 for i in image.shape[2:]] __snake_case : List[str] = sample[0, -1, -3:, -3:].flatten().cpu() __snake_case : Dict = torch.tensor(lowerCamelCase ) __snake_case : Dict = 3E-3 if torch_device != "mps" else 1E-2 assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=lowerCamelCase )
81
0
"""simple docstring""" # using dfs for finding eulerian path traversal def __A ( a_ :int , a_ :Dict , a_ :str , a_ :Optional[int]=None) -> List[str]: __a : Any = (path or []) + [u] for v in graph[u]: if visited_edge[u][v] is False: __a , __a : Union[str, Any] = True, True __a : List[Any] = dfs(a_ , a_ , a_ , a_) return path def __A ( a_ :int , a_ :int) -> Optional[int]: __a : Any = 0 __a : Optional[int] = -1 for i in range(a_): if i not in graph.keys(): continue if len(graph[i]) % 2 == 1: odd_degree_nodes += 1 __a : int = i if odd_degree_nodes == 0: return 1, odd_node if odd_degree_nodes == 2: return 2, odd_node return 3, odd_node def __A ( a_ :List[str] , a_ :Tuple) -> Tuple: __a : List[str] = [[False for _ in range(max_node + 1)] for _ in range(max_node + 1)] __a , __a : Any = check_circuit_or_path(a_ , a_) if check == 3: print('''graph is not Eulerian''') print('''no path''') return __a : Any = 1 if check == 2: __a : str = odd_node print('''graph has a Euler path''') if check == 1: print('''graph has a Euler cycle''') __a : Any = dfs(a_ , a_ , a_) print(a_) def __A ( ) -> List[str]: __a : List[Any] = {1: [2, 3, 4], 2: [1, 3], 3: [1, 2], 4: [1, 5], 5: [4]} __a : Any = {1: [2, 3, 4, 5], 2: [1, 3], 3: [1, 2], 4: [1, 5], 5: [1, 4]} __a : List[Any] = {1: [2, 3, 4], 2: [1, 3, 4], 3: [1, 2], 4: [1, 2, 5], 5: [4]} __a : str = {1: [2, 3], 2: [1, 3], 3: [1, 2]} __a : List[Any] = { 1: [], 2: [] # all degree is zero } __a : Tuple = 10 check_euler(a_ , a_) check_euler(a_ , a_) check_euler(a_ , a_) check_euler(a_ , a_) check_euler(a_ , a_) if __name__ == "__main__": main()
52
import warnings from ...utils import logging from .image_processing_flava import FlavaImageProcessor _snake_case : Optional[int] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[Any] , *lowerCamelCase : Any , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class FlavaFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use FlavaImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _snake_case : List[str] = logging.get_logger(__name__) _snake_case : Optional[int] = { 'facebook/convnextv2-tiny-1k-224': 'https://huggingface.co/facebook/convnextv2-tiny-1k-224/resolve/main/config.json', } class _UpperCAmelCase ( _UpperCamelCase , _UpperCamelCase ): """simple docstring""" a_ = """convnextv2""" def __init__( self : List[str] , lowerCAmelCase_ : Dict=3 , lowerCAmelCase_ : Optional[Any]=4 , lowerCAmelCase_ : str=4 , lowerCAmelCase_ : Union[str, Any]=None , lowerCAmelCase_ : List[str]=None , lowerCAmelCase_ : int="gelu" , lowerCAmelCase_ : Any=0.02 , lowerCAmelCase_ : Tuple=1e-12 , lowerCAmelCase_ : Dict=0.0 , lowerCAmelCase_ : Tuple=2_2_4 , lowerCAmelCase_ : Tuple=None , lowerCAmelCase_ : Any=None , **lowerCAmelCase_ : int , ) -> Union[str, Any]: super().__init__(**lowerCAmelCase_ ) __lowerCAmelCase = num_channels __lowerCAmelCase = patch_size __lowerCAmelCase = num_stages __lowerCAmelCase = [9_6, 1_9_2, 3_8_4, 7_6_8] if hidden_sizes is None else hidden_sizes __lowerCAmelCase = [3, 3, 9, 3] if depths is None else depths __lowerCAmelCase = hidden_act __lowerCAmelCase = initializer_range __lowerCAmelCase = layer_norm_eps __lowerCAmelCase = drop_path_rate __lowerCAmelCase = image_size __lowerCAmelCase = ['stem'] + [f"""stage{idx}""" for idx in range(1 , len(self.depths ) + 1 )] __lowerCAmelCase , __lowerCAmelCase = get_aligned_output_features_output_indices( out_features=lowerCAmelCase_ , out_indices=lowerCAmelCase_ , stage_names=self.stage_names )
53
import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : Tuple ) -> Optional[Any]: __snake_case : Dict = tempfile.mkdtemp() __snake_case : Any = SamImageProcessor() __snake_case : Optional[int] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : Optional[Any] , **lowerCamelCase : Optional[int] ) -> Optional[Any]: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[Any] ) -> Dict: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : int ) -> List[Any]: __snake_case : List[str] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : int = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : List[Any] ) -> Dict: __snake_case : int = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Dict = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Optional[Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : List[str] ) -> Tuple: __snake_case : int = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Optional[int] = self.prepare_image_inputs() __snake_case : List[str] = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : Dict = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_torch def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Tuple = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[str] = [torch.ones((1, 3, 5, 5) )] __snake_case : Tuple = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : int = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , torch.tensor(lowerCamelCase ) , torch.tensor(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : List[str] = [np.ones((1, 3, 5, 5) )] __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : str = [[1, 0], [0, 1]] with self.assertRaises(lowerCamelCase ): __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) @require_vision @require_tf class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[Any] ) -> Union[str, Any]: __snake_case : int = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : str , **lowerCamelCase : Any ) -> Tuple: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> Any: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : str ) -> List[Any]: __snake_case : int = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : Dict = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : int ) -> List[str]: __snake_case : List[Any] = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Optional[Any] = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Tuple = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : Union[str, Any] ) -> List[Any]: __snake_case : str = self.get_image_processor() __snake_case : Union[str, Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : int = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : List[str] = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_tf def __snake_case ( self : Any ) -> Optional[int]: __snake_case : List[str] = self.get_image_processor() __snake_case : Dict = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Union[str, Any] = [tf.ones((1, 3, 5, 5) )] __snake_case : List[Any] = [[1764, 2646]] __snake_case : Dict = [[683, 1024]] __snake_case : List[str] = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , tf.convert_to_tensor(lowerCamelCase ) , tf.convert_to_tensor(lowerCamelCase ) , return_tensors="tf" , ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : Union[str, Any] = [np.ones((1, 3, 5, 5) )] __snake_case : List[str] = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Tuple = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): __snake_case : Dict = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[str] ) -> str: __snake_case : Optional[int] = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : List[str] , **lowerCamelCase : Any ) -> Dict: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> List[Any]: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : str = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : List[Any] = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : str = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[Any] = np.random.randint(0 , 2 , size=(1, 3, 5, 5) ).astype(np.floataa ) __snake_case : Dict = [tf.convert_to_tensor(lowerCamelCase )] __snake_case : List[Any] = [torch.tensor(lowerCamelCase )] __snake_case : Optional[Any] = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : Union[str, Any] = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) __snake_case : Dict = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="pt" ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def __snake_case ( self : List[Any] ) -> List[str]: __snake_case : Any = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : Any = image_processor(lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Optional[Any] = processor(images=lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Tuple = image_processor(lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() __snake_case : List[Any] = processor(images=lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) )
81
0
from math import pi, sqrt, tan def a__ ( lowercase__ ): '''simple docstring''' if side_length < 0: raise ValueError("surface_area_cube() only accepts non-negative values" ) return 6 * side_length**2 def a__ ( lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' if length < 0 or breadth < 0 or height < 0: raise ValueError("surface_area_cuboid() only accepts non-negative values" ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def a__ ( lowercase__ ): '''simple docstring''' if radius < 0: raise ValueError("surface_area_sphere() only accepts non-negative values" ) return 4 * pi * radius**2 def a__ ( lowercase__ ): '''simple docstring''' if radius < 0: raise ValueError("surface_area_hemisphere() only accepts non-negative values" ) return 3 * pi * radius**2 def a__ ( lowercase__ , lowercase__ ): '''simple docstring''' if radius < 0 or height < 0: raise ValueError("surface_area_cone() only accepts non-negative values" ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def a__ ( lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( "surface_area_conical_frustum() only accepts non-negative values" ) UpperCAmelCase_ =(height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def a__ ( lowercase__ , lowercase__ ): '''simple docstring''' if radius < 0 or height < 0: raise ValueError("surface_area_cylinder() only accepts non-negative values" ) return 2 * pi * radius * (height + radius) def a__ ( lowercase__ , lowercase__ ): '''simple docstring''' if torus_radius < 0 or tube_radius < 0: raise ValueError("surface_area_torus() only accepts non-negative values" ) if torus_radius < tube_radius: raise ValueError( "surface_area_torus() does not support spindle or self intersecting tori" ) return 4 * pow(lowercase__ , 2 ) * torus_radius * tube_radius def a__ ( lowercase__ , lowercase__ ): '''simple docstring''' if length < 0 or width < 0: raise ValueError("area_rectangle() only accepts non-negative values" ) return length * width def a__ ( lowercase__ ): '''simple docstring''' if side_length < 0: raise ValueError("area_square() only accepts non-negative values" ) return side_length**2 def a__ ( lowercase__ , lowercase__ ): '''simple docstring''' if base < 0 or height < 0: raise ValueError("area_triangle() only accepts non-negative values" ) return (base * height) / 2 def a__ ( lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError("area_triangle_three_sides() only accepts non-negative values" ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError("Given three sides do not form a triangle" ) UpperCAmelCase_ =(sidea + sidea + sidea) / 2 UpperCAmelCase_ =sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def a__ ( lowercase__ , lowercase__ ): '''simple docstring''' if base < 0 or height < 0: raise ValueError("area_parallelogram() only accepts non-negative values" ) return base * height def a__ ( lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' if basea < 0 or basea < 0 or height < 0: raise ValueError("area_trapezium() only accepts non-negative values" ) return 1 / 2 * (basea + basea) * height def a__ ( lowercase__ ): '''simple docstring''' if radius < 0: raise ValueError("area_circle() only accepts non-negative values" ) return pi * radius**2 def a__ ( lowercase__ , lowercase__ ): '''simple docstring''' if radius_x < 0 or radius_y < 0: raise ValueError("area_ellipse() only accepts non-negative values" ) return pi * radius_x * radius_y def a__ ( lowercase__ , lowercase__ ): '''simple docstring''' if diagonal_a < 0 or diagonal_a < 0: raise ValueError("area_rhombus() only accepts non-negative values" ) return 1 / 2 * diagonal_a * diagonal_a def a__ ( lowercase__ , lowercase__ ): '''simple docstring''' if not isinstance(lowercase__ , lowercase__ ) or sides < 3: raise ValueError( "area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides" ) elif length < 0: raise ValueError( "area_reg_polygon() only accepts non-negative values as \ length of a side" ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print("""[DEMO] Areas of various geometric shapes: \n""") print(f"""Rectangle: {area_rectangle(10, 20) = }""") print(f"""Square: {area_square(10) = }""") print(f"""Triangle: {area_triangle(10, 10) = }""") print(f"""Triangle: {area_triangle_three_sides(5, 12, 13) = }""") print(f"""Parallelogram: {area_parallelogram(10, 20) = }""") print(f"""Rhombus: {area_rhombus(10, 20) = }""") print(f"""Trapezium: {area_trapezium(10, 20, 30) = }""") print(f"""Circle: {area_circle(20) = }""") print(f"""Ellipse: {area_ellipse(10, 20) = }""") print("""\nSurface Areas of various geometric shapes: \n""") print(f"""Cube: {surface_area_cube(20) = }""") print(f"""Cuboid: {surface_area_cuboid(10, 20, 30) = }""") print(f"""Sphere: {surface_area_sphere(20) = }""") print(f"""Hemisphere: {surface_area_hemisphere(20) = }""") print(f"""Cone: {surface_area_cone(10, 20) = }""") print(f"""Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }""") print(f"""Cylinder: {surface_area_cylinder(10, 20) = }""") print(f"""Torus: {surface_area_torus(20, 10) = }""") print(f"""Equilateral Triangle: {area_reg_polygon(3, 10) = }""") print(f"""Square: {area_reg_polygon(4, 10) = }""") print(f"""Reqular Pentagon: {area_reg_polygon(5, 10) = }""")
54
import warnings from ...utils import logging from .image_processing_deit import DeiTImageProcessor _snake_case : List[str] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[str] , *lowerCamelCase : str , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class DeiTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use DeiTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roberta import RobertaTokenizer SCREAMING_SNAKE_CASE :str = logging.get_logger(__name__) SCREAMING_SNAKE_CASE :Dict = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_file': 'tokenizer.json'} SCREAMING_SNAKE_CASE :Optional[Any] = { 'vocab_file': { 'roberta-base': 'https://huggingface.co/roberta-base/resolve/main/vocab.json', 'roberta-large': 'https://huggingface.co/roberta-large/resolve/main/vocab.json', 'roberta-large-mnli': 'https://huggingface.co/roberta-large-mnli/resolve/main/vocab.json', 'distilroberta-base': 'https://huggingface.co/distilroberta-base/resolve/main/vocab.json', 'roberta-base-openai-detector': 'https://huggingface.co/roberta-base-openai-detector/resolve/main/vocab.json', 'roberta-large-openai-detector': ( 'https://huggingface.co/roberta-large-openai-detector/resolve/main/vocab.json' ), }, 'merges_file': { 'roberta-base': 'https://huggingface.co/roberta-base/resolve/main/merges.txt', 'roberta-large': 'https://huggingface.co/roberta-large/resolve/main/merges.txt', 'roberta-large-mnli': 'https://huggingface.co/roberta-large-mnli/resolve/main/merges.txt', 'distilroberta-base': 'https://huggingface.co/distilroberta-base/resolve/main/merges.txt', 'roberta-base-openai-detector': 'https://huggingface.co/roberta-base-openai-detector/resolve/main/merges.txt', 'roberta-large-openai-detector': ( 'https://huggingface.co/roberta-large-openai-detector/resolve/main/merges.txt' ), }, 'tokenizer_file': { 'roberta-base': 'https://huggingface.co/roberta-base/resolve/main/tokenizer.json', 'roberta-large': 'https://huggingface.co/roberta-large/resolve/main/tokenizer.json', 'roberta-large-mnli': 'https://huggingface.co/roberta-large-mnli/resolve/main/tokenizer.json', 'distilroberta-base': 'https://huggingface.co/distilroberta-base/resolve/main/tokenizer.json', 'roberta-base-openai-detector': ( 'https://huggingface.co/roberta-base-openai-detector/resolve/main/tokenizer.json' ), 'roberta-large-openai-detector': ( 'https://huggingface.co/roberta-large-openai-detector/resolve/main/tokenizer.json' ), }, } SCREAMING_SNAKE_CASE :List[Any] = { 'roberta-base': 512, 'roberta-large': 512, 'roberta-large-mnli': 512, 'distilroberta-base': 512, 'roberta-base-openai-detector': 512, 'roberta-large-openai-detector': 512, } class UpperCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' snake_case_ = VOCAB_FILES_NAMES snake_case_ = PRETRAINED_VOCAB_FILES_MAP snake_case_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case_ = ["input_ids", "attention_mask"] snake_case_ = RobertaTokenizer def __init__( self : Union[str, Any] ,A : List[str]=None ,A : Union[str, Any]=None ,A : Union[str, Any]=None ,A : Any="replace" ,A : str="<s>" ,A : List[Any]="</s>" ,A : Any="</s>" ,A : Optional[int]="<s>" ,A : Union[str, Any]="<unk>" ,A : Dict="<pad>" ,A : Union[str, Any]="<mask>" ,A : str=False ,A : List[str]=True ,**A : Optional[int] ,): super().__init__( A ,A ,tokenizer_file=A ,errors=A ,bos_token=A ,eos_token=A ,sep_token=A ,cls_token=A ,unk_token=A ,pad_token=A ,mask_token=A ,add_prefix_space=A ,trim_offsets=A ,**A ,) __A = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" ,A ) != add_prefix_space: __A = getattr(A ,pre_tok_state.pop("type" ) ) __A = add_prefix_space __A = pre_tok_class(**A ) __A = add_prefix_space __A = "post_processor" __A = getattr(self.backend_tokenizer ,A ,A ) if tokenizer_component_instance: __A = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: __A = tuple(state["sep"] ) if "cls" in state: __A = tuple(state["cls"] ) __A = False if state.get("add_prefix_space" ,A ) != add_prefix_space: __A = add_prefix_space __A = True if state.get("trim_offsets" ,A ) != trim_offsets: __A = trim_offsets __A = True if changes_to_apply: __A = getattr(A ,state.pop("type" ) ) __A = component_class(**A ) setattr(self.backend_tokenizer ,A ,A ) @property def UpperCamelCase_ ( self : Tuple ): if self._mask_token is None: if self.verbose: logger.error("Using mask_token, but it is not set yet." ) return None return str(self._mask_token ) @mask_token.setter def UpperCamelCase_ ( self : Optional[int] ,A : List[Any] ): __A = AddedToken(A ,lstrip=A ,rstrip=A ) if isinstance(A ,A ) else value __A = value def UpperCamelCase_ ( self : Union[str, Any] ,*A : List[str] ,**A : Optional[int] ): __A = kwargs.get("is_split_into_words" ,A ) assert self.add_prefix_space or not is_split_into_words, ( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*A ,**A ) def UpperCamelCase_ ( self : int ,*A : str ,**A : Optional[Any] ): __A = kwargs.get("is_split_into_words" ,A ) assert self.add_prefix_space or not is_split_into_words, ( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._encode_plus(*A ,**A ) def UpperCamelCase_ ( self : int ,A : str ,A : Optional[str] = None ): __A = self._tokenizer.model.save(A ,name=A ) return tuple(A ) def UpperCamelCase_ ( self : Any ,A : Any ,A : Any=None ): __A = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def UpperCamelCase_ ( self : Optional[int] ,A : List[int] ,A : Optional[List[int]] = None ): __A = [self.sep_token_id] __A = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
55
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case : Union[str, Any] = { "configuration_owlvit": [ "OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "OwlViTConfig", "OwlViTOnnxConfig", "OwlViTTextConfig", "OwlViTVisionConfig", ], "processing_owlvit": ["OwlViTProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[Any] = ["OwlViTFeatureExtractor"] _snake_case : Optional[int] = ["OwlViTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "OwlViTModel", "OwlViTPreTrainedModel", "OwlViTTextModel", "OwlViTVisionModel", "OwlViTForObjectDetection", ] if TYPE_CHECKING: from .configuration_owlvit import ( OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, OwlViTConfig, OwlViTOnnxConfig, OwlViTTextConfig, OwlViTVisionConfig, ) from .processing_owlvit import OwlViTProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_owlvit import OwlViTFeatureExtractor from .image_processing_owlvit import OwlViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_owlvit import ( OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST, OwlViTForObjectDetection, OwlViTModel, OwlViTPreTrainedModel, OwlViTTextModel, OwlViTVisionModel, ) else: import sys _snake_case : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
0
'''simple docstring''' def _a (lowercase__ : str , lowercase__ : str ) -> Dict: """simple docstring""" assert x is not None assert y is not None __snake_case = len(lowercase__ ) __snake_case = len(lowercase__ ) # declaring the array for storing the dp values __snake_case = [[0] * (n + 1) for _ in range(m + 1 )] # noqa: E741 for i in range(1 , m + 1 ): for j in range(1 , n + 1 ): __snake_case = 1 if x[i - 1] == y[j - 1] else 0 __snake_case = max(l[i - 1][j] , l[i][j - 1] , l[i - 1][j - 1] + match ) __snake_case = '' __snake_case , __snake_case = m, n while i > 0 and j > 0: __snake_case = 1 if x[i - 1] == y[j - 1] else 0 if l[i][j] == l[i - 1][j - 1] + match: if match == 1: __snake_case = x[i - 1] + seq i -= 1 j -= 1 elif l[i][j] == l[i - 1][j]: i -= 1 else: j -= 1 return l[m][n], seq if __name__ == "__main__": _a : Dict = "AGGTAB" _a : Any = "GXTXAYB" _a : Dict = 4 _a : Optional[int] = "GTAB" _a , _a : Union[str, Any] = longest_common_subsequence(a, b) print("len =", ln, ", sub-sequence =", subseq) import doctest doctest.testmod()
56
from ....configuration_utils import PretrainedConfig from ....utils import logging _snake_case : int = logging.get_logger(__name__) _snake_case : str = { "Visual-Attention-Network/van-base": ( "https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json" ), } class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[Any] = "van" def __init__( self : Optional[int] , lowerCamelCase : Any=224 , lowerCamelCase : str=3 , lowerCamelCase : Any=[7, 3, 3, 3] , lowerCamelCase : Dict=[4, 2, 2, 2] , lowerCamelCase : List[Any]=[64, 128, 320, 512] , lowerCamelCase : str=[3, 3, 12, 3] , lowerCamelCase : Dict=[8, 8, 4, 4] , lowerCamelCase : Any="gelu" , lowerCamelCase : Optional[int]=0.02 , lowerCamelCase : Tuple=1E-6 , lowerCamelCase : Optional[int]=1E-2 , lowerCamelCase : int=0.0 , lowerCamelCase : Optional[Any]=0.0 , **lowerCamelCase : Optional[int] , ) -> int: super().__init__(**lowerCamelCase ) __snake_case : Union[str, Any] = image_size __snake_case : Any = num_channels __snake_case : Any = patch_sizes __snake_case : List[Any] = strides __snake_case : str = hidden_sizes __snake_case : Dict = depths __snake_case : Optional[int] = mlp_ratios __snake_case : Dict = hidden_act __snake_case : Union[str, Any] = initializer_range __snake_case : List[str] = layer_norm_eps __snake_case : Optional[int] = layer_scale_init_value __snake_case : List[Any] = drop_path_rate __snake_case : int = dropout_rate
81
0
from abc import ABC, abstractmethod from typing import List, Optional class _lowerCAmelCase( UpperCAmelCase_ ): """simple docstring""" def __init__( self ): # test for the above condition self.test() def _a ( self ): UpperCamelCase_: List[str] = 0 UpperCamelCase_: List[Any] = False while not completed: if counter == 1: self.reset() UpperCamelCase_: Union[str, Any] = self.advance() if not self.does_advance(_lowerCamelCase ): raise Exception( 'Custom Constraint is not defined correctly. self.does_advance(self.advance()) must be true.' ) UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_: int = self.update(_lowerCamelCase ) counter += 1 if counter > 1_0_0_0_0: raise Exception('update() does not fulfill the constraint.' ) if self.remaining() != 0: raise Exception('Custom Constraint is not defined correctly.' ) @abstractmethod def _a ( self ): raise NotImplementedError( f'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''' ) @abstractmethod def _a ( self , _lowerCamelCase ): raise NotImplementedError( f'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''' ) @abstractmethod def _a ( self , _lowerCamelCase ): raise NotImplementedError( f'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''' ) @abstractmethod def _a ( self ): raise NotImplementedError( f'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''' ) @abstractmethod def _a ( self ): raise NotImplementedError( f'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''' ) @abstractmethod def _a ( self , _lowerCamelCase=False ): raise NotImplementedError( f'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''' ) class _lowerCAmelCase( UpperCAmelCase_ ): """simple docstring""" def __init__( self , _lowerCamelCase ): super(_lowerCamelCase , self ).__init__() if not isinstance(_lowerCamelCase , _lowerCamelCase ) or len(_lowerCamelCase ) == 0: raise ValueError(f'''`token_ids` has to be a non-empty list, but is {token_ids}.''' ) if any((not isinstance(_lowerCamelCase , _lowerCamelCase ) or token_id < 0) for token_id in token_ids ): raise ValueError(f'''Each list in `token_ids` has to be a list of positive integers, but is {token_ids}.''' ) UpperCamelCase_: Optional[Any] = token_ids UpperCamelCase_: Optional[int] = len(self.token_ids ) UpperCamelCase_: Union[str, Any] = -1 # the index of the currently fulfilled step UpperCamelCase_: List[str] = False def _a ( self ): if self.completed: return None return self.token_ids[self.fulfilled_idx + 1] def _a ( self , _lowerCamelCase ): if not isinstance(_lowerCamelCase , _lowerCamelCase ): raise ValueError(f'''`token_id` has to be an `int`, but is {token_id} of type {type(_lowerCamelCase )}''' ) if self.completed: return False return token_id == self.token_ids[self.fulfilled_idx + 1] def _a ( self , _lowerCamelCase ): if not isinstance(_lowerCamelCase , _lowerCamelCase ): raise ValueError(f'''`token_id` has to be an `int`, but is {token_id} of type {type(_lowerCamelCase )}''' ) UpperCamelCase_: Dict = False UpperCamelCase_: Optional[Any] = False UpperCamelCase_: int = False if self.does_advance(_lowerCamelCase ): self.fulfilled_idx += 1 UpperCamelCase_: Tuple = True if self.fulfilled_idx == (self.seqlen - 1): UpperCamelCase_: Tuple = True UpperCamelCase_: List[Any] = completed else: # failed to make progress. UpperCamelCase_: Tuple = True self.reset() return stepped, completed, reset def _a ( self ): UpperCamelCase_: str = False UpperCamelCase_: List[Any] = 0 def _a ( self ): return self.seqlen - (self.fulfilled_idx + 1) def _a ( self , _lowerCamelCase=False ): UpperCamelCase_: Union[str, Any] = PhrasalConstraint(self.token_ids ) if stateful: UpperCamelCase_: Any = self.seqlen UpperCamelCase_: Any = self.fulfilled_idx UpperCamelCase_: List[Any] = self.completed return new_constraint class _lowerCAmelCase: """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=True ): UpperCamelCase_: Dict = max([len(_lowerCamelCase ) for one in nested_token_ids] ) UpperCamelCase_: Dict = {} for token_ids in nested_token_ids: UpperCamelCase_: str = root for tidx, token_id in enumerate(_lowerCamelCase ): if token_id not in level: UpperCamelCase_: Optional[Any] = {} UpperCamelCase_: Dict = level[token_id] if no_subsets and self.has_subsets(_lowerCamelCase , _lowerCamelCase ): raise ValueError( 'Each list in `nested_token_ids` can\'t be a complete subset of another list, but is' f''' {nested_token_ids}.''' ) UpperCamelCase_: int = root def _a ( self , _lowerCamelCase ): UpperCamelCase_: Union[str, Any] = self.trie for current_token in current_seq: UpperCamelCase_: Optional[Any] = start[current_token] UpperCamelCase_: Dict = list(start.keys() ) return next_tokens def _a ( self , _lowerCamelCase ): UpperCamelCase_: Tuple = self.next_tokens(_lowerCamelCase ) return len(_lowerCamelCase ) == 0 def _a ( self , _lowerCamelCase ): UpperCamelCase_: Union[str, Any] = list(root.values() ) if len(_lowerCamelCase ) == 0: return 1 else: return sum([self.count_leaves(_lowerCamelCase ) for nn in next_nodes] ) def _a ( self , _lowerCamelCase , _lowerCamelCase ): UpperCamelCase_: str = self.count_leaves(_lowerCamelCase ) return len(_lowerCamelCase ) != leaf_count class _lowerCAmelCase( UpperCAmelCase_ ): """simple docstring""" def __init__( self , _lowerCamelCase ): super(_lowerCamelCase , self ).__init__() if not isinstance(_lowerCamelCase , _lowerCamelCase ) or len(_lowerCamelCase ) == 0: raise ValueError(f'''`nested_token_ids` has to be a non-empty list, but is {nested_token_ids}.''' ) if any(not isinstance(_lowerCamelCase , _lowerCamelCase ) for token_ids in nested_token_ids ): raise ValueError(f'''`nested_token_ids` has to be a list of lists, but is {nested_token_ids}.''' ) if any( any((not isinstance(_lowerCamelCase , _lowerCamelCase ) or token_id < 0) for token_id in token_ids ) for token_ids in nested_token_ids ): raise ValueError( f'''Each list in `nested_token_ids` has to be a list of positive integers, but is {nested_token_ids}.''' ) UpperCamelCase_: Any = DisjunctiveTrie(_lowerCamelCase ) UpperCamelCase_: Dict = nested_token_ids UpperCamelCase_: Union[str, Any] = self.trie.max_height UpperCamelCase_: Any = [] UpperCamelCase_: Optional[Any] = False def _a ( self ): UpperCamelCase_: Optional[int] = self.trie.next_tokens(self.current_seq ) if len(_lowerCamelCase ) == 0: return None else: return token_list def _a ( self , _lowerCamelCase ): if not isinstance(_lowerCamelCase , _lowerCamelCase ): raise ValueError(f'''`token_id` is supposed to be type `int`, but is {token_id} of type {type(_lowerCamelCase )}''' ) UpperCamelCase_: Union[str, Any] = self.trie.next_tokens(self.current_seq ) return token_id in next_tokens def _a ( self , _lowerCamelCase ): if not isinstance(_lowerCamelCase , _lowerCamelCase ): raise ValueError(f'''`token_id` is supposed to be type `int`, but is {token_id} of type {type(_lowerCamelCase )}''' ) UpperCamelCase_: int = False UpperCamelCase_: Dict = False UpperCamelCase_: int = False if self.does_advance(_lowerCamelCase ): self.current_seq.append(_lowerCamelCase ) UpperCamelCase_: List[str] = True else: UpperCamelCase_: Tuple = True self.reset() UpperCamelCase_: Any = self.trie.reached_leaf(self.current_seq ) UpperCamelCase_: Union[str, Any] = completed return stepped, completed, reset def _a ( self ): UpperCamelCase_: Dict = False UpperCamelCase_: List[str] = [] def _a ( self ): if self.completed: # since this can be completed without reaching max height return 0 else: return self.seqlen - len(self.current_seq ) def _a ( self , _lowerCamelCase=False ): UpperCamelCase_: Optional[int] = DisjunctiveConstraint(self.token_ids ) if stateful: UpperCamelCase_: str = self.seqlen UpperCamelCase_: Any = self.current_seq UpperCamelCase_: Any = self.completed return new_constraint class _lowerCAmelCase: """simple docstring""" def __init__( self , _lowerCamelCase ): UpperCamelCase_: Tuple = constraints # max # of steps required to fulfill a given constraint UpperCamelCase_: List[Any] = max([c.seqlen for c in constraints] ) UpperCamelCase_: Optional[int] = len(_lowerCamelCase ) UpperCamelCase_: Optional[Any] = False self.init_state() def _a ( self ): UpperCamelCase_: str = [] UpperCamelCase_: Union[str, Any] = None UpperCamelCase_: Tuple = [constraint.copy(stateful=_lowerCamelCase ) for constraint in self.constraints] def _a ( self ): UpperCamelCase_: Tuple = 0 if self.inprogress_constraint: # extra points for having a constraint mid-fulfilled add += self.max_seqlen - self.inprogress_constraint.remaining() return (len(self.complete_constraints ) * self.max_seqlen) + add def _a ( self ): UpperCamelCase_: List[str] = [] if self.inprogress_constraint is None: for constraint in self.pending_constraints: # "pending" == "unfulfilled yet" UpperCamelCase_: Tuple = constraint.advance() if isinstance(_lowerCamelCase , _lowerCamelCase ): token_list.append(_lowerCamelCase ) elif isinstance(_lowerCamelCase , _lowerCamelCase ): token_list.extend(_lowerCamelCase ) else: UpperCamelCase_: Dict = self.inprogress_constraint.advance() if isinstance(_lowerCamelCase , _lowerCamelCase ): token_list.append(_lowerCamelCase ) elif isinstance(_lowerCamelCase , _lowerCamelCase ): token_list.extend(_lowerCamelCase ) if len(_lowerCamelCase ) == 0: return None else: return token_list def _a ( self , _lowerCamelCase ): self.init_state() if token_ids is not None: for token in token_ids: # completes or steps **one** constraint UpperCamelCase_ ,UpperCamelCase_: Any = self.add(_lowerCamelCase ) # the entire list of constraints are fulfilled if self.completed: break def _a ( self , _lowerCamelCase ): if not isinstance(_lowerCamelCase , _lowerCamelCase ): raise ValueError(f'''`token_id` should be an `int`, but is `{token_id}`.''' ) UpperCamelCase_ ,UpperCamelCase_: int = False, False if self.completed: UpperCamelCase_: List[Any] = True UpperCamelCase_: List[Any] = False return complete, stepped if self.inprogress_constraint is not None: # In the middle of fulfilling a constraint. If the `token_id` *does* makes an incremental progress to current # job, simply update the state UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_: Optional[Any] = self.inprogress_constraint.update(_lowerCamelCase ) if reset: # 1. If the next token breaks the progress, then we must restart. # e.g. constraint = "I love pies" and sequence so far is "I love" but `token_id` == "books". # But that doesn't mean we self.init_state(), since we only reset the state for this particular # constraint, not the full list of constraints. self.pending_constraints.append(self.inprogress_constraint.copy(stateful=_lowerCamelCase ) ) UpperCamelCase_: str = None if complete: # 2. If the next token completes the constraint, move it to completed list, set # inprogress to None. If there are no pending constraints either, then this full list of constraints # is complete. self.complete_constraints.append(self.inprogress_constraint ) UpperCamelCase_: List[Any] = None if len(self.pending_constraints ) == 0: # we're done! UpperCamelCase_: Any = True else: # Not in the middle of fulfilling a constraint. So does this `token_id` helps us step towards any of our list # of constraints? for cidx, pending_constraint in enumerate(self.pending_constraints ): if pending_constraint.does_advance(_lowerCamelCase ): UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_: int = pending_constraint.update(_lowerCamelCase ) if not stepped: raise Exception( '`constraint.update(token_id)` is not yielding incremental progress, ' 'even though `constraint.does_advance(token_id)` is true.' ) if complete: self.complete_constraints.append(_lowerCamelCase ) UpperCamelCase_: Union[str, Any] = None if not complete and stepped: UpperCamelCase_: List[str] = pending_constraint if complete or stepped: # If we made any progress at all, then it's at least not a "pending constraint". UpperCamelCase_: List[Any] = ( self.pending_constraints[:cidx] + self.pending_constraints[cidx + 1 :] ) if len(self.pending_constraints ) == 0 and self.inprogress_constraint is None: # If there's no longer any pending after this and no inprogress either, then we must be # complete. UpperCamelCase_: str = True break # prevent accidentally stepping through multiple constraints with just one token. return complete, stepped def _a ( self , _lowerCamelCase=True ): UpperCamelCase_: List[Any] = ConstraintListState(self.constraints ) # we actually never though self.constraints objects # throughout this process. So it's at initialization state. if stateful: UpperCamelCase_: Optional[int] = [ constraint.copy(stateful=_lowerCamelCase ) for constraint in self.complete_constraints ] if self.inprogress_constraint is not None: UpperCamelCase_: Tuple = self.inprogress_constraint.copy(stateful=_lowerCamelCase ) UpperCamelCase_: Optional[int] = [constraint.copy() for constraint in self.pending_constraints] return new_state
57
import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer _snake_case : Union[str, Any] = logging.getLogger(__name__) def lowerCAmelCase_ ( ): __snake_case : int = argparse.ArgumentParser( description="Prepare TFRecord shards from pre-tokenized samples of the wikitext dataset." ) parser.add_argument( "--dataset_name" , type=__lowerCamelCase , default="wikitext" , help="Name of the training. Explore datasets at: hf.co/datasets." , ) parser.add_argument( "--dataset_config" , type=__lowerCamelCase , default="wikitext-103-raw-v1" , help="Configuration name of the dataset." ) parser.add_argument( "--tokenizer_name_or_path" , type=__lowerCamelCase , default="sayakpaul/unigram-tokenizer-wikitext" , help="Tokenizer identifier. Can be a local filepath or a Hub identifier." , ) parser.add_argument( "--shard_size" , type=__lowerCamelCase , default=1_0_0_0 , help="Number of entries to go in a single shard." , ) parser.add_argument("--split" , type=__lowerCamelCase , default="train" , choices=["train", "test", "validation"] ) parser.add_argument( "--limit" , default=__lowerCamelCase , type=__lowerCamelCase , help="Limit the number of shards (used for debugging)." , ) parser.add_argument( "--max_length" , type=__lowerCamelCase , default=5_1_2 , help="Maximum sequence length. For training on TPUs, it helps to have a maximum" " sequence length that is a multiple of 8." , ) parser.add_argument( "--output_dir" , default="tf-tpu" , type=__lowerCamelCase , help="Output directory where the TFRecord shards will be saved. If the" " path is appended with `gs://` ('gs://tf-tpu', for example) then the TFRecord" " shards will be directly saved to a Google Cloud Storage bucket." , ) __snake_case : List[str] = parser.parse_args() return args def lowerCAmelCase_ ( __lowerCamelCase ): def fn(__lowerCamelCase ): return tokenizer(examples["text"] ) return fn def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = [] for i in range(len(tokenized_data["input_ids"] ) ): __snake_case : Tuple = { "input_ids": tf.train.Feature(intaa_list=tf.train.IntaaList(value=tokenized_data["input_ids"][i] ) ), "attention_mask": tf.train.Feature( intaa_list=tf.train.IntaaList(value=tokenized_data["attention_mask"][i] ) ), } __snake_case : List[Any] = tf.train.Features(feature=__lowerCamelCase ) __snake_case : str = tf.train.Example(features=__lowerCamelCase ) __snake_case : List[str] = example.SerializeToString() records.append(__lowerCamelCase ) return records def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Optional[int] = datasets.load_dataset(args.dataset_name , args.dataset_config , split=args.split ) if args.limit is not None: __snake_case : Optional[Any] = min(len(__lowerCamelCase ) , args.limit ) __snake_case : Dict = dataset.select(range(__lowerCamelCase ) ) print(F'Limiting the dataset to {args.limit} entries.' ) __snake_case : Dict = AutoTokenizer.from_pretrained(args.tokenizer_name_or_path ) # Handle output directory creation. # For serializing into a Google Cloud Storage Bucket, one needs to first # create a bucket. if "gs" not in args.output_dir: if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) __snake_case : Dict = os.path.join(args.output_dir , args.split ) if not os.path.exists(__lowerCamelCase ): os.makedirs(__lowerCamelCase ) else: __snake_case : str = os.path.join(args.output_dir , args.split ) # Tokenize the whole dataset at once. __snake_case : Any = tokenize_function(__lowerCamelCase ) __snake_case : Optional[Any] = dataset.map(__lowerCamelCase , batched=__lowerCamelCase , num_proc=4 , remove_columns=["text"] ) # We need to concatenate all our texts together, and then split the result # into chunks of a fixed size, which we will call block_size. To do this, we # will use the map method again, with the option batched=True. When we use batched=True, # the function we pass to map() will be passed multiple inputs at once, allowing us # to group them into more or fewer examples than we had in the input. # This allows us to create our new fixed-length samples. The advantage of this # method is that we don't lose a whole lot of content from the dataset compared to the # case where we simply tokenize with a pre-defined max_length. def group_texts(__lowerCamelCase ): # Concatenate all texts. __snake_case : List[str] = {k: sum(examples[k] , [] ) for k in examples.keys()} __snake_case : List[Any] = len(concatenated_examples[list(examples.keys() )[0]] ) # We drop the small remainder, though you could add padding instead if the model supports it # In this, as in all things, we advise you to follow your heart 🫀 __snake_case : Any = (total_length // args.max_length) * args.max_length # Split by chunks of max_len. __snake_case : int = { k: [t[i : i + args.max_length] for i in range(0 , __lowerCamelCase , args.max_length )] for k, t in concatenated_examples.items() } return result __snake_case : Any = dataset_tokenized.map(__lowerCamelCase , batched=__lowerCamelCase , batch_size=1_0_0_0 , num_proc=4 ) __snake_case : Optional[Any] = 0 __snake_case : Optional[Any] = 0 for shard in range(0 , len(__lowerCamelCase ) , args.shard_size ): __snake_case : List[str] = grouped_dataset[shard : shard + args.shard_size] __snake_case : Any = len(dataset_snapshot["input_ids"] ) __snake_case : List[Any] = os.path.join(__lowerCamelCase , F'dataset-{shard_count}-{records_containing}.tfrecord' ) __snake_case : Optional[Any] = get_serialized_examples(__lowerCamelCase ) with tf.io.TFRecordWriter(__lowerCamelCase ) as out_file: for i in range(len(__lowerCamelCase ) ): __snake_case : Union[str, Any] = serialized_examples[i] out_file.write(__lowerCamelCase ) print("Wrote file {} containing {} records".format(__lowerCamelCase , __lowerCamelCase ) ) shard_count += 1 total_records += records_containing with open(F'split-{args.split}-records-count.txt' , "w" ) as f: print(F'Total {args.split} records: {total_records}' , file=__lowerCamelCase ) if __name__ == "__main__": _snake_case : List[Any] = parse_args() main(args)
81
0
"""simple docstring""" from __future__ import annotations import csv import requests from bsa import BeautifulSoup def __lowerCAmelCase ( __UpperCamelCase : str = "" ): '''simple docstring''' snake_case_ : Optional[int] = url or """https://www.imdb.com/chart/top/?ref_=nv_mv_250""" snake_case_ : Dict = BeautifulSoup(requests.get(__UpperCamelCase ).text , """html.parser""" ) snake_case_ : str = soup.find_all("""td""" , attrs="""titleColumn""" ) snake_case_ : Union[str, Any] = soup.find_all("""td""" , class_="""ratingColumn imdbRating""" ) return { title.a.text: float(rating.strong.text ) for title, rating in zip(__UpperCamelCase , __UpperCamelCase ) } def __lowerCAmelCase ( __UpperCamelCase : str = "IMDb_Top_250_Movies.csv" ): '''simple docstring''' snake_case_ : int = get_imdb_top_aaa_movies() with open(__UpperCamelCase , """w""" , newline="""""" ) as out_file: snake_case_ : str = csv.writer(__UpperCamelCase ) writer.writerow(["""Movie title""", """IMDb rating"""] ) for title, rating in movies.items(): writer.writerow([title, rating] ) if __name__ == "__main__": write_movies()
58
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) _snake_case : List[Any] = "\\n Text data.\n Second line of data." _snake_case : Tuple = "file" @pytest.fixture(scope="session" ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / (FILE_PATH + ".zstd") __snake_case : Optional[Any] = bytes(__lowerCamelCase , "utf-8" ) with zstd.open(__lowerCamelCase , "wb" ) as f: f.write(__lowerCamelCase ) return path @pytest.fixture def lowerCAmelCase_ ( __lowerCamelCase ): with open(os.path.join(tmpfs.local_root_dir , __lowerCamelCase ) , "w" ) as f: f.write(__lowerCamelCase ) return FILE_PATH @pytest.mark.parametrize("compression_format" , ["gzip", "xz", "zstd"] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Optional[int] = {"gzip": gz_file, "xz": xz_file, "zstd": zstd_path} __snake_case : str = input_paths[compression_format] __snake_case : Optional[Any] = tmp_path / "cache" __snake_case : Optional[int] = DownloadConfig(cache_dir=__lowerCamelCase , extract_compressed_file=__lowerCamelCase ) __snake_case : Union[str, Any] = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) with open(__lowerCamelCase ) as f: __snake_case : Dict = f.read() with open(__lowerCamelCase ) as f: __snake_case : Tuple = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("default_extracted" , [True, False] ) @pytest.mark.parametrize("default_cache_dir" , [True, False] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Tuple = "custom_cache" __snake_case : List[str] = "custom_extracted_dir" __snake_case : Any = tmp_path / "custom_extracted_path" if default_extracted: __snake_case : List[Any] = ("downloads" if default_cache_dir else custom_cache_dir, "extracted") else: monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_DIR" , __lowerCamelCase ) monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_PATH" , str(__lowerCamelCase ) ) __snake_case : Optional[Any] = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) __snake_case : Optional[int] = xz_file __snake_case : Optional[int] = ( DownloadConfig(extract_compressed_file=__lowerCamelCase ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=__lowerCamelCase ) ) __snake_case : str = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) assert Path(__lowerCamelCase ).parent.parts[-2:] == expected def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : Optional[Any] = str(Path(__lowerCamelCase ).resolve() ) assert cached_path(__lowerCamelCase ) == text_file # relative path __snake_case : Any = str(Path(__lowerCamelCase ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(__lowerCamelCase ) == text_file def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : List[Any] = str(tmp_path.resolve() / "__missing_file__.txt" ) with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) # relative path __snake_case : Optional[int] = "./__missing_file__.txt" with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : str = get_from_cache(F'tmp://{tmpfs_file}' ) with open(__lowerCamelCase ) as f: __snake_case : Union[str, Any] = f.read() assert output_file_content == FILE_CONTENT @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( ): with pytest.raises(__lowerCamelCase ): cached_path("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[Any] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): http_get("https://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): http_head("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[str] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): ftp_get("ftp://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): ftp_head("ftp://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): fsspec_get("s3://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): fsspec_head("s3://huggingface.co" )
81
0
import random from typing import Any def lowerCAmelCase_ ( __a ) -> list[Any]: """simple docstring""" for _ in range(len(__a ) ): lowerCamelCase__: Tuple =random.randint(0 , len(__a ) - 1 ) lowerCamelCase__: str =random.randint(0 , len(__a ) - 1 ) lowerCamelCase__ , lowerCamelCase__: Optional[int] =data[b], data[a] return data if __name__ == "__main__": __A = [0, 1, 2, 3, 4, 5, 6, 7] __A = ["python", "says", "hello", "!"] print("Fisher-Yates Shuffle:") print("List", integers, strings) print("FY Shuffle", fisher_yates_shuffle(integers), fisher_yates_shuffle(strings))
59
_snake_case : Optional[int] = {"a": ["c", "b"], "b": ["d", "e"], "c": [], "d": [], "e": []} _snake_case : Dict = ["a", "b", "c", "d", "e"] def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : List[str] = start # add current to visited visited.append(__lowerCamelCase ) __snake_case : List[Any] = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: __snake_case : Tuple = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # if all neighbors visited add current to sort sort.append(__lowerCamelCase ) # if all vertices haven't been visited select a new one to visit if len(__lowerCamelCase ) != len(__lowerCamelCase ): for vertice in vertices: if vertice not in visited: __snake_case : int = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # return sort return sort if __name__ == "__main__": _snake_case : List[Any] = topological_sort("a", [], []) print(sort)
81
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available lowerCAmelCase_ = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = ['''MLukeTokenizer'''] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mluke import MLukeTokenizer else: import sys lowerCAmelCase_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
60
def lowerCAmelCase_ ( ): return [ a * b * (1_0_0_0 - a - b) for a in range(1 , 9_9_9 ) for b in range(__lowerCamelCase , 9_9_9 ) if (a * a + b * b == (1_0_0_0 - a - b) ** 2) ][0] if __name__ == "__main__": print(f'''{solution() = }''')
81
0
import inspect import unittest from transformers import YolosConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import YolosForObjectDetection, YolosModel from transformers.models.yolos.modeling_yolos import YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __lowerCamelCase : """simple docstring""" def __init__( self : Any , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : List[str]=13 , SCREAMING_SNAKE_CASE__ : Any=[30, 30] , SCREAMING_SNAKE_CASE__ : Union[str, Any]=2 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=3 , SCREAMING_SNAKE_CASE__ : int=True , SCREAMING_SNAKE_CASE__ : Optional[int]=True , SCREAMING_SNAKE_CASE__ : List[Any]=32 , SCREAMING_SNAKE_CASE__ : Optional[int]=5 , SCREAMING_SNAKE_CASE__ : Tuple=4 , SCREAMING_SNAKE_CASE__ : Tuple=37 , SCREAMING_SNAKE_CASE__ : List[Any]="gelu" , SCREAMING_SNAKE_CASE__ : Optional[Any]=0.1 , SCREAMING_SNAKE_CASE__ : str=0.1 , SCREAMING_SNAKE_CASE__ : List[str]=10 , SCREAMING_SNAKE_CASE__ : Tuple=0.02 , SCREAMING_SNAKE_CASE__ : Dict=3 , SCREAMING_SNAKE_CASE__ : str=None , SCREAMING_SNAKE_CASE__ : List[Any]=8 , SCREAMING_SNAKE_CASE__ : Any=10 , ) -> Tuple: lowerCAmelCase__ = parent lowerCAmelCase__ = batch_size lowerCAmelCase__ = image_size lowerCAmelCase__ = patch_size lowerCAmelCase__ = num_channels lowerCAmelCase__ = is_training lowerCAmelCase__ = use_labels lowerCAmelCase__ = hidden_size lowerCAmelCase__ = num_hidden_layers lowerCAmelCase__ = num_attention_heads lowerCAmelCase__ = intermediate_size lowerCAmelCase__ = hidden_act lowerCAmelCase__ = hidden_dropout_prob lowerCAmelCase__ = attention_probs_dropout_prob lowerCAmelCase__ = type_sequence_label_size lowerCAmelCase__ = initializer_range lowerCAmelCase__ = num_labels lowerCAmelCase__ = scope lowerCAmelCase__ = n_targets lowerCAmelCase__ = num_detection_tokens # we set the expected sequence length (which is used in several tests) # expected sequence length = num_patches + 1 (we add 1 for the [CLS] token) + num_detection_tokens lowerCAmelCase__ = (image_size[1] // patch_size) * (image_size[0] // patch_size) lowerCAmelCase__ = num_patches + 1 + self.num_detection_tokens def a ( self : Dict ) -> Tuple: lowerCAmelCase__ = floats_tensor([self.batch_size, self.num_channels, self.image_size[0], self.image_size[1]] ) lowerCAmelCase__ = None if self.use_labels: # labels is a list of Dict (each Dict being the labels for a given example in the batch) lowerCAmelCase__ = [] for i in range(self.batch_size ): lowerCAmelCase__ = {} lowerCAmelCase__ = torch.randint( high=self.num_labels , size=(self.n_targets,) , device=SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = torch.rand(self.n_targets , 4 , device=SCREAMING_SNAKE_CASE__ ) labels.append(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = self.get_config() return config, pixel_values, labels def a ( self : List[str] ) -> List[str]: return YolosConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=SCREAMING_SNAKE_CASE__ , initializer_range=self.initializer_range , num_detection_tokens=self.num_detection_tokens , num_labels=self.num_labels , ) def a ( self : int , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Dict ) -> Optional[int]: lowerCAmelCase__ = YolosModel(config=SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.expected_seq_len, self.hidden_size) ) def a ( self : List[Any] , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : str ) -> Union[str, Any]: lowerCAmelCase__ = YolosForObjectDetection(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() lowerCAmelCase__ = model(pixel_values=SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_detection_tokens, self.num_labels + 1) ) self.parent.assertEqual(result.pred_boxes.shape , (self.batch_size, self.num_detection_tokens, 4) ) lowerCAmelCase__ = model(pixel_values=SCREAMING_SNAKE_CASE__ , labels=SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_detection_tokens, self.num_labels + 1) ) self.parent.assertEqual(result.pred_boxes.shape , (self.batch_size, self.num_detection_tokens, 4) ) def a ( self : Optional[Any] ) -> Any: lowerCAmelCase__ = self.prepare_config_and_inputs() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = config_and_inputs lowerCAmelCase__ = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , unittest.TestCase ): """simple docstring""" snake_case__ = (YolosModel, YolosForObjectDetection) if is_torch_available() else () snake_case__ = ( {"feature-extraction": YolosModel, "object-detection": YolosForObjectDetection} if is_torch_available() else {} ) snake_case__ = False snake_case__ = False snake_case__ = False snake_case__ = False def a ( self : List[str] , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : Dict=False ) -> Dict: lowerCAmelCase__ = super()._prepare_for_class(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , return_labels=SCREAMING_SNAKE_CASE__ ) if return_labels: if model_class.__name__ == "YolosForObjectDetection": lowerCAmelCase__ = [] for i in range(self.model_tester.batch_size ): lowerCAmelCase__ = {} lowerCAmelCase__ = torch.ones( size=(self.model_tester.n_targets,) , device=SCREAMING_SNAKE_CASE__ , dtype=torch.long ) lowerCAmelCase__ = torch.ones( self.model_tester.n_targets , 4 , device=SCREAMING_SNAKE_CASE__ , dtype=torch.float ) labels.append(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = labels return inputs_dict def a ( self : Tuple ) -> Dict: lowerCAmelCase__ = YolosModelTester(self ) lowerCAmelCase__ = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE__ , has_text_modality=SCREAMING_SNAKE_CASE__ , hidden_size=37 ) def a ( self : List[str] ) -> Tuple: self.config_tester.run_common_tests() def a ( self : int ) -> List[str]: # YOLOS does not use inputs_embeds pass def a ( self : Optional[Any] ) -> Any: lowerCAmelCase__ , lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ = model_class(SCREAMING_SNAKE_CASE__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCAmelCase__ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(SCREAMING_SNAKE_CASE__ , nn.Linear ) ) def a ( self : List[str] ) -> Optional[Any]: lowerCAmelCase__ , lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ = model_class(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ = [*signature.parameters.keys()] lowerCAmelCase__ = ["pixel_values"] self.assertListEqual(arg_names[:1] , SCREAMING_SNAKE_CASE__ ) def a ( self : Union[str, Any] ) -> int: lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE__ ) def a ( self : str ) -> Dict: lowerCAmelCase__ , lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ = True # in YOLOS, the seq_len is different lowerCAmelCase__ = self.model_tester.expected_seq_len for model_class in self.all_model_classes: lowerCAmelCase__ = True lowerCAmelCase__ = False lowerCAmelCase__ = True lowerCAmelCase__ = model_class(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() with torch.no_grad(): lowerCAmelCase__ = model(**self._prepare_for_class(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ) lowerCAmelCase__ = outputs.attentions self.assertEqual(len(SCREAMING_SNAKE_CASE__ ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] lowerCAmelCase__ = True lowerCAmelCase__ = model_class(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() with torch.no_grad(): lowerCAmelCase__ = model(**self._prepare_for_class(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ) lowerCAmelCase__ = outputs.attentions self.assertEqual(len(SCREAMING_SNAKE_CASE__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) lowerCAmelCase__ = len(SCREAMING_SNAKE_CASE__ ) # Check attention is always last and order is fine lowerCAmelCase__ = True lowerCAmelCase__ = True lowerCAmelCase__ = model_class(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() with torch.no_grad(): lowerCAmelCase__ = model(**self._prepare_for_class(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ) lowerCAmelCase__ = 1 self.assertEqual(out_len + added_hidden_states , len(SCREAMING_SNAKE_CASE__ ) ) lowerCAmelCase__ = outputs.attentions self.assertEqual(len(SCREAMING_SNAKE_CASE__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) def a ( self : Optional[Any] ) -> Any: def check_hidden_states_output(SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : int ): lowerCAmelCase__ = model_class(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() with torch.no_grad(): lowerCAmelCase__ = model(**self._prepare_for_class(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ) lowerCAmelCase__ = outputs.hidden_states lowerCAmelCase__ = getattr( self.model_tester , "expected_num_hidden_layers" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(SCREAMING_SNAKE_CASE__ ) , SCREAMING_SNAKE_CASE__ ) # YOLOS has a different seq_length lowerCAmelCase__ = self.model_tester.expected_seq_len self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) lowerCAmelCase__ , lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ = True check_hidden_states_output(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ = True check_hidden_states_output(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def a ( self : str ) -> Union[str, Any]: lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_object_detection(*SCREAMING_SNAKE_CASE__ ) @slow def a ( self : Union[str, Any] ) -> Optional[Any]: for model_name in YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ = YolosModel.from_pretrained(SCREAMING_SNAKE_CASE__ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE__ ) def _A ( ): """simple docstring""" lowerCAmelCase__ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class __lowerCamelCase ( unittest.TestCase ): """simple docstring""" @cached_property def a ( self : Optional[Any] ) -> List[str]: return AutoImageProcessor.from_pretrained("hustvl/yolos-small" ) if is_vision_available() else None @slow def a ( self : Dict ) -> int: lowerCAmelCase__ = YolosForObjectDetection.from_pretrained("hustvl/yolos-small" ).to(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = self.default_image_processor lowerCAmelCase__ = prepare_img() lowerCAmelCase__ = image_processor(images=SCREAMING_SNAKE_CASE__ , return_tensors="pt" ).to(SCREAMING_SNAKE_CASE__ ) # forward pass with torch.no_grad(): lowerCAmelCase__ = model(inputs.pixel_values ) # verify outputs lowerCAmelCase__ = torch.Size((1, 100, 92) ) self.assertEqual(outputs.logits.shape , SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = torch.tensor( [[-24.0_248, -10.3_024, -14.8_290], [-42.0_392, -16.8_200, -27.4_334], [-27.2_743, -11.8_154, -18.7_148]] , device=SCREAMING_SNAKE_CASE__ , ) lowerCAmelCase__ = torch.tensor( [[0.2_559, 0.5_455, 0.4_706], [0.2_989, 0.7_279, 0.1_875], [0.7_732, 0.4_017, 0.4_462]] , device=SCREAMING_SNAKE_CASE__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , SCREAMING_SNAKE_CASE__ , atol=1e-4 ) ) self.assertTrue(torch.allclose(outputs.pred_boxes[0, :3, :3] , SCREAMING_SNAKE_CASE__ , atol=1e-4 ) ) # verify postprocessing lowerCAmelCase__ = image_processor.post_process_object_detection( SCREAMING_SNAKE_CASE__ , threshold=0.3 , target_sizes=[image.size[::-1]] )[0] lowerCAmelCase__ = torch.tensor([0.9_994, 0.9_790, 0.9_964, 0.9_972, 0.9_861] ).to(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = [75, 75, 17, 63, 17] lowerCAmelCase__ = torch.tensor([335.0_609, 79.3_848, 375.4_216, 187.2_495] ).to(SCREAMING_SNAKE_CASE__ ) self.assertEqual(len(results["scores"] ) , 5 ) self.assertTrue(torch.allclose(results["scores"] , SCREAMING_SNAKE_CASE__ , atol=1e-4 ) ) self.assertSequenceEqual(results["labels"].tolist() , SCREAMING_SNAKE_CASE__ ) self.assertTrue(torch.allclose(results["boxes"][0, :] , SCREAMING_SNAKE_CASE__ ) )
61
from math import sqrt def lowerCAmelCase_ ( __lowerCamelCase = 1_0_0_0_0_0_0 ): __snake_case : int = 0 __snake_case : int = 0 __snake_case : int while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(__lowerCamelCase , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(f'''{solution() = }''')
81
0
import copy import os from typing import TYPE_CHECKING, List, Union if TYPE_CHECKING: pass from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case = logging.get_logger(__name__) snake_case = { """kakaobrain/align-base""": """https://huggingface.co/kakaobrain/align-base/resolve/main/config.json""", } class SCREAMING_SNAKE_CASE ( lowerCAmelCase ): '''simple docstring''' UpperCamelCase_ : int = '''align_text_model''' def __init__( self : Dict , UpperCAmelCase_ : List[str]=3_0522 , UpperCAmelCase_ : Dict=768 , UpperCAmelCase_ : str=12 , UpperCAmelCase_ : Optional[Any]=12 , UpperCAmelCase_ : List[str]=3072 , UpperCAmelCase_ : Dict="gelu" , UpperCAmelCase_ : int=0.1 , UpperCAmelCase_ : Any=0.1 , UpperCAmelCase_ : Dict=512 , UpperCAmelCase_ : str=2 , UpperCAmelCase_ : Dict=0.02 , UpperCAmelCase_ : Optional[int]=1E-12 , UpperCAmelCase_ : List[str]=0 , UpperCAmelCase_ : List[Any]="absolute" , UpperCAmelCase_ : List[Any]=True , **UpperCAmelCase_ : Optional[int] , ): super().__init__(**UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : str = vocab_size SCREAMING_SNAKE_CASE : int = hidden_size SCREAMING_SNAKE_CASE : Dict = num_hidden_layers SCREAMING_SNAKE_CASE : List[str] = num_attention_heads SCREAMING_SNAKE_CASE : str = hidden_act SCREAMING_SNAKE_CASE : Dict = intermediate_size SCREAMING_SNAKE_CASE : List[Any] = hidden_dropout_prob SCREAMING_SNAKE_CASE : str = attention_probs_dropout_prob SCREAMING_SNAKE_CASE : str = max_position_embeddings SCREAMING_SNAKE_CASE : Dict = type_vocab_size SCREAMING_SNAKE_CASE : Optional[int] = initializer_range SCREAMING_SNAKE_CASE : List[Any] = layer_norm_eps SCREAMING_SNAKE_CASE : str = position_embedding_type SCREAMING_SNAKE_CASE : Optional[int] = use_cache SCREAMING_SNAKE_CASE : List[str] = pad_token_id @classmethod def _A ( cls : Any , UpperCAmelCase_ : Union[str, os.PathLike] , **UpperCAmelCase_ : Union[str, Any] ): cls._set_token_in_kwargs(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : List[str] = cls.get_config_dict(UpperCAmelCase_ , **UpperCAmelCase_ ) # get the text config dict if we are loading from AlignConfig if config_dict.get("model_type" ) == "align": SCREAMING_SNAKE_CASE : Optional[int] = config_dict["text_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(UpperCAmelCase_ , **UpperCAmelCase_ ) class SCREAMING_SNAKE_CASE ( lowerCAmelCase ): '''simple docstring''' UpperCamelCase_ : int = '''align_vision_model''' def __init__( self : Union[str, Any] , UpperCAmelCase_ : int = 3 , UpperCAmelCase_ : int = 600 , UpperCAmelCase_ : float = 2.0 , UpperCAmelCase_ : float = 3.1 , UpperCAmelCase_ : int = 8 , UpperCAmelCase_ : List[int] = [3, 3, 5, 3, 5, 5, 3] , UpperCAmelCase_ : List[int] = [32, 16, 24, 40, 80, 112, 192] , UpperCAmelCase_ : List[int] = [16, 24, 40, 80, 112, 192, 320] , UpperCAmelCase_ : List[int] = [] , UpperCAmelCase_ : List[int] = [1, 2, 2, 2, 1, 2, 1] , UpperCAmelCase_ : List[int] = [1, 2, 2, 3, 3, 4, 1] , UpperCAmelCase_ : List[int] = [1, 6, 6, 6, 6, 6, 6] , UpperCAmelCase_ : float = 0.25 , UpperCAmelCase_ : str = "swish" , UpperCAmelCase_ : int = 2560 , UpperCAmelCase_ : str = "mean" , UpperCAmelCase_ : float = 0.02 , UpperCAmelCase_ : float = 0.001 , UpperCAmelCase_ : float = 0.99 , UpperCAmelCase_ : float = 0.2 , **UpperCAmelCase_ : List[str] , ): super().__init__(**UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : Any = num_channels SCREAMING_SNAKE_CASE : Dict = image_size SCREAMING_SNAKE_CASE : List[Any] = width_coefficient SCREAMING_SNAKE_CASE : Union[str, Any] = depth_coefficient SCREAMING_SNAKE_CASE : List[Any] = depth_divisor SCREAMING_SNAKE_CASE : Tuple = kernel_sizes SCREAMING_SNAKE_CASE : int = in_channels SCREAMING_SNAKE_CASE : Union[str, Any] = out_channels SCREAMING_SNAKE_CASE : str = depthwise_padding SCREAMING_SNAKE_CASE : Dict = strides SCREAMING_SNAKE_CASE : List[str] = num_block_repeats SCREAMING_SNAKE_CASE : List[str] = expand_ratios SCREAMING_SNAKE_CASE : Union[str, Any] = squeeze_expansion_ratio SCREAMING_SNAKE_CASE : List[str] = hidden_act SCREAMING_SNAKE_CASE : Any = hidden_dim SCREAMING_SNAKE_CASE : Dict = pooling_type SCREAMING_SNAKE_CASE : int = initializer_range SCREAMING_SNAKE_CASE : Tuple = batch_norm_eps SCREAMING_SNAKE_CASE : int = batch_norm_momentum SCREAMING_SNAKE_CASE : Any = drop_connect_rate SCREAMING_SNAKE_CASE : Optional[Any] = sum(UpperCAmelCase_ ) * 4 @classmethod def _A ( cls : int , UpperCAmelCase_ : Union[str, os.PathLike] , **UpperCAmelCase_ : Union[str, Any] ): cls._set_token_in_kwargs(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : str = cls.get_config_dict(UpperCAmelCase_ , **UpperCAmelCase_ ) # get the vision config dict if we are loading from AlignConfig if config_dict.get("model_type" ) == "align": SCREAMING_SNAKE_CASE : List[Any] = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(UpperCAmelCase_ , **UpperCAmelCase_ ) class SCREAMING_SNAKE_CASE ( lowerCAmelCase ): '''simple docstring''' UpperCamelCase_ : Dict = '''align''' UpperCamelCase_ : Optional[int] = True def __init__( self : str , UpperCAmelCase_ : Optional[Any]=None , UpperCAmelCase_ : str=None , UpperCAmelCase_ : Optional[int]=640 , UpperCAmelCase_ : List[Any]=1.0 , UpperCAmelCase_ : str=0.02 , **UpperCAmelCase_ : Optional[Any] , ): super().__init__(**UpperCAmelCase_ ) if text_config is None: SCREAMING_SNAKE_CASE : Union[str, Any] = {} logger.info("text_config is None. Initializing the AlignTextConfig with default values." ) if vision_config is None: SCREAMING_SNAKE_CASE : Tuple = {} logger.info("vision_config is None. Initializing the AlignVisionConfig with default values." ) SCREAMING_SNAKE_CASE : List[str] = AlignTextConfig(**UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : Union[str, Any] = AlignVisionConfig(**UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : List[str] = projection_dim SCREAMING_SNAKE_CASE : Tuple = temperature_init_value SCREAMING_SNAKE_CASE : str = initializer_range @classmethod def _A ( cls : Any , UpperCAmelCase_ : AlignTextConfig , UpperCAmelCase_ : AlignVisionConfig , **UpperCAmelCase_ : List[str] ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **UpperCAmelCase_ ) def _A ( self : List[str] ): SCREAMING_SNAKE_CASE : Dict = copy.deepcopy(self.__dict__ ) SCREAMING_SNAKE_CASE : List[str] = self.text_config.to_dict() SCREAMING_SNAKE_CASE : List[Any] = self.vision_config.to_dict() SCREAMING_SNAKE_CASE : str = self.__class__.model_type return output
62
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class a (_lowerCAmelCase ): """simple docstring""" def __snake_case ( self : str ) -> str: __snake_case : Tuple = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCamelCase , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "neck_hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "num_attention_heads" ) ) class a : """simple docstring""" def __init__( self : Optional[int] , lowerCamelCase : List[str] , lowerCamelCase : Tuple=13 , lowerCamelCase : str=32 , lowerCamelCase : Dict=2 , lowerCamelCase : List[str]=3 , lowerCamelCase : Any=640 , lowerCamelCase : Optional[Any]=4 , lowerCamelCase : Tuple="silu" , lowerCamelCase : int=3 , lowerCamelCase : Dict=32 , lowerCamelCase : str=0.1 , lowerCamelCase : Optional[int]=0.1 , lowerCamelCase : Optional[Any]=0.1 , lowerCamelCase : Dict=0.02 , lowerCamelCase : Union[str, Any]=True , lowerCamelCase : Optional[int]=True , lowerCamelCase : Union[str, Any]=10 , lowerCamelCase : int=None , ) -> str: __snake_case : Optional[Any] = parent __snake_case : Optional[Any] = batch_size __snake_case : Any = image_size __snake_case : List[Any] = patch_size __snake_case : Any = num_channels __snake_case : Union[str, Any] = last_hidden_size __snake_case : Any = num_attention_heads __snake_case : Any = hidden_act __snake_case : Tuple = conv_kernel_size __snake_case : Any = output_stride __snake_case : Any = hidden_dropout_prob __snake_case : List[Any] = attention_probs_dropout_prob __snake_case : Optional[Any] = classifier_dropout_prob __snake_case : Union[str, Any] = use_labels __snake_case : Optional[int] = is_training __snake_case : Dict = num_labels __snake_case : Any = initializer_range __snake_case : Optional[int] = scope def __snake_case ( self : str ) -> Union[str, Any]: __snake_case : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : List[Any] = None __snake_case : Optional[int] = None if self.use_labels: __snake_case : Optional[int] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : Any = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __snake_case : Optional[Any] = self.get_config() return config, pixel_values, labels, pixel_labels def __snake_case ( self : Any ) -> Union[str, Any]: return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __snake_case ( self : Any , lowerCamelCase : Tuple , lowerCamelCase : int , lowerCamelCase : Tuple , lowerCamelCase : Optional[int] ) -> Dict: __snake_case : List[Any] = MobileViTModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[str] = model(lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[Any] , lowerCamelCase : List[str] , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] , lowerCamelCase : Tuple ) -> List[str]: __snake_case : str = self.num_labels __snake_case : List[Any] = MobileViTForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : Dict , lowerCamelCase : Any , lowerCamelCase : Any , lowerCamelCase : Optional[Any] , lowerCamelCase : Dict ) -> Dict: __snake_case : Union[str, Any] = self.num_labels __snake_case : Optional[int] = MobileViTForSemanticSegmentation(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Tuple = model(lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[int] ) -> List[Any]: __snake_case : Optional[Any] = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case , __snake_case : Union[str, Any] = config_and_inputs __snake_case : Dict = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) __UpperCAmelCase : Optional[Any] = ( { "feature-extraction": MobileViTModel, "image-classification": MobileViTForImageClassification, "image-segmentation": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) __UpperCAmelCase : List[str] = False __UpperCAmelCase : int = False __UpperCAmelCase : Optional[int] = False __UpperCAmelCase : Optional[int] = False def __snake_case ( self : Optional[int] ) -> Dict: __snake_case : Tuple = MobileViTModelTester(self ) __snake_case : Any = MobileViTConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase ) def __snake_case ( self : Optional[int] ) -> Dict: self.config_tester.run_common_tests() @unittest.skip(reason="MobileViT does not use inputs_embeds" ) def __snake_case ( self : Dict ) -> Any: pass @unittest.skip(reason="MobileViT does not support input and output embeddings" ) def __snake_case ( self : Dict ) -> List[Any]: pass @unittest.skip(reason="MobileViT does not output attentions" ) def __snake_case ( self : int ) -> Dict: pass def __snake_case ( self : int ) -> Union[str, Any]: __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Optional[Any] = model_class(lowerCamelCase ) __snake_case : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : int = [*signature.parameters.keys()] __snake_case : List[str] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __snake_case ( self : int ) -> Tuple: pass def __snake_case ( self : Any ) -> Tuple: __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def __snake_case ( self : Any ) -> str: def check_hidden_states_output(lowerCamelCase : Union[str, Any] , lowerCamelCase : Dict , lowerCamelCase : Any ): __snake_case : int = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : int = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : Union[str, Any] = outputs.hidden_states __snake_case : int = 5 self.assertEqual(len(lowerCamelCase ) , lowerCamelCase ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. __snake_case : List[Any] = 2 for i in range(len(lowerCamelCase ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) __snake_case , __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : str = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : List[Any] = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Any ) -> Any: __snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) def __snake_case ( self : List[str] ) -> List[str]: __snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowerCamelCase ) @slow def __snake_case ( self : List[str] ) -> Any: for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : List[str] = MobileViTModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def lowerCAmelCase_ ( ): __snake_case : Optional[int] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a (unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self : str ) -> Dict: return MobileViTImageProcessor.from_pretrained("apple/mobilevit-xx-small" ) if is_vision_available() else None @slow def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : Tuple = MobileViTForImageClassification.from_pretrained("apple/mobilevit-xx-small" ).to(lowerCamelCase ) __snake_case : Optional[Any] = self.default_image_processor __snake_case : Union[str, Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Dict = model(**lowerCamelCase ) # verify the logits __snake_case : Union[str, Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([-1.93_64, -1.23_27, -0.46_53] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : str ) -> Optional[int]: __snake_case : Optional[int] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : str = model.to(lowerCamelCase ) __snake_case : int = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Optional[int] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : List[str] = model(**lowerCamelCase ) __snake_case : Union[str, Any] = outputs.logits # verify the logits __snake_case : Tuple = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , lowerCamelCase ) __snake_case : Union[str, Any] = torch.tensor( [ [[6.97_13, 6.97_86, 7.24_22], [7.28_93, 7.28_25, 7.44_46], [7.65_80, 7.87_97, 7.94_20]], [[-10.68_69, -10.32_50, -10.34_71], [-10.42_28, -9.98_68, -9.71_32], [-11.04_05, -11.02_21, -10.73_18]], [[-3.30_89, -2.85_39, -2.67_40], [-3.27_06, -2.56_21, -2.51_08], [-3.25_34, -2.66_15, -2.66_51]], ] , device=lowerCamelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : Union[str, Any] ) -> Optional[int]: __snake_case : Optional[Any] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Tuple = model.to(lowerCamelCase ) __snake_case : Dict = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : List[Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Any = model(**lowerCamelCase ) __snake_case : Dict = outputs.logits.detach().cpu() __snake_case : Any = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase , target_sizes=[(50, 60)] ) __snake_case : int = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , lowerCamelCase ) __snake_case : List[str] = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase ) __snake_case : Optional[int] = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , lowerCamelCase )
81
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) a : List[str] = {"configuration_xglm": ["XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP", "XGLMConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Dict = ["XGLMTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[Any] = ["XGLMTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[Any] = [ "XGLM_PRETRAINED_MODEL_ARCHIVE_LIST", "XGLMForCausalLM", "XGLMModel", "XGLMPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = [ "FlaxXGLMForCausalLM", "FlaxXGLMModel", "FlaxXGLMPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = [ "TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXGLMForCausalLM", "TFXGLMModel", "TFXGLMPreTrainedModel", ] if TYPE_CHECKING: from .configuration_xglm import XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XGLMConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm import XGLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm_fast import XGLMTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xglm import XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, XGLMForCausalLM, XGLMModel, XGLMPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xglm import FlaxXGLMForCausalLM, FlaxXGLMModel, FlaxXGLMPreTrainedModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, TFXGLMPreTrainedModel, ) else: import sys a : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure)
63
import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor _snake_case : int = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[int] , *lowerCamelCase : List[Any] , **lowerCamelCase : int ) -> None: warnings.warn( "The class ImageGPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use ImageGPTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
def A__ ( snake_case_ : list ): SCREAMING_SNAKE_CASE__: Dict= 0 while len(snake_case_ ) > 1: SCREAMING_SNAKE_CASE__: Any= 0 # Consider two files with minimum cost to be merged for _ in range(2 ): SCREAMING_SNAKE_CASE__: Optional[int]= files.index(min(snake_case_ ) ) temp += files[min_index] files.pop(snake_case_ ) files.append(snake_case_ ) optimal_merge_cost += temp return optimal_merge_cost if __name__ == "__main__": import doctest doctest.testmod()
64
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _snake_case : Optional[Any] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[str] = ["pixel_values"] def __init__( self : List[Any] , lowerCamelCase : bool = True , lowerCamelCase : Union[int, float] = 1 / 255 , lowerCamelCase : bool = True , lowerCamelCase : int = 8 , **lowerCamelCase : Tuple , ) -> None: super().__init__(**lowerCamelCase ) __snake_case : Dict = do_rescale __snake_case : Dict = rescale_factor __snake_case : Optional[Any] = do_pad __snake_case : Tuple = pad_size def __snake_case ( self : Dict , lowerCamelCase : np.ndarray , lowerCamelCase : float , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase : Optional[int] ) -> np.ndarray: return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def __snake_case ( self : Optional[Any] , lowerCamelCase : np.ndarray , lowerCamelCase : int , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None ) -> Tuple: __snake_case , __snake_case : List[str] = get_image_size(lowerCamelCase ) __snake_case : Optional[Any] = (old_height // size + 1) * size - old_height __snake_case : List[Any] = (old_width // size + 1) * size - old_width return pad(lowerCamelCase , ((0, pad_height), (0, pad_width)) , mode="symmetric" , data_format=lowerCamelCase ) def __snake_case ( self : Tuple , lowerCamelCase : ImageInput , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[float] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[int] = None , lowerCamelCase : Optional[Union[str, TensorType]] = None , lowerCamelCase : Union[str, ChannelDimension] = ChannelDimension.FIRST , **lowerCamelCase : Union[str, Any] , ) -> List[str]: __snake_case : int = do_rescale if do_rescale is not None else self.do_rescale __snake_case : Dict = rescale_factor if rescale_factor is not None else self.rescale_factor __snake_case : str = do_pad if do_pad is not None else self.do_pad __snake_case : Any = pad_size if pad_size is not None else self.pad_size __snake_case : int = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) # All transformations expect numpy arrays. __snake_case : str = [to_numpy_array(lowerCamelCase ) for image in images] if do_rescale: __snake_case : Optional[int] = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_pad: __snake_case : Optional[Any] = [self.pad(lowerCamelCase , size=lowerCamelCase ) for image in images] __snake_case : int = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __snake_case : Union[str, Any] = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
81
0
"""simple docstring""" import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import BatchEncoding, PreTrainedTokenizer from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = '▁' __UpperCAmelCase = { 'vocab_file': 'vocab.json', 'spm_file': 'sentencepiece.bpe.model', 'tokenizer_config_file': 'tokenizer_config.json', } __UpperCAmelCase = { 'vocab_file': { 'facebook/m2m100_418M': 'https://huggingface.co/facebook/m2m100_418M/resolve/main/vocab.json', 'facebook/m2m100_1.2B': 'https://huggingface.co/facebook/m2m100_1.2B/resolve/main/vocab.json', }, 'spm_file': { 'facebook/m2m100_418M': 'https://huggingface.co/facebook/m2m100_418M/resolve/main/sentencepiece.bpe.model', 'facebook/m2m100_1.2B': 'https://huggingface.co/facebook/m2m100_1.2B/resolve/main/sentencepiece.bpe.model', }, 'tokenizer_config_file': { 'facebook/m2m100_418M': 'https://huggingface.co/facebook/m2m100_418M/resolve/main/tokenizer_config.json', 'facebook/m2m100_1.2B': 'https://huggingface.co/facebook/m2m100_1.2B/resolve/main/tokenizer_config.json', }, } __UpperCAmelCase = { 'facebook/m2m100_418M': 1024, } # fmt: off __UpperCAmelCase = { 'm2m100': ['af', 'am', 'ar', 'ast', 'az', 'ba', 'be', 'bg', 'bn', 'br', 'bs', 'ca', 'ceb', 'cs', 'cy', 'da', 'de', 'el', 'en', 'es', 'et', 'fa', 'ff', 'fi', 'fr', 'fy', 'ga', 'gd', 'gl', 'gu', 'ha', 'he', 'hi', 'hr', 'ht', 'hu', 'hy', 'id', 'ig', 'ilo', 'is', 'it', 'ja', 'jv', 'ka', 'kk', 'km', 'kn', 'ko', 'lb', 'lg', 'ln', 'lo', 'lt', 'lv', 'mg', 'mk', 'ml', 'mn', 'mr', 'ms', 'my', 'ne', 'nl', 'no', 'ns', 'oc', 'or', 'pa', 'pl', 'ps', 'pt', 'ro', 'ru', 'sd', 'si', 'sk', 'sl', 'so', 'sq', 'sr', 'ss', 'su', 'sv', 'sw', 'ta', 'th', 'tl', 'tn', 'tr', 'uk', 'ur', 'uz', 'vi', 'wo', 'xh', 'yi', 'yo', 'zh', 'zu'], 'wmt21': ['en', 'ha', 'is', 'ja', 'cs', 'ru', 'zh', 'de'] } class __lowercase ( __lowerCamelCase ): snake_case_ = VOCAB_FILES_NAMES snake_case_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case_ = PRETRAINED_VOCAB_FILES_MAP snake_case_ = ["""input_ids""", """attention_mask"""] snake_case_ = [] snake_case_ = [] def __init__( self : int ,A : List[Any] ,A : str ,A : List[Any]=None ,A : Dict=None ,A : str="<s>" ,A : int="</s>" ,A : List[Any]="</s>" ,A : Optional[Any]="<pad>" ,A : List[str]="<unk>" ,A : Optional[Any]="m2m100" ,A : Optional[Dict[str, Any]] = None ,A : List[str]=8 ,**A : Optional[Any] ,): '''simple docstring''' UpperCAmelCase__ : List[Any] = {} if sp_model_kwargs is None else sp_model_kwargs UpperCAmelCase__ : List[str] = language_codes UpperCAmelCase__ : Dict = FAIRSEQ_LANGUAGE_CODES[language_codes] UpperCAmelCase__ : Optional[int] = {lang_code: f"__{lang_code}__" for lang_code in fairseq_language_code} UpperCAmelCase__ : Tuple = kwargs.get("""additional_special_tokens""" ,[] ) kwargs["additional_special_tokens"] += [ self.get_lang_token(A ) for lang_code in fairseq_language_code if self.get_lang_token(A ) not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=A ,tgt_lang=A ,bos_token=A ,eos_token=A ,sep_token=A ,unk_token=A ,pad_token=A ,language_codes=A ,sp_model_kwargs=self.sp_model_kwargs ,num_madeup_words=A ,**A ,) UpperCAmelCase__ : str = vocab_file UpperCAmelCase__ : Dict = load_json(A ) UpperCAmelCase__ : Optional[int] = {v: k for k, v in self.encoder.items()} UpperCAmelCase__ : Dict = spm_file UpperCAmelCase__ : List[Any] = load_spm(A ,self.sp_model_kwargs ) UpperCAmelCase__ : Dict = len(self.encoder ) UpperCAmelCase__ : Dict = { self.get_lang_token(A ): self.encoder_size + i for i, lang_code in enumerate(A ) } UpperCAmelCase__ : Union[str, Any] = {lang_code: self.encoder_size + i for i, lang_code in enumerate(A )} UpperCAmelCase__ : Union[str, Any] = {v: k for k, v in self.lang_token_to_id.items()} UpperCAmelCase__ : List[str] = src_lang if src_lang is not None else """en""" UpperCAmelCase__ : int = tgt_lang UpperCAmelCase__ : int = self.get_lang_id(self._src_lang ) self.set_src_lang_special_tokens(self._src_lang ) UpperCAmelCase__ : List[str] = num_madeup_words @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return len(self.encoder ) + len(self.lang_token_to_id ) @property def __lowercase ( self : Dict ): '''simple docstring''' return self._src_lang @src_lang.setter def __lowercase ( self : List[str] ,A : str ): '''simple docstring''' UpperCAmelCase__ : Tuple = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __lowercase ( self : Optional[Any] ,A : str ): '''simple docstring''' return self.sp_model.encode(A ,out_type=A ) def __lowercase ( self : List[str] ,A : Dict ): '''simple docstring''' if token in self.lang_token_to_id: return self.lang_token_to_id[token] return self.encoder.get(A ,self.encoder[self.unk_token] ) def __lowercase ( self : Any ,A : int ): '''simple docstring''' if index in self.id_to_lang_token: return self.id_to_lang_token[index] return self.decoder.get(A ,self.unk_token ) def __lowercase ( self : Tuple ,A : List[str] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = [] UpperCAmelCase__ : str = """""" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(A ) + token UpperCAmelCase__ : str = [] else: current_sub_tokens.append(A ) out_string += self.sp_model.decode(A ) return out_string.strip() def __lowercase ( self : Optional[int] ,A : List[int] ,A : Optional[List[int]] = None ,A : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=A ,token_ids_a=A ,already_has_special_tokens=A ) UpperCAmelCase__ : Union[str, Any] = [1] * len(self.prefix_tokens ) UpperCAmelCase__ : Optional[int] = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(A )) + suffix_ones return prefix_ones + ([0] * len(A )) + ([0] * len(A )) + suffix_ones def __lowercase ( self : Tuple ,A : List[int] ,A : Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __lowercase ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = {self.convert_ids_to_tokens(A ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Any ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.__dict__.copy() UpperCAmelCase__ : Dict = None return state def __setstate__( self : int ,A : Dict ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = d # for backward compatibility if not hasattr(self ,"""sp_model_kwargs""" ): UpperCAmelCase__ : Union[str, Any] = {} UpperCAmelCase__ : List[Any] = load_spm(self.spm_file ,self.sp_model_kwargs ) def __lowercase ( self : Any ,A : str ,A : Optional[str] = None ): '''simple docstring''' UpperCAmelCase__ : List[str] = Path(A ) if not save_dir.is_dir(): raise OSError(f"{save_directory} should be a directory" ) UpperCAmelCase__ : Tuple = save_dir / ( (filename_prefix + """-""" if filename_prefix else """""") + self.vocab_files_names["""vocab_file"""] ) UpperCAmelCase__ : Any = save_dir / ( (filename_prefix + """-""" if filename_prefix else """""") + self.vocab_files_names["""spm_file"""] ) save_json(self.encoder ,A ) if os.path.abspath(self.spm_file ) != os.path.abspath(A ) and os.path.isfile(self.spm_file ): copyfile(self.spm_file ,A ) elif not os.path.isfile(self.spm_file ): with open(A ,"""wb""" ) as fi: UpperCAmelCase__ : List[str] = self.sp_model.serialized_model_proto() fi.write(A ) return (str(A ), str(A )) def __lowercase ( self : str ,A : List[str] ,A : str = "en" ,A : Optional[List[str]] = None ,A : str = "ro" ,**A : List[Any] ,): '''simple docstring''' UpperCAmelCase__ : List[Any] = src_lang UpperCAmelCase__ : str = tgt_lang self.set_src_lang_special_tokens(self.src_lang ) return super().prepare_seqaseq_batch(A ,A ,**A ) def __lowercase ( self : Any ,A : Union[str, Any] ,A : Optional[str] ,A : Optional[str] ,**A : List[Any] ): '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError("""Translation requires a `src_lang` and a `tgt_lang` for this model""" ) UpperCAmelCase__ : List[Any] = src_lang UpperCAmelCase__ : List[str] = self(A ,add_special_tokens=A ,**A ) UpperCAmelCase__ : List[Any] = self.get_lang_id(A ) UpperCAmelCase__ : List[str] = tgt_lang_id return inputs def __lowercase ( self : Union[str, Any] ): '''simple docstring''' self.set_src_lang_special_tokens(self.src_lang ) def __lowercase ( self : List[Any] ): '''simple docstring''' self.set_tgt_lang_special_tokens(self.tgt_lang ) def __lowercase ( self : int ,A : str ): '''simple docstring''' UpperCAmelCase__ : int = self.get_lang_token(A ) UpperCAmelCase__ : List[str] = self.lang_token_to_id[lang_token] UpperCAmelCase__ : Union[str, Any] = [self.cur_lang_id] UpperCAmelCase__ : int = [self.eos_token_id] def __lowercase ( self : Dict ,A : str ): '''simple docstring''' UpperCAmelCase__ : int = self.get_lang_token(A ) UpperCAmelCase__ : List[Any] = self.lang_token_to_id[lang_token] UpperCAmelCase__ : Optional[int] = [self.cur_lang_id] UpperCAmelCase__ : str = [self.eos_token_id] def __lowercase ( self : int ,A : str ): '''simple docstring''' return self.lang_code_to_token[lang] def __lowercase ( self : Dict ,A : str ): '''simple docstring''' UpperCAmelCase__ : int = self.get_lang_token(A ) return self.lang_token_to_id[lang_token] def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = sentencepiece.SentencePieceProcessor(**__UpperCamelCase ) spm.Load(str(__UpperCamelCase ) ) return spm def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' with open(__UpperCamelCase , """r""" ) as f: return json.load(__UpperCamelCase ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' with open(__UpperCamelCase , """w""" ) as f: json.dump(__UpperCamelCase , __UpperCamelCase , indent=2 )
65
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case : Any = { "configuration_informer": [ "INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "InformerConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "InformerForPrediction", "InformerModel", "InformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys _snake_case : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
0
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyImgaImgPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.text_encoder import MCLIPConfig, MultilingualCLIP from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : List[str] = KandinskyImgaImgPipeline _UpperCamelCase : List[str] = ["prompt", "image_embeds", "negative_image_embeds", "image"] _UpperCamelCase : Dict = [ "prompt", "negative_prompt", "image_embeds", "negative_image_embeds", "image", ] _UpperCamelCase : Optional[int] = [ "generator", "height", "width", "strength", "guidance_scale", "negative_prompt", "num_inference_steps", "return_dict", "guidance_scale", "num_images_per_prompt", "output_type", "return_dict", ] _UpperCamelCase : Tuple = False @property def __a ( self ): return 3_2 @property def __a ( self ): return 3_2 @property def __a ( self ): return self.time_input_dim @property def __a ( self ): return self.time_input_dim * 4 @property def __a ( self ): return 1_0_0 @property def __a ( self ): _lowercase : int = XLMRobertaTokenizerFast.from_pretrained('YiYiXu/tiny-random-mclip-base' ) return tokenizer @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Tuple = MCLIPConfig( numDims=self.cross_attention_dim , transformerDimensions=self.text_embedder_hidden_size , hidden_size=self.text_embedder_hidden_size , intermediate_size=3_7 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=1_0_0_5 , ) _lowercase : Any = MultilingualCLIP(_lowerCAmelCase ) _lowercase : List[str] = text_encoder.eval() return text_encoder @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Tuple = { 'in_channels': 4, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'text_image', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'text_image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } _lowercase : Tuple = UNetaDConditionModel(**_lowerCAmelCase ) return model @property def __a ( self ): return { "block_out_channels": [3_2, 6_4], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 1_2, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[Any] = VQModel(**self.dummy_movq_kwargs ) return model def __a ( self ): _lowercase : Dict = self.dummy_text_encoder _lowercase : Dict = self.dummy_tokenizer _lowercase : Any = self.dummy_unet _lowercase : Optional[int] = self.dummy_movq _lowercase : Tuple = { 'num_train_timesteps': 1_0_0_0, 'beta_schedule': 'linear', 'beta_start': 0.0_00_85, 'beta_end': 0.0_12, 'clip_sample': False, 'set_alpha_to_one': False, 'steps_offset': 0, 'prediction_type': 'epsilon', 'thresholding': False, } _lowercase : int = DDIMScheduler(**_lowerCAmelCase ) _lowercase : Tuple = { 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): _lowercase : str = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : Union[str, Any] = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(seed + 1 ) ).to(_lowerCAmelCase ) # create init_image _lowercase : Tuple = floats_tensor((1, 3, 6_4, 6_4) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : Union[str, Any] = image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowercase : Union[str, Any] = Image.fromarray(np.uinta(_lowerCAmelCase ) ).convert('RGB' ).resize((2_5_6, 2_5_6) ) if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : str = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Optional[int] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : Optional[int] = { 'prompt': 'horse', 'image': init_image, 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'generator': generator, 'height': 6_4, 'width': 6_4, 'num_inference_steps': 1_0, 'guidance_scale': 7.0, 'strength': 0.2, 'output_type': 'np', } return inputs def __a ( self ): _lowercase : List[Any] = 'cpu' _lowercase : Union[str, Any] = self.get_dummy_components() _lowercase : str = self.pipeline_class(**_lowerCAmelCase ) _lowercase : List[str] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Dict = pipe(**self.get_dummy_inputs(_lowerCAmelCase ) ) _lowercase : Tuple = output.images _lowercase : Union[str, Any] = pipe( **self.get_dummy_inputs(_lowerCAmelCase ) , return_dict=_lowerCAmelCase , )[0] _lowercase : Optional[int] = image[0, -3:, -3:, -1] _lowercase : Any = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) _lowercase : Union[str, Any] = np.array( [0.61_47_49_43, 0.6_07_35_39, 0.43_30_85_44, 0.5_92_82_69, 0.47_49_35_95, 0.46_75_59_73, 0.4_61_38_38, 0.45_36_87_97, 0.50_11_92_33] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), F""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), F""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" @slow @require_torch_gpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ): _lowercase : Any = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/kandinsky_img2img_frog.npy' ) _lowercase : Optional[Any] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/cat.png' ) _lowercase : Optional[int] = 'A red cartoon frog, 4k' _lowercase : Optional[Any] = KandinskyPriorPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-1-prior' , torch_dtype=torch.floataa ) pipe_prior.to(_lowerCAmelCase ) _lowercase : Tuple = KandinskyImgaImgPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-1' , torch_dtype=torch.floataa ) _lowercase : Tuple = pipeline.to(_lowerCAmelCase ) pipeline.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Tuple = torch.Generator(device='cpu' ).manual_seed(0 ) _lowercase , _lowercase : Any = pipe_prior( _lowerCAmelCase , generator=_lowerCAmelCase , num_inference_steps=5 , negative_prompt='' , ).to_tuple() _lowercase : Any = pipeline( _lowerCAmelCase , image=_lowerCAmelCase , image_embeds=_lowerCAmelCase , negative_image_embeds=_lowerCAmelCase , generator=_lowerCAmelCase , num_inference_steps=1_0_0 , height=7_6_8 , width=7_6_8 , strength=0.2 , output_type='np' , ) _lowercase : Union[str, Any] = output.images[0] assert image.shape == (7_6_8, 7_6_8, 3) assert_mean_pixel_difference(_lowerCAmelCase , _lowerCAmelCase )
66
from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging _snake_case : Any = logging.get_logger(__name__) # pylint: disable=invalid-name class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[Any] , lowerCamelCase : CLIPSegForImageSegmentation , lowerCamelCase : CLIPSegProcessor , lowerCamelCase : AutoencoderKL , lowerCamelCase : CLIPTextModel , lowerCamelCase : CLIPTokenizer , lowerCamelCase : UNetaDConditionModel , lowerCamelCase : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , lowerCamelCase : StableDiffusionSafetyChecker , lowerCamelCase : CLIPImageProcessor , ) -> Tuple: super().__init__() if hasattr(scheduler.config , "steps_offset" ) and scheduler.config.steps_offset != 1: __snake_case : Tuple = ( F'The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`' F' should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure ' "to update the config accordingly as leaving `steps_offset` might led to incorrect results" " in future versions. If you have downloaded this checkpoint from the Hugging Face Hub," " it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`" " file" ) deprecate("steps_offset!=1" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : Any = dict(scheduler.config ) __snake_case : List[Any] = 1 __snake_case : Tuple = FrozenDict(lowerCamelCase ) if hasattr(scheduler.config , "skip_prk_steps" ) and scheduler.config.skip_prk_steps is False: __snake_case : List[str] = ( F'The configuration file of this scheduler: {scheduler} has not set the configuration' " `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make" " sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to" " incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face" " Hub, it would be very nice if you could open a Pull request for the" " `scheduler/scheduler_config.json` file" ) deprecate("skip_prk_steps not set" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : List[str] = dict(scheduler.config ) __snake_case : List[str] = True __snake_case : Any = FrozenDict(lowerCamelCase ) if safety_checker is None: logger.warning( F'You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure' " that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered" " results in services or applications open to the public. Both the diffusers team and Hugging Face" " strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling" " it only for use-cases that involve analyzing network behavior or auditing its results. For more" " information, please have a look at https://github.com/huggingface/diffusers/pull/254 ." ) self.register_modules( segmentation_model=lowerCamelCase , segmentation_processor=lowerCamelCase , vae=lowerCamelCase , text_encoder=lowerCamelCase , tokenizer=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase , safety_checker=lowerCamelCase , feature_extractor=lowerCamelCase , ) def __snake_case ( self : Dict , lowerCamelCase : Optional[Union[str, int]] = "auto" ) -> Tuple: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __snake_case : Any = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowerCamelCase ) def __snake_case ( self : List[Any] ) -> Any: self.enable_attention_slicing(lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> str: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) __snake_case : Optional[int] = torch.device("cuda" ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(lowerCamelCase , lowerCamelCase ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __snake_case ( self : int ) -> Any: if self.device != torch.device("meta" ) or not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowerCamelCase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self : List[Any] , lowerCamelCase : Union[str, List[str]] , lowerCamelCase : Union[torch.FloatTensor, PIL.Image.Image] , lowerCamelCase : str , lowerCamelCase : int = 512 , lowerCamelCase : int = 512 , lowerCamelCase : int = 50 , lowerCamelCase : float = 7.5 , lowerCamelCase : Optional[Union[str, List[str]]] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : float = 0.0 , lowerCamelCase : Optional[torch.Generator] = None , lowerCamelCase : Optional[torch.FloatTensor] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , lowerCamelCase : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , lowerCamelCase : int = 1 , **lowerCamelCase : Dict , ) -> List[str]: __snake_case : Tuple = self.segmentation_processor( text=[text] , images=[image] , padding="max_length" , return_tensors="pt" ).to(self.device ) __snake_case : str = self.segmentation_model(**lowerCamelCase ) __snake_case : Dict = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() __snake_case : List[Any] = self.numpy_to_pil(lowerCamelCase )[0].resize(image.size ) # Run inpainting pipeline with the generated mask __snake_case : Tuple = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=lowerCamelCase , image=lowerCamelCase , mask_image=lowerCamelCase , height=lowerCamelCase , width=lowerCamelCase , num_inference_steps=lowerCamelCase , guidance_scale=lowerCamelCase , negative_prompt=lowerCamelCase , num_images_per_prompt=lowerCamelCase , eta=lowerCamelCase , generator=lowerCamelCase , latents=lowerCamelCase , output_type=lowerCamelCase , return_dict=lowerCamelCase , callback=lowerCamelCase , callback_steps=lowerCamelCase , )
81
0
from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case = logging.get_logger(__name__) snake_case = { """transfo-xl-wt103""": """https://huggingface.co/transfo-xl-wt103/resolve/main/config.json""", } class A_ ( UpperCAmelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = '''transfo-xl''' SCREAMING_SNAKE_CASE_ : Optional[Any] = ['''mems'''] SCREAMING_SNAKE_CASE_ : Optional[Any] = { '''n_token''': '''vocab_size''', '''hidden_size''': '''d_model''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : Optional[Any] ,__A : Union[str, Any]=26_7735 ,__A : List[Any]=[2_0000, 4_0000, 20_0000] ,__A : Dict=1024 ,__A : str=1024 ,__A : Dict=16 ,__A : int=64 ,__A : Dict=4096 ,__A : List[Any]=4 ,__A : Optional[int]=False ,__A : Union[str, Any]=18 ,__A : Tuple=1600 ,__A : str=1000 ,__A : Dict=True ,__A : Dict=True ,__A : int=0 ,__A : Optional[int]=-1 ,__A : int=True ,__A : List[str]=0.1 ,__A : Optional[int]=0.0 ,__A : str=True ,__A : Tuple="normal" ,__A : Union[str, Any]=0.01 ,__A : Tuple=0.01 ,__A : Any=0.02 ,__A : Union[str, Any]=1e-5 ,__A : List[Any]=0 ,**__A : str ,) -> List[Any]: _lowercase = vocab_size _lowercase = [] self.cutoffs.extend(__A ) if proj_share_all_but_first: _lowercase = [False] + [True] * len(self.cutoffs ) else: _lowercase = [False] + [False] * len(self.cutoffs ) _lowercase = d_model _lowercase = d_embed _lowercase = d_head _lowercase = d_inner _lowercase = div_val _lowercase = pre_lnorm _lowercase = n_layer _lowercase = n_head _lowercase = mem_len _lowercase = same_length _lowercase = attn_type _lowercase = clamp_len _lowercase = sample_softmax _lowercase = adaptive _lowercase = dropout _lowercase = dropatt _lowercase = untie_r _lowercase = init _lowercase = init_range _lowercase = proj_init_std _lowercase = init_std _lowercase = layer_norm_epsilon super().__init__(eos_token_id=__A ,**__A ) @property def __UpperCAmelCase ( self : str ) -> Optional[int]: # Message copied from Transformer-XL documentation logger.info(F"""The model {self.model_type} is one of the few models that has no sequence length limit.""" ) return -1 @max_position_embeddings.setter def __UpperCAmelCase ( self : Any ,__A : Dict ) -> Optional[Any]: # Message copied from Transformer-XL documentation raise NotImplementedError( F"""The model {self.model_type} is one of the few models that has no sequence length limit.""" )
67
class a : """simple docstring""" def __init__( self : Tuple , lowerCamelCase : list ) -> None: __snake_case : str = set_counts __snake_case : Union[str, Any] = max(lowerCamelCase ) __snake_case : List[Any] = len(lowerCamelCase ) __snake_case : Tuple = [1] * num_sets __snake_case : Dict = list(range(lowerCamelCase ) ) def __snake_case ( self : str , lowerCamelCase : int , lowerCamelCase : int ) -> bool: __snake_case : List[Any] = self.get_parent(lowerCamelCase ) __snake_case : Tuple = self.get_parent(lowerCamelCase ) if src_parent == dst_parent: return False if self.ranks[dst_parent] >= self.ranks[src_parent]: self.set_counts[dst_parent] += self.set_counts[src_parent] __snake_case : List[str] = 0 __snake_case : List[Any] = dst_parent if self.ranks[dst_parent] == self.ranks[src_parent]: self.ranks[dst_parent] += 1 __snake_case : Dict = self.set_counts[dst_parent] else: self.set_counts[src_parent] += self.set_counts[dst_parent] __snake_case : Union[str, Any] = 0 __snake_case : Optional[int] = src_parent __snake_case : Tuple = self.set_counts[src_parent] __snake_case : str = max(self.max_set , lowerCamelCase ) return True def __snake_case ( self : int , lowerCamelCase : int ) -> int: if self.parents[disj_set] == disj_set: return disj_set __snake_case : Optional[int] = self.get_parent(self.parents[disj_set] ) return self.parents[disj_set]
81
0
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json import os from ...utils.constants import SAGEMAKER_PARALLEL_EC2_INSTANCES, TORCH_DYNAMO_MODES from ...utils.dataclasses import ComputeEnvironment, SageMakerDistributedType from ...utils.imports import is_botoa_available from .config_args import SageMakerConfig from .config_utils import ( DYNAMO_BACKENDS, _ask_field, _ask_options, _convert_dynamo_backend, _convert_mixed_precision, _convert_sagemaker_distributed_mode, _convert_yes_no_to_bool, ) if is_botoa_available(): import botoa # noqa: F401 def lowercase__ ( A_: Union[str, Any] ) -> List[Any]: """simple docstring""" __UpperCAmelCase =botoa.client("""iam""" ) __UpperCAmelCase ={ """Version""": """2012-10-17""", """Statement""": [ {"""Effect""": """Allow""", """Principal""": {"""Service""": """sagemaker.amazonaws.com"""}, """Action""": """sts:AssumeRole"""} ], } try: # create the role, associated with the chosen trust policy iam_client.create_role( RoleName=A_ , AssumeRolePolicyDocument=json.dumps(A_ , indent=2 ) ) __UpperCAmelCase ={ """Version""": """2012-10-17""", """Statement""": [ { """Effect""": """Allow""", """Action""": [ """sagemaker:*""", """ecr:GetDownloadUrlForLayer""", """ecr:BatchGetImage""", """ecr:BatchCheckLayerAvailability""", """ecr:GetAuthorizationToken""", """cloudwatch:PutMetricData""", """cloudwatch:GetMetricData""", """cloudwatch:GetMetricStatistics""", """cloudwatch:ListMetrics""", """logs:CreateLogGroup""", """logs:CreateLogStream""", """logs:DescribeLogStreams""", """logs:PutLogEvents""", """logs:GetLogEvents""", """s3:CreateBucket""", """s3:ListBucket""", """s3:GetBucketLocation""", """s3:GetObject""", """s3:PutObject""", ], """Resource""": """*""", } ], } # attach policy to role iam_client.put_role_policy( RoleName=A_ , PolicyName=F'''{role_name}_policy_permission''' , PolicyDocument=json.dumps(A_ , indent=2 ) , ) except iam_client.exceptions.EntityAlreadyExistsException: print(F'''role {role_name} already exists. Using existing one''' ) def lowercase__ ( A_: Dict ) -> Any: """simple docstring""" __UpperCAmelCase =botoa.client("""iam""" ) return iam_client.get_role(RoleName=A_ )["Role"]["Arn"] def lowercase__ ( ) -> Union[str, Any]: """simple docstring""" __UpperCAmelCase =_ask_options( """How do you want to authorize?""" , ["""AWS Profile""", """Credentials (AWS_ACCESS_KEY_ID, AWS_SECRET_ACCESS_KEY) """] , A_ , ) __UpperCAmelCase =None if credentials_configuration == 0: __UpperCAmelCase =_ask_field("""Enter your AWS Profile name: [default] """ , default="""default""" ) __UpperCAmelCase =aws_profile else: print( """Note you will need to provide AWS_ACCESS_KEY_ID and AWS_SECRET_ACCESS_KEY when you launch you training script with,""" """`accelerate launch --aws_access_key_id XXX --aws_secret_access_key YYY`""" ) __UpperCAmelCase =_ask_field("""AWS Access Key ID: """ ) __UpperCAmelCase =aws_access_key_id __UpperCAmelCase =_ask_field("""AWS Secret Access Key: """ ) __UpperCAmelCase =aws_secret_access_key __UpperCAmelCase =_ask_field("""Enter your AWS Region: [us-east-1]""" , default="""us-east-1""" ) __UpperCAmelCase =aws_region __UpperCAmelCase =_ask_options( """Do you already have an IAM Role for executing Amazon SageMaker Training Jobs?""" , ["""Provide IAM Role name""", """Create new IAM role using credentials"""] , A_ , ) if role_management == 0: __UpperCAmelCase =_ask_field("""Enter your IAM role name: """ ) else: __UpperCAmelCase ="""accelerate_sagemaker_execution_role""" print(F'''Accelerate will create an iam role "{iam_role_name}" using the provided credentials''' ) _create_iam_role_for_sagemaker(A_ ) __UpperCAmelCase =_ask_field( """Do you want to use custom Docker image? [yes/NO]: """ , _convert_yes_no_to_bool , default=A_ , error_message="""Please enter yes or no.""" , ) __UpperCAmelCase =None if is_custom_docker_image: __UpperCAmelCase =_ask_field("""Enter your Docker image: """ , lambda A_ : str(A_ ).lower() ) __UpperCAmelCase =_ask_field( """Do you want to provide SageMaker input channels with data locations? [yes/NO]: """ , _convert_yes_no_to_bool , default=A_ , error_message="""Please enter yes or no.""" , ) __UpperCAmelCase =None if is_sagemaker_inputs_enabled: __UpperCAmelCase =_ask_field( """Enter the path to the SageMaker inputs TSV file with columns (channel_name, data_location): """ , lambda A_ : str(A_ ).lower() , ) __UpperCAmelCase =_ask_field( """Do you want to enable SageMaker metrics? [yes/NO]: """ , _convert_yes_no_to_bool , default=A_ , error_message="""Please enter yes or no.""" , ) __UpperCAmelCase =None if is_sagemaker_metrics_enabled: __UpperCAmelCase =_ask_field( """Enter the path to the SageMaker metrics TSV file with columns (metric_name, metric_regex): """ , lambda A_ : str(A_ ).lower() , ) __UpperCAmelCase =_ask_options( """What is the distributed mode?""" , ["""No distributed training""", """Data parallelism"""] , _convert_sagemaker_distributed_mode , ) __UpperCAmelCase ={} __UpperCAmelCase =_ask_field( """Do you wish to optimize your script with torch dynamo?[yes/NO]:""" , _convert_yes_no_to_bool , default=A_ , error_message="""Please enter yes or no.""" , ) if use_dynamo: __UpperCAmelCase ="""dynamo_""" __UpperCAmelCase =_ask_options( """Which dynamo backend would you like to use?""" , [x.lower() for x in DYNAMO_BACKENDS] , _convert_dynamo_backend , default=2 , ) __UpperCAmelCase =_ask_field( """Do you want to customize the defaults sent to torch.compile? [yes/NO]: """ , _convert_yes_no_to_bool , default=A_ , error_message="""Please enter yes or no.""" , ) if use_custom_options: __UpperCAmelCase =_ask_options( """Which mode do you want to use?""" , A_ , lambda A_ : TORCH_DYNAMO_MODES[int(A_ )] , default="""default""" , ) __UpperCAmelCase =_ask_field( """Do you want the fullgraph mode or it is ok to break model into several subgraphs? [yes/NO]: """ , _convert_yes_no_to_bool , default=A_ , error_message="""Please enter yes or no.""" , ) __UpperCAmelCase =_ask_field( """Do you want to enable dynamic shape tracing? [yes/NO]: """ , _convert_yes_no_to_bool , default=A_ , error_message="""Please enter yes or no.""" , ) __UpperCAmelCase ="""Which EC2 instance type you want to use for your training?""" if distributed_type != SageMakerDistributedType.NO: __UpperCAmelCase =_ask_options( A_ , A_ , lambda A_ : SAGEMAKER_PARALLEL_EC2_INSTANCES[int(A_ )] ) else: eca_instance_query += "? [ml.p3.2xlarge]:" __UpperCAmelCase =_ask_field(A_ , lambda A_ : str(A_ ).lower() , default="""ml.p3.2xlarge""" ) __UpperCAmelCase =1 if distributed_type in (SageMakerDistributedType.DATA_PARALLEL, SageMakerDistributedType.MODEL_PARALLEL): __UpperCAmelCase =_ask_field( """How many machines do you want use? [1]: """ , A_ , default=1 , ) __UpperCAmelCase =_ask_options( """Do you wish to use FP16 or BF16 (mixed precision)?""" , ["""no""", """fp16""", """bf16""", """fp8"""] , _convert_mixed_precision , ) if use_dynamo and mixed_precision == "no": print( """Torch dynamo used without mixed precision requires TF32 to be efficient. Accelerate will enable it by default when launching your scripts.""" ) return SageMakerConfig( image_uri=A_ , compute_environment=ComputeEnvironment.AMAZON_SAGEMAKER , distributed_type=A_ , use_cpu=A_ , dynamo_config=A_ , eca_instance_type=A_ , profile=A_ , region=A_ , iam_role_name=A_ , mixed_precision=A_ , num_machines=A_ , sagemaker_inputs_file=A_ , sagemaker_metrics_file=A_ , )
68
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class a (unittest.TestCase ): """simple docstring""" def __init__( self : int , lowerCamelCase : Any , lowerCamelCase : int=7 , lowerCamelCase : str=3 , lowerCamelCase : Optional[int]=30 , lowerCamelCase : Dict=400 , lowerCamelCase : str=True , lowerCamelCase : str=None , lowerCamelCase : Any=True , lowerCamelCase : Union[str, Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[str]=True , lowerCamelCase : Optional[int]=1 / 255 , lowerCamelCase : Any=True , ) -> str: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __snake_case : Optional[Any] = size if size is not None else {"shortest_edge": 18, "longest_edge": 1333} __snake_case : Optional[Any] = parent __snake_case : List[Any] = batch_size __snake_case : Optional[int] = num_channels __snake_case : str = min_resolution __snake_case : int = max_resolution __snake_case : int = do_resize __snake_case : Tuple = size __snake_case : Any = do_normalize __snake_case : int = image_mean __snake_case : Tuple = image_std __snake_case : Dict = do_rescale __snake_case : Optional[Any] = rescale_factor __snake_case : str = do_pad def __snake_case ( self : Any ) -> int: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def __snake_case ( self : Dict , lowerCamelCase : List[Any] , lowerCamelCase : Optional[Any]=False ) -> List[str]: if not batched: __snake_case : Dict = image_inputs[0] if isinstance(lowerCamelCase , Image.Image ): __snake_case , __snake_case : Dict = image.size else: __snake_case , __snake_case : List[str] = image.shape[1], image.shape[2] if w < h: __snake_case : Optional[int] = int(self.size["shortest_edge"] * h / w ) __snake_case : int = self.size["shortest_edge"] elif w > h: __snake_case : List[str] = self.size["shortest_edge"] __snake_case : Optional[Any] = int(self.size["shortest_edge"] * w / h ) else: __snake_case : List[Any] = self.size["shortest_edge"] __snake_case : Any = self.size["shortest_edge"] else: __snake_case : int = [] for image in image_inputs: __snake_case , __snake_case : List[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[0] )[0] __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class a (_lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ConditionalDetrImageProcessor if is_vision_available() else None def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : Optional[Any] = ConditionalDetrImageProcessingTester(self ) @property def __snake_case ( self : Any ) -> str: return self.image_processor_tester.prepare_image_processor_dict() def __snake_case ( self : Optional[Any] ) -> Optional[int]: __snake_case : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase , "image_mean" ) ) self.assertTrue(hasattr(lowerCamelCase , "image_std" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_normalize" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase , "size" ) ) def __snake_case ( self : Any ) -> Dict: __snake_case : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 1333} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) __snake_case : str = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=lowerCamelCase ) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> Dict: pass def __snake_case ( self : Tuple ) -> str: # Initialize image_processing __snake_case : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , Image.Image ) # Test not batched input __snake_case : int = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) __snake_case : Dict = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> str: # Initialize image_processing __snake_case : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __snake_case : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , numpify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , np.ndarray ) # Test not batched input __snake_case : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : List[Any] = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> List[str]: # Initialize image_processing __snake_case : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , torchify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , torch.Tensor ) # Test not batched input __snake_case : Optional[Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : List[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : int = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def __snake_case ( self : Any ) -> Optional[int]: # prepare image and target __snake_case : List[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : List[Any] = {"image_id": 39769, "annotations": target} # encode them __snake_case : List[str] = ConditionalDetrImageProcessor.from_pretrained("microsoft/conditional-detr-resnet-50" ) __snake_case : List[str] = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : Tuple = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Optional[int] = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : List[Any] = torch.tensor([58_87.96_00, 1_12_50.20_61, 48_93_53.84_38, 83_71_22.75_00, 14_79_67.51_56, 16_57_32.34_38] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : Optional[int] = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([0.55_03, 0.27_65, 0.06_04, 0.22_15] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Optional[Any] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : List[Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : Dict = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify orig_size __snake_case : int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Tuple = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) ) @slow def __snake_case ( self : str ) -> Tuple: # prepare image, target and masks_path __snake_case : Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : str = {"file_name": "000000039769.png", "image_id": 39769, "segments_info": target} __snake_case : Optional[int] = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them __snake_case : int = ConditionalDetrImageProcessor(format="coco_panoptic" ) __snake_case : str = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , masks_path=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : List[str] = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Dict = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : Any = torch.tensor([14_79_79.68_75, 16_55_27.04_69, 48_46_38.59_38, 1_12_92.93_75, 58_79.65_62, 76_34.11_47] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : str = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : Optional[Any] = torch.tensor([0.26_25, 0.54_37, 0.46_88, 0.86_25] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Tuple = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : Dict = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : int = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify masks __snake_case : List[Any] = 822873 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , lowerCamelCase ) # verify orig_size __snake_case : List[str] = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Any = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) )
81
0
'''simple docstring''' from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class SCREAMING_SNAKE_CASE__ : __SCREAMING_SNAKE_CASE = 42 # [batch_size x 3] __SCREAMING_SNAKE_CASE = 42 # [batch_size x 3] __SCREAMING_SNAKE_CASE = 42 # [batch_size x 3] __SCREAMING_SNAKE_CASE = 42 # [batch_size x 3] __SCREAMING_SNAKE_CASE = 42 __SCREAMING_SNAKE_CASE = 42 __SCREAMING_SNAKE_CASE = 42 __SCREAMING_SNAKE_CASE = 42 __SCREAMING_SNAKE_CASE = 42 def A ( self : List[str] ): """simple docstring""" assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def A ( self : Dict ): """simple docstring""" return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def A ( self : Optional[Any] ): """simple docstring""" return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def A ( self : Optional[int] ): """simple docstring""" __snake_case = torch.arange(self.height * self.width ) __snake_case = torch.stack( [ pixel_indices % self.width, torch.div(a_ , self.width , rounding_mode="trunc" ), ] , axis=1 , ) return coords @property def A ( self : Optional[int] ): """simple docstring""" __snake_case , *__snake_case = self.shape __snake_case = int(np.prod(a_ ) ) __snake_case = self.get_image_coords() __snake_case = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) __snake_case = self.get_camera_rays(a_ ) __snake_case = rays.view(a_ , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def A ( self : Optional[int] , a_ : torch.Tensor ): """simple docstring""" __snake_case , *__snake_case , __snake_case = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] __snake_case = coords.view(a_ , -1 , 2 ) __snake_case = self.resolution() __snake_case = self.fov() __snake_case = (flat.float() / (res - 1)) * 2 - 1 __snake_case = fracs * torch.tan(fov / 2 ) __snake_case = fracs.view(a_ , -1 , 2 ) __snake_case = ( self.z.view(a_ , 1 , 3 ) + self.x.view(a_ , 1 , 3 ) * fracs[:, :, :1] + self.y.view(a_ , 1 , 3 ) * fracs[:, :, 1:] ) __snake_case = directions / directions.norm(dim=-1 , keepdim=a_ ) __snake_case = torch.stack( [ torch.broadcast_to(self.origin.view(a_ , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(a_ , *a_ , 2 , 3 ) def A ( self : List[str] , a_ : int , a_ : int ): """simple docstring""" assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=a_ , height=a_ , x_fov=self.x_fov , y_fov=self.y_fov , ) def __UpperCAmelCase ( _UpperCAmelCase : int ) -> DifferentiableProjectiveCamera: __snake_case = [] __snake_case = [] __snake_case = [] __snake_case = [] for theta in np.linspace(0 , 2 * np.pi , num=20 ): __snake_case = np.array([np.sin(_UpperCAmelCase ), np.cos(_UpperCAmelCase ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) __snake_case = -z * 4 __snake_case = np.array([np.cos(_UpperCAmelCase ), -np.sin(_UpperCAmelCase ), 0.0] ) __snake_case = np.cross(_UpperCAmelCase , _UpperCAmelCase ) origins.append(_UpperCAmelCase ) xs.append(_UpperCAmelCase ) ys.append(_UpperCAmelCase ) zs.append(_UpperCAmelCase ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(_UpperCAmelCase , axis=0 ) ).float() , x=torch.from_numpy(np.stack(_UpperCAmelCase , axis=0 ) ).float() , y=torch.from_numpy(np.stack(_UpperCAmelCase , axis=0 ) ).float() , z=torch.from_numpy(np.stack(_UpperCAmelCase , axis=0 ) ).float() , width=_UpperCAmelCase , height=_UpperCAmelCase , x_fov=0.7 , y_fov=0.7 , shape=(1, len(_UpperCAmelCase )) , )
69
from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from ... import AutoBackbone from ...modeling_outputs import SemanticSegmenterOutput from ...modeling_utils import PreTrainedModel from ...utils import add_start_docstrings, add_start_docstrings_to_model_forward, replace_return_docstrings from ...utils.backbone_utils import BackboneMixin from .configuration_upernet import UperNetConfig _snake_case : Optional[Any] = [ "openmmlab/upernet-convnext-tiny", # See all UperNet models at https://huggingface.co/models?filter=upernet ] # General docstring _snake_case : Dict = "UperNetConfig" class a (nn.Module ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : Union[int, Tuple[int, int]] , lowerCamelCase : Union[int, Tuple[int, int], str] = 0 , lowerCamelCase : bool = False , lowerCamelCase : Union[int, Tuple[int, int]] = 1 , ) -> None: super().__init__() __snake_case : Union[str, Any] = nn.Convad( in_channels=lowerCamelCase , out_channels=lowerCamelCase , kernel_size=lowerCamelCase , padding=lowerCamelCase , bias=lowerCamelCase , dilation=lowerCamelCase , ) __snake_case : Dict = nn.BatchNormad(lowerCamelCase ) __snake_case : List[Any] = nn.ReLU() def __snake_case ( self : List[Any] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : Dict = self.conv(lowerCamelCase ) __snake_case : int = self.batch_norm(lowerCamelCase ) __snake_case : Optional[Any] = self.activation(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : str , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : int ) -> None: super().__init__() __snake_case : Tuple = [ nn.AdaptiveAvgPoolad(lowerCamelCase ), UperNetConvModule(lowerCamelCase , lowerCamelCase , kernel_size=1 ), ] for i, layer in enumerate(self.layers ): self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : Dict , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : List[str] = input for layer in self.layers: __snake_case : Tuple = layer(lowerCamelCase ) return hidden_state class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Tuple[int, ...] , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : bool ) -> None: super().__init__() __snake_case : Dict = pool_scales __snake_case : List[str] = align_corners __snake_case : List[Any] = in_channels __snake_case : str = channels __snake_case : Optional[Any] = [] for i, pool_scale in enumerate(lowerCamelCase ): __snake_case : Union[str, Any] = UperNetPyramidPoolingBlock(pool_scale=lowerCamelCase , in_channels=lowerCamelCase , channels=lowerCamelCase ) self.blocks.append(lowerCamelCase ) self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> List[torch.Tensor]: __snake_case : Tuple = [] for ppm in self.blocks: __snake_case : Any = ppm(lowerCamelCase ) __snake_case : List[Any] = nn.functional.interpolate( lowerCamelCase , size=x.size()[2:] , mode="bilinear" , align_corners=self.align_corners ) ppm_outs.append(lowerCamelCase ) return ppm_outs class a (nn.Module ): """simple docstring""" def __init__( self : int , lowerCamelCase : str , lowerCamelCase : str ) -> List[str]: super().__init__() __snake_case : Dict = config __snake_case : List[str] = config.pool_scales # e.g. (1, 2, 3, 6) __snake_case : Tuple = in_channels __snake_case : str = config.hidden_size __snake_case : List[str] = False __snake_case : Any = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) # PSP Module __snake_case : Tuple = UperNetPyramidPoolingModule( self.pool_scales , self.in_channels[-1] , self.channels , align_corners=self.align_corners , ) __snake_case : List[str] = UperNetConvModule( self.in_channels[-1] + len(self.pool_scales ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) # FPN Module __snake_case : List[Any] = nn.ModuleList() __snake_case : Dict = nn.ModuleList() for in_channels in self.in_channels[:-1]: # skip the top layer __snake_case : Union[str, Any] = UperNetConvModule(lowerCamelCase , self.channels , kernel_size=1 ) __snake_case : Optional[int] = UperNetConvModule(self.channels , self.channels , kernel_size=3 , padding=1 ) self.lateral_convs.append(lowerCamelCase ) self.fpn_convs.append(lowerCamelCase ) __snake_case : int = UperNetConvModule( len(self.in_channels ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) def __snake_case ( self : List[str] ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> str: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : List[Any] , lowerCamelCase : Tuple ) -> Optional[int]: __snake_case : str = inputs[-1] __snake_case : int = [x] psp_outs.extend(self.psp_modules(lowerCamelCase ) ) __snake_case : Tuple = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Union[str, Any] = self.bottleneck(lowerCamelCase ) return output def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # build laterals __snake_case : Any = [lateral_conv(encoder_hidden_states[i] ) for i, lateral_conv in enumerate(self.lateral_convs )] laterals.append(self.psp_forward(lowerCamelCase ) ) # build top-down path __snake_case : Dict = len(lowerCamelCase ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Union[str, Any] = laterals[i - 1].shape[2:] __snake_case : Optional[int] = laterals[i - 1] + nn.functional.interpolate( laterals[i] , size=lowerCamelCase , mode="bilinear" , align_corners=self.align_corners ) # build outputs __snake_case : str = [self.fpn_convs[i](laterals[i] ) for i in range(used_backbone_levels - 1 )] # append psp feature fpn_outs.append(laterals[-1] ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Tuple = nn.functional.interpolate( fpn_outs[i] , size=fpn_outs[0].shape[2:] , mode="bilinear" , align_corners=self.align_corners ) __snake_case : str = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Optional[Any] = self.fpn_bottleneck(lowerCamelCase ) __snake_case : Tuple = self.classifier(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Optional[int] , lowerCamelCase : int = 2 , lowerCamelCase : int = 3 , lowerCamelCase : Union[int, Tuple[int, int]] = 1 ) -> None: super().__init__() __snake_case : List[Any] = config __snake_case : List[str] = config.auxiliary_in_channels __snake_case : List[Any] = config.auxiliary_channels __snake_case : Tuple = config.auxiliary_num_convs __snake_case : int = config.auxiliary_concat_input __snake_case : Optional[int] = in_index __snake_case : Tuple = (kernel_size // 2) * dilation __snake_case : Optional[int] = [] convs.append( UperNetConvModule( self.in_channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) for i in range(self.num_convs - 1 ): convs.append( UperNetConvModule( self.channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) if self.num_convs == 0: __snake_case : Union[str, Any] = nn.Identity() else: __snake_case : Any = nn.Sequential(*lowerCamelCase ) if self.concat_input: __snake_case : int = UperNetConvModule( self.in_channels + self.channels , self.channels , kernel_size=lowerCamelCase , padding=kernel_size // 2 ) __snake_case : Optional[Any] = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) def __snake_case ( self : Dict ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : Tuple ) -> Optional[int]: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : Optional[int] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # just take the relevant feature maps __snake_case : List[str] = encoder_hidden_states[self.in_index] __snake_case : Optional[Any] = self.convs(lowerCamelCase ) if self.concat_input: __snake_case : Optional[Any] = self.conv_cat(torch.cat([hidden_states, output] , dim=1 ) ) __snake_case : Union[str, Any] = self.classifier(lowerCamelCase ) return output class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = UperNetConfig __UpperCAmelCase : int = "pixel_values" __UpperCAmelCase : str = True def __snake_case ( self : Union[str, Any] , lowerCamelCase : List[Any] ) -> List[Any]: if isinstance(lowerCamelCase , lowerCamelCase ): module.backbone.init_weights() module.decode_head.init_weights() module.auxiliary_head.init_weights() def __snake_case ( self : Optional[Any] ) -> List[str]: self.backbone.init_weights() self.decode_head.init_weights() self.auxiliary_head.init_weights() def __snake_case ( self : List[Any] , lowerCamelCase : Any , lowerCamelCase : Optional[Any]=False ) -> Dict: if isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : Union[str, Any] = value _snake_case : Dict = R"\n Parameters:\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n config ([`UperNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" _snake_case : Tuple = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using\n [`AutoImageProcessor`]. See [`SegformerImageProcessor.__call__`] for details.\n output_attentions (`bool`, *optional*):\n Whether or not to return the attentions tensors of all attention layers in case the backbone has them. See\n `attentions` under returned tensors for more detail.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers of the backbone. See `hidden_states` under\n returned tensors for more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n" @add_start_docstrings( "UperNet framework leveraging any vision backbone e.g. for ADE20k, CityScapes." , _lowerCAmelCase , ) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Dict , lowerCamelCase : int ) -> Optional[int]: super().__init__(lowerCamelCase ) __snake_case : Any = AutoBackbone.from_config(config.backbone_config ) # Semantic segmentation head(s) __snake_case : Union[str, Any] = UperNetHead(lowerCamelCase , in_channels=self.backbone.channels ) __snake_case : Any = UperNetFCNHead(lowerCamelCase ) if config.use_auxiliary_head else None # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UPERNET_INPUTS_DOCSTRING.format("batch_size, sequence_length" ) ) @replace_return_docstrings(output_type=lowerCamelCase , config_class=_CONFIG_FOR_DOC ) def __snake_case ( self : Union[str, Any] , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , ) -> Union[tuple, SemanticSegmenterOutput]: __snake_case : Any = return_dict if return_dict is not None else self.config.use_return_dict __snake_case : int = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __snake_case : str = output_attentions if output_attentions is not None else self.config.output_attentions __snake_case : Tuple = self.backbone.forward_with_filtered_kwargs( lowerCamelCase , output_hidden_states=lowerCamelCase , output_attentions=lowerCamelCase ) __snake_case : List[Any] = outputs.feature_maps __snake_case : List[Any] = self.decode_head(lowerCamelCase ) __snake_case : List[str] = nn.functional.interpolate(lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : Optional[int] = None if self.auxiliary_head is not None: __snake_case : Dict = self.auxiliary_head(lowerCamelCase ) __snake_case : Dict = nn.functional.interpolate( lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : int = None if labels is not None: if self.config.num_labels == 1: raise ValueError("The number of labels should be greater than one" ) else: # compute weighted loss __snake_case : Any = CrossEntropyLoss(ignore_index=self.config.loss_ignore_index ) __snake_case : Union[str, Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = main_loss + self.config.auxiliary_loss_weight * auxiliary_loss if not return_dict: if output_hidden_states: __snake_case : Any = (logits,) + outputs[1:] else: __snake_case : str = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return SemanticSegmenterOutput( loss=lowerCamelCase , logits=lowerCamelCase , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
81
0
from __future__ import annotations import numpy as np from numpy import floataa from numpy.typing import NDArray def _SCREAMING_SNAKE_CASE ( lowercase : NDArray[floataa] , lowercase : NDArray[floataa] , lowercase : list[int] , lowercase : int , ): '''simple docstring''' lowerCamelCase_ , lowerCamelCase_ = coefficient_matrix.shape lowerCamelCase_ , lowerCamelCase_ = constant_matrix.shape if rowsa != colsa: lowerCamelCase_ = f"""Coefficient matrix dimensions must be nxn but received {rowsa}x{colsa}""" raise ValueError(lowercase ) if colsa != 1: lowerCamelCase_ = f"""Constant matrix must be nx1 but received {rowsa}x{colsa}""" raise ValueError(lowercase ) if rowsa != rowsa: lowerCamelCase_ = ( 'Coefficient and constant matrices dimensions must be nxn and nx1 but ' f"""received {rowsa}x{colsa} and {rowsa}x{colsa}""" ) raise ValueError(lowercase ) if len(lowercase ) != rowsa: lowerCamelCase_ = ( 'Number of initial values must be equal to number of rows in coefficient ' f"""matrix but received {len(lowercase )} and {rowsa}""" ) raise ValueError(lowercase ) if iterations <= 0: raise ValueError('Iterations must be at least 1' ) lowerCamelCase_ = np.concatenate( (coefficient_matrix, constant_matrix) , axis=1 ) lowerCamelCase_ , lowerCamelCase_ = table.shape strictly_diagonally_dominant(lowercase ) # Iterates the whole matrix for given number of times for _ in range(lowercase ): lowerCamelCase_ = [] for row in range(lowercase ): lowerCamelCase_ = 0 for col in range(lowercase ): if col == row: lowerCamelCase_ = table[row][col] elif col == cols - 1: lowerCamelCase_ = table[row][col] else: temp += (-1) * table[row][col] * init_val[col] lowerCamelCase_ = (temp + val) / denom new_val.append(lowercase ) lowerCamelCase_ = new_val return [float(lowercase ) for i in new_val] def _SCREAMING_SNAKE_CASE ( lowercase : NDArray[floataa] ): '''simple docstring''' lowerCamelCase_ , lowerCamelCase_ = table.shape lowerCamelCase_ = True for i in range(0 , lowercase ): lowerCamelCase_ = 0 for j in range(0 , cols - 1 ): if i == j: continue else: total += table[i][j] if table[i][i] <= total: raise ValueError('Coefficient matrix is not strictly diagonally dominant' ) return is_diagonally_dominant # Test Cases if __name__ == "__main__": import doctest doctest.testmod()
70
from __future__ import annotations def lowerCAmelCase_ ( __lowerCamelCase ): if not nums: return 0 __snake_case : Any = nums[0] __snake_case : str = 0 for num in nums[1:]: __snake_case , __snake_case : List[str] = ( max_excluding + num, max(__lowerCamelCase , __lowerCamelCase ), ) return max(__lowerCamelCase , __lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
81
0
'''simple docstring''' import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def a__ ( _SCREAMING_SNAKE_CASE : Dict ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : Tuple = image.size UpperCAmelCase_ , UpperCAmelCase_ : Dict = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 UpperCAmelCase_ : Union[str, Any] = image.resize((w, h) , resample=PIL_INTERPOLATION["lanczos"] ) UpperCAmelCase_ : List[str] = np.array(_SCREAMING_SNAKE_CASE ).astype(np.floataa ) / 255.0 UpperCAmelCase_ : Tuple = image[None].transpose(0 , 3 , 1 , 2 ) UpperCAmelCase_ : Dict = torch.from_numpy(_SCREAMING_SNAKE_CASE ) return 2.0 * image - 1.0 class _snake_case (__SCREAMING_SNAKE_CASE): def __init__( self ,_snake_case ,_snake_case ,_snake_case ,): super().__init__() self.register_modules(vqvae=_snake_case ,unet=_snake_case ,scheduler=_snake_case ) @torch.no_grad() def __call__( self ,_snake_case = None ,_snake_case = 1 ,_snake_case = 1_00 ,_snake_case = 0.0 ,_snake_case = None ,_snake_case = "pil" ,_snake_case = True ,): if isinstance(_snake_case ,PIL.Image.Image ): UpperCAmelCase_ : Dict = 1 elif isinstance(_snake_case ,torch.Tensor ): UpperCAmelCase_ : int = image.shape[0] else: raise ValueError(f'''`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(_snake_case )}''' ) if isinstance(_snake_case ,PIL.Image.Image ): UpperCAmelCase_ : Optional[int] = preprocess(_snake_case ) UpperCAmelCase_ , UpperCAmelCase_ : Any = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image UpperCAmelCase_ : str = (batch_size, self.unet.config.in_channels // 2, height, width) UpperCAmelCase_ : List[Any] = next(self.unet.parameters() ).dtype UpperCAmelCase_ : int = randn_tensor(_snake_case ,generator=_snake_case ,device=self.device ,dtype=_snake_case ) UpperCAmelCase_ : List[Any] = image.to(device=self.device ,dtype=_snake_case ) # set timesteps and move to the correct device self.scheduler.set_timesteps(_snake_case ,device=self.device ) UpperCAmelCase_ : Any = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler UpperCAmelCase_ : Optional[int] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] UpperCAmelCase_ : Union[str, Any] = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) UpperCAmelCase_ : str = {} if accepts_eta: UpperCAmelCase_ : Dict = eta for t in self.progress_bar(_snake_case ): # concat latents and low resolution image in the channel dimension. UpperCAmelCase_ : Optional[Any] = torch.cat([latents, image] ,dim=1 ) UpperCAmelCase_ : Tuple = self.scheduler.scale_model_input(_snake_case ,_snake_case ) # predict the noise residual UpperCAmelCase_ : List[Any] = self.unet(_snake_case ,_snake_case ).sample # compute the previous noisy sample x_t -> x_t-1 UpperCAmelCase_ : str = self.scheduler.step(_snake_case ,_snake_case ,_snake_case ,**_snake_case ).prev_sample # decode the image latents with the VQVAE UpperCAmelCase_ : List[Any] = self.vqvae.decode(_snake_case ).sample UpperCAmelCase_ : Tuple = torch.clamp(_snake_case ,-1.0 ,1.0 ) UpperCAmelCase_ : Any = image / 2 + 0.5 UpperCAmelCase_ : str = image.cpu().permute(0 ,2 ,3 ,1 ).numpy() if output_type == "pil": UpperCAmelCase_ : Optional[Any] = self.numpy_to_pil(_snake_case ) if not return_dict: return (image,) return ImagePipelineOutput(images=_snake_case )
71
from __future__ import annotations from typing import Any def lowerCAmelCase_ ( __lowerCamelCase ): create_state_space_tree(__lowerCamelCase , [] , 0 ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): if index == len(__lowerCamelCase ): print(__lowerCamelCase ) return create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.pop() if __name__ == "__main__": _snake_case : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["A", "B", "C"]) generate_all_subsequences(seq)
81
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _UpperCAmelCase : List[Any] = { '''configuration_clip''': [ '''CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''CLIPConfig''', '''CLIPOnnxConfig''', '''CLIPTextConfig''', '''CLIPVisionConfig''', ], '''processing_clip''': ['''CLIPProcessor'''], '''tokenization_clip''': ['''CLIPTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : int = ['''CLIPTokenizerFast'''] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Tuple = ['''CLIPFeatureExtractor'''] _UpperCAmelCase : List[str] = ['''CLIPImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Dict = [ '''CLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''CLIPModel''', '''CLIPPreTrainedModel''', '''CLIPTextModel''', '''CLIPTextModelWithProjection''', '''CLIPVisionModel''', '''CLIPVisionModelWithProjection''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[str] = [ '''TF_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFCLIPModel''', '''TFCLIPPreTrainedModel''', '''TFCLIPTextModel''', '''TFCLIPVisionModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[int] = [ '''FlaxCLIPModel''', '''FlaxCLIPPreTrainedModel''', '''FlaxCLIPTextModel''', '''FlaxCLIPTextPreTrainedModel''', '''FlaxCLIPVisionModel''', '''FlaxCLIPVisionPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_clip import ( CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, CLIPConfig, CLIPOnnxConfig, CLIPTextConfig, CLIPVisionConfig, ) from .processing_clip import CLIPProcessor from .tokenization_clip import CLIPTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_clip_fast import CLIPTokenizerFast try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_clip import CLIPFeatureExtractor from .image_processing_clip import CLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_clip import ( CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, CLIPModel, CLIPPreTrainedModel, CLIPTextModel, CLIPTextModelWithProjection, CLIPVisionModel, CLIPVisionModelWithProjection, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_clip import ( TF_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFCLIPModel, TFCLIPPreTrainedModel, TFCLIPTextModel, TFCLIPVisionModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_clip import ( FlaxCLIPModel, FlaxCLIPPreTrainedModel, FlaxCLIPTextModel, FlaxCLIPTextPreTrainedModel, FlaxCLIPVisionModel, FlaxCLIPVisionPreTrainedModel, ) else: import sys _UpperCAmelCase : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
72
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case , __snake_case : List[Any] = image.size __snake_case , __snake_case : Tuple = (x - x % 3_2 for x in (w, h)) # resize to integer multiple of 32 __snake_case : str = image.resize((w, h) , resample=PIL_INTERPOLATION["lanczos"] ) __snake_case : int = np.array(__lowerCamelCase ).astype(np.floataa ) / 2_5_5.0 __snake_case : Union[str, Any] = image[None].transpose(0 , 3 , 1 , 2 ) __snake_case : Union[str, Any] = torch.from_numpy(__lowerCamelCase ) return 2.0 * image - 1.0 class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : VQModel , lowerCamelCase : UNetaDModel , lowerCamelCase : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ) -> Union[str, Any]: super().__init__() self.register_modules(vqvae=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase ) @torch.no_grad() def __call__( self : List[str] , lowerCamelCase : Union[torch.Tensor, PIL.Image.Image] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : Optional[int] = 100 , lowerCamelCase : Optional[float] = 0.0 , lowerCamelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , ) -> Union[Tuple, ImagePipelineOutput]: if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : Any = 1 elif isinstance(lowerCamelCase , torch.Tensor ): __snake_case : Any = image.shape[0] else: raise ValueError(F'`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(lowerCamelCase )}' ) if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : List[Any] = preprocess(lowerCamelCase ) __snake_case , __snake_case : int = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image __snake_case : str = (batch_size, self.unet.config.in_channels // 2, height, width) __snake_case : str = next(self.unet.parameters() ).dtype __snake_case : Tuple = randn_tensor(lowerCamelCase , generator=lowerCamelCase , device=self.device , dtype=lowerCamelCase ) __snake_case : List[Any] = image.to(device=self.device , dtype=lowerCamelCase ) # set timesteps and move to the correct device self.scheduler.set_timesteps(lowerCamelCase , device=self.device ) __snake_case : str = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler __snake_case : Dict = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] __snake_case : Union[str, Any] = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __snake_case : int = {} if accepts_eta: __snake_case : List[str] = eta for t in self.progress_bar(lowerCamelCase ): # concat latents and low resolution image in the channel dimension. __snake_case : Union[str, Any] = torch.cat([latents, image] , dim=1 ) __snake_case : Optional[Any] = self.scheduler.scale_model_input(lowerCamelCase , lowerCamelCase ) # predict the noise residual __snake_case : int = self.unet(lowerCamelCase , lowerCamelCase ).sample # compute the previous noisy sample x_t -> x_t-1 __snake_case : Union[str, Any] = self.scheduler.step(lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase ).prev_sample # decode the image latents with the VQVAE __snake_case : List[Any] = self.vqvae.decode(lowerCamelCase ).sample __snake_case : Dict = torch.clamp(lowerCamelCase , -1.0 , 1.0 ) __snake_case : Any = image / 2 + 0.5 __snake_case : Tuple = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __snake_case : Tuple = self.numpy_to_pil(lowerCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCamelCase )
81
0
a_ : dict[tuple[int, int, int], int] = {} def lowerCamelCase__ (_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase): # if we are absent twice, or late 3 consecutive days, # no further prize strings are possible if late == 3 or absent == 2: return 0 # if we have no days left, and have not failed any other rules, # we have a prize string if days == 0: return 1 # No easy solution, so now we need to do the recursive calculation # First, check if the combination is already in the cache, and # if yes, return the stored value from there since we already # know the number of possible prize strings from this point on SCREAMING_SNAKE_CASE = (days, absent, late) if key in cache: return cache[key] # now we calculate the three possible ways that can unfold from # this point on, depending on our attendance today # 1) if we are late (but not absent), the "absent" counter stays as # it is, but the "late" counter increases by one SCREAMING_SNAKE_CASE = _calculate(days - 1 , _UpperCAmelCase , late + 1) # 2) if we are absent, the "absent" counter increases by 1, and the # "late" counter resets to 0 SCREAMING_SNAKE_CASE = _calculate(days - 1 , absent + 1 , 0) # 3) if we are on time, this resets the "late" counter and keeps the # absent counter SCREAMING_SNAKE_CASE = _calculate(days - 1 , _UpperCAmelCase , 0) SCREAMING_SNAKE_CASE = state_late + state_absent + state_ontime SCREAMING_SNAKE_CASE = prizestrings return prizestrings def lowerCamelCase__ (_UpperCAmelCase = 30): return _calculate(_UpperCAmelCase , absent=0 , late=0) if __name__ == "__main__": print(solution())
73
import gc import unittest import torch from parameterized import parameterized from diffusers import AutoencoderKL from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = AutoencoderKL __UpperCAmelCase : Optional[Any] = "sample" __UpperCAmelCase : Optional[int] = 1e-2 @property def __snake_case ( self : Dict ) -> Optional[Any]: __snake_case : Optional[Any] = 4 __snake_case : Tuple = 3 __snake_case : List[str] = (32, 32) __snake_case : str = floats_tensor((batch_size, num_channels) + sizes ).to(lowerCamelCase ) return {"sample": image} @property def __snake_case ( self : Union[str, Any] ) -> Tuple: return (3, 32, 32) @property def __snake_case ( self : int ) -> int: return (3, 32, 32) def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Optional[Any] = { "block_out_channels": [32, 64], "in_channels": 3, "out_channels": 3, "down_block_types": ["DownEncoderBlock2D", "DownEncoderBlock2D"], "up_block_types": ["UpDecoderBlock2D", "UpDecoderBlock2D"], "latent_channels": 4, } __snake_case : Any = self.dummy_input return init_dict, inputs_dict def __snake_case ( self : str ) -> Dict: pass def __snake_case ( self : Tuple ) -> List[str]: pass @unittest.skipIf(torch_device == "mps" , "Gradient checkpointing skipped on MPS" ) def __snake_case ( self : Any ) -> Optional[Any]: # enable deterministic behavior for gradient checkpointing __snake_case , __snake_case : int = self.prepare_init_args_and_inputs_for_common() __snake_case : str = self.model_class(**lowerCamelCase ) model.to(lowerCamelCase ) assert not model.is_gradient_checkpointing and model.training __snake_case : str = model(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model.zero_grad() __snake_case : Any = torch.randn_like(lowerCamelCase ) __snake_case : str = (out - labels).mean() loss.backward() # re-instantiate the model now enabling gradient checkpointing __snake_case : Optional[int] = self.model_class(**lowerCamelCase ) # clone model model_a.load_state_dict(model.state_dict() ) model_a.to(lowerCamelCase ) model_a.enable_gradient_checkpointing() assert model_a.is_gradient_checkpointing and model_a.training __snake_case : int = model_a(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model_a.zero_grad() __snake_case : Union[str, Any] = (out_a - labels).mean() loss_a.backward() # compare the output and parameters gradients self.assertTrue((loss - loss_a).abs() < 1E-5 ) __snake_case : Optional[int] = dict(model.named_parameters() ) __snake_case : List[Any] = dict(model_a.named_parameters() ) for name, param in named_params.items(): self.assertTrue(torch_all_close(param.grad.data , named_params_a[name].grad.data , atol=5E-5 ) ) def __snake_case ( self : List[Any] ) -> Optional[int]: __snake_case , __snake_case : Optional[Any] = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" , output_loading_info=lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(lowerCamelCase ) __snake_case : Optional[Any] = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def __snake_case ( self : Optional[Any] ) -> Union[str, Any]: __snake_case : Tuple = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" ) __snake_case : Dict = model.to(lowerCamelCase ) model.eval() if torch_device == "mps": __snake_case : int = torch.manual_seed(0 ) else: __snake_case : str = torch.Generator(device=lowerCamelCase ).manual_seed(0 ) __snake_case : List[str] = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) __snake_case : Union[str, Any] = image.to(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , sample_posterior=lowerCamelCase , generator=lowerCamelCase ).sample __snake_case : List[Any] = output[0, -1, -3:, -3:].flatten().cpu() # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. if torch_device == "mps": __snake_case : Union[str, Any] = torch.tensor( [ -4.0078E-01, -3.8323E-04, -1.2681E-01, -1.1462E-01, 2.0095E-01, 1.0893E-01, -8.8247E-02, -3.0361E-01, -9.8644E-03, ] ) elif torch_device == "cpu": __snake_case : Tuple = torch.tensor( [-0.13_52, 0.08_78, 0.04_19, -0.08_18, -0.10_69, 0.06_88, -0.14_58, -0.44_46, -0.00_26] ) else: __snake_case : List[str] = torch.tensor( [-0.24_21, 0.46_42, 0.25_07, -0.04_38, 0.06_82, 0.31_60, -0.20_18, -0.07_27, 0.24_85] ) self.assertTrue(torch_all_close(lowerCamelCase , lowerCamelCase , rtol=1E-2 ) ) @slow class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : int , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] ) -> List[str]: return F'gaussian_noise_s={seed}_shape={"_".join([str(lowerCamelCase ) for s in shape] )}.npy' def __snake_case ( self : List[Any] ) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __snake_case ( self : Tuple , lowerCamelCase : List[Any]=0 , lowerCamelCase : Tuple=(4, 3, 512, 512) , lowerCamelCase : Optional[int]=False ) -> str: __snake_case : List[Any] = torch.floataa if fpaa else torch.floataa __snake_case : Tuple = torch.from_numpy(load_hf_numpy(self.get_file_format(lowerCamelCase , lowerCamelCase ) ) ).to(lowerCamelCase ).to(lowerCamelCase ) return image def __snake_case ( self : Optional[Any] , lowerCamelCase : int="CompVis/stable-diffusion-v1-4" , lowerCamelCase : int=False ) -> int: __snake_case : str = "fp16" if fpaa else None __snake_case : int = torch.floataa if fpaa else torch.floataa __snake_case : int = AutoencoderKL.from_pretrained( lowerCamelCase , subfolder="vae" , torch_dtype=lowerCamelCase , revision=lowerCamelCase , ) model.to(lowerCamelCase ).eval() return model def __snake_case ( self : str , lowerCamelCase : int=0 ) -> Optional[Any]: if torch_device == "mps": return torch.manual_seed(lowerCamelCase ) return torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) @parameterized.expand( [ # fmt: off [33, [-0.16_03, 0.98_78, -0.04_95, -0.07_90, -0.27_09, 0.83_75, -0.20_60, -0.08_24], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_76, 0.11_68, 0.13_32, -0.48_40, -0.25_08, -0.07_91, -0.04_93, -0.40_89], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[str] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] ) -> List[Any]: __snake_case : Optional[Any] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) __snake_case : Tuple = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : List[Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : int = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [33, [-0.05_13, 0.02_89, 1.37_99, 0.21_66, -0.25_73, -0.08_71, 0.51_03, -0.09_99]], [47, [-0.41_28, -0.13_20, -0.37_04, 0.19_65, -0.41_16, -0.23_32, -0.33_40, 0.22_47]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : Any , lowerCamelCase : List[str] , lowerCamelCase : List[str] ) -> Tuple: __snake_case : Any = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Any = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.16_09, 0.98_66, -0.04_87, -0.07_77, -0.27_16, 0.83_68, -0.20_55, -0.08_14], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_77, 0.11_47, 0.13_33, -0.48_41, -0.25_06, -0.08_05, -0.04_91, -0.40_85], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Any , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) with torch.no_grad(): __snake_case : int = model(lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Union[str, Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : List[str] = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [13, [-0.20_51, -0.18_03, -0.23_11, -0.21_14, -0.32_92, -0.35_74, -0.29_53, -0.33_23]], [37, [-0.26_32, -0.26_25, -0.21_99, -0.27_41, -0.45_39, -0.49_90, -0.37_20, -0.49_25]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : List[str] , lowerCamelCase : Tuple , lowerCamelCase : Any ) -> Optional[Any]: __snake_case : List[str] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : str = sample[-1, -2:, :2, -2:].flatten().cpu() __snake_case : Optional[int] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-3 ) @parameterized.expand( [ # fmt: off [27, [-0.03_69, 0.02_07, -0.07_76, -0.06_82, -0.17_47, -0.19_30, -0.14_65, -0.20_39]], [16, [-0.16_28, -0.21_34, -0.27_47, -0.26_42, -0.37_74, -0.44_04, -0.36_87, -0.42_77]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : str , lowerCamelCase : Optional[int] , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[str] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : Union[str, Any] = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Optional[Any] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=5E-3 ) @parameterized.expand([(13,), (16,), (27,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> Tuple: __snake_case : Dict = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : Any = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Any = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-1 ) @parameterized.expand([(13,), (16,), (37,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : List[Any] , lowerCamelCase : Any ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : Union[str, Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : List[Any] = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Dict = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.30_01, 0.09_18, -2.69_84, -3.97_20, -3.20_99, -5.03_53, 1.73_38, -0.20_65, 3.42_67]], [47, [-1.50_30, -4.38_71, -6.03_55, -9.11_57, -1.66_61, -2.78_53, 2.16_07, -5.08_23, 2.56_33]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Dict ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : int = self.get_sd_image(lowerCamelCase ) __snake_case : int = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model.encode(lowerCamelCase ).latent_dist __snake_case : Dict = dist.sample(generator=lowerCamelCase ) assert list(sample.shape ) == [image.shape[0], 4] + [i // 8 for i in image.shape[2:]] __snake_case : List[str] = sample[0, -1, -3:, -3:].flatten().cpu() __snake_case : Dict = torch.tensor(lowerCamelCase ) __snake_case : Dict = 3E-3 if torch_device != "mps" else 1E-2 assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=lowerCamelCase )
81
0
# # This a `torch.distributed` diagnostics script that checks that all GPUs in the cluster (one or # many nodes) can talk to each other via nccl and allocate gpu memory. # # To run first adjust the number of processes and nodes: # # python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # You may need to add --master_addr $MASTER_ADDR --master_port $MASTER_PORT if using a custom addr:port # # You can also use the rdzv API: --rdzv_endpoint $MASTER_ADDR:$MASTER_PORT --rdzv_backend c10d # # use torch.distributed.launch instead of torch.distributed.run for torch < 1.9 # # If you get a hanging in `barrier` calls you have some network issues, you may try to debug this with: # # NCCL_DEBUG=INFO python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # which should tell you what's going on behind the scenes. # # # This script can be run via `srun` in the SLURM environment as well. Here is a SLURM script that # runs on 2 nodes of 4 gpus per node: # # #SBATCH --job-name=test-nodes # name # #SBATCH --nodes=2 # nodes # #SBATCH --ntasks-per-node=1 # crucial - only 1 task per dist per node! # #SBATCH --cpus-per-task=10 # number of cores per tasks # #SBATCH --gres=gpu:4 # number of gpus # #SBATCH --time 0:05:00 # maximum execution time (HH:MM:SS) # #SBATCH --output=%x-%j.out # output file name # # GPUS_PER_NODE=4 # MASTER_ADDR=$(scontrol show hostnames $SLURM_JOB_NODELIST | head -n 1) # MASTER_PORT=6000 # # srun --jobid $SLURM_JOBID bash -c 'python -m torch.distributed.run \ # --nproc_per_node $GPUS_PER_NODE --nnodes $SLURM_NNODES --node_rank $SLURM_PROCID \ # --master_addr $MASTER_ADDR --master_port $MASTER_PORT \ # torch-distributed-gpu-test.py' # import fcntl import os import socket import torch import torch.distributed as dist def a__ ( *snake_case ): """simple docstring""" with open(snake_case , '''r''' ) as fh: fcntl.flock(snake_case , fcntl.LOCK_EX ) try: print(*snake_case ) finally: fcntl.flock(snake_case , fcntl.LOCK_UN ) lowercase_ = int(os.environ["""LOCAL_RANK"""]) torch.cuda.set_device(local_rank) lowercase_ = torch.device("""cuda""", local_rank) lowercase_ = socket.gethostname() lowercase_ = f'''[{hostname}-{local_rank}]''' try: # test distributed dist.init_process_group("""nccl""") dist.all_reduce(torch.ones(1).to(device), op=dist.ReduceOp.SUM) dist.barrier() # test cuda is available and can allocate memory torch.cuda.is_available() torch.ones(1).cuda(local_rank) # global rank lowercase_ = dist.get_rank() lowercase_ = dist.get_world_size() printflock(f'''{gpu} is OK (global rank: {rank}/{world_size})''') dist.barrier() if rank == 0: printflock(f'''pt={torch.__version__}, cuda={torch.version.cuda}, nccl={torch.cuda.nccl.version()}''') except Exception: printflock(f'''{gpu} is broken''') raise
74
import warnings from ...utils import logging from .image_processing_flava import FlavaImageProcessor _snake_case : Optional[int] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[Any] , *lowerCamelCase : Any , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class FlavaFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use FlavaImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
'''simple docstring''' from __future__ import annotations import numpy as np def a__ ( lowerCAmelCase__ ) -> List[Any]: return np.maximum(0 , lowerCAmelCase__ ) if __name__ == "__main__": print(np.array(relu([-1, 0, 5]))) # --> [0, 0, 5]
75
import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : Tuple ) -> Optional[Any]: __snake_case : Dict = tempfile.mkdtemp() __snake_case : Any = SamImageProcessor() __snake_case : Optional[int] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : Optional[Any] , **lowerCamelCase : Optional[int] ) -> Optional[Any]: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[Any] ) -> Dict: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : int ) -> List[Any]: __snake_case : List[str] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : int = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : List[Any] ) -> Dict: __snake_case : int = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Dict = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Optional[Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : List[str] ) -> Tuple: __snake_case : int = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Optional[int] = self.prepare_image_inputs() __snake_case : List[str] = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : Dict = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_torch def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Tuple = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[str] = [torch.ones((1, 3, 5, 5) )] __snake_case : Tuple = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : int = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , torch.tensor(lowerCamelCase ) , torch.tensor(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : List[str] = [np.ones((1, 3, 5, 5) )] __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : str = [[1, 0], [0, 1]] with self.assertRaises(lowerCamelCase ): __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) @require_vision @require_tf class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[Any] ) -> Union[str, Any]: __snake_case : int = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : str , **lowerCamelCase : Any ) -> Tuple: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> Any: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : str ) -> List[Any]: __snake_case : int = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : Dict = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : int ) -> List[str]: __snake_case : List[Any] = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Optional[Any] = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Tuple = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : Union[str, Any] ) -> List[Any]: __snake_case : str = self.get_image_processor() __snake_case : Union[str, Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : int = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : List[str] = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_tf def __snake_case ( self : Any ) -> Optional[int]: __snake_case : List[str] = self.get_image_processor() __snake_case : Dict = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Union[str, Any] = [tf.ones((1, 3, 5, 5) )] __snake_case : List[Any] = [[1764, 2646]] __snake_case : Dict = [[683, 1024]] __snake_case : List[str] = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , tf.convert_to_tensor(lowerCamelCase ) , tf.convert_to_tensor(lowerCamelCase ) , return_tensors="tf" , ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : Union[str, Any] = [np.ones((1, 3, 5, 5) )] __snake_case : List[str] = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Tuple = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): __snake_case : Dict = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[str] ) -> str: __snake_case : Optional[int] = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : List[str] , **lowerCamelCase : Any ) -> Dict: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> List[Any]: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : str = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : List[Any] = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : str = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[Any] = np.random.randint(0 , 2 , size=(1, 3, 5, 5) ).astype(np.floataa ) __snake_case : Dict = [tf.convert_to_tensor(lowerCamelCase )] __snake_case : List[Any] = [torch.tensor(lowerCamelCase )] __snake_case : Optional[Any] = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : Union[str, Any] = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) __snake_case : Dict = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="pt" ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def __snake_case ( self : List[Any] ) -> List[str]: __snake_case : Any = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : Any = image_processor(lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Optional[Any] = processor(images=lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Tuple = image_processor(lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() __snake_case : List[Any] = processor(images=lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) )
81
0
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_convbert import ConvBertTokenizer a_ = logging.get_logger(__name__) a_ = {'vocab_file': 'vocab.txt'} a_ = { 'vocab_file': { 'YituTech/conv-bert-base': 'https://huggingface.co/YituTech/conv-bert-base/resolve/main/vocab.txt', 'YituTech/conv-bert-medium-small': ( 'https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/vocab.txt' ), 'YituTech/conv-bert-small': 'https://huggingface.co/YituTech/conv-bert-small/resolve/main/vocab.txt', } } a_ = { 'YituTech/conv-bert-base': 5_1_2, 'YituTech/conv-bert-medium-small': 5_1_2, 'YituTech/conv-bert-small': 5_1_2, } a_ = { 'YituTech/conv-bert-base': {'do_lower_case': True}, 'YituTech/conv-bert-medium-small': {'do_lower_case': True}, 'YituTech/conv-bert-small': {'do_lower_case': True}, } class UpperCAmelCase_ ( snake_case ): UpperCamelCase =VOCAB_FILES_NAMES UpperCamelCase =PRETRAINED_VOCAB_FILES_MAP UpperCamelCase =PRETRAINED_INIT_CONFIGURATION UpperCamelCase =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase =ConvBertTokenizer def __init__( self , UpperCamelCase_=None , UpperCamelCase_=None , UpperCamelCase_=True , UpperCamelCase_="[UNK]" , UpperCamelCase_="[SEP]" , UpperCamelCase_="[PAD]" , UpperCamelCase_="[CLS]" , UpperCamelCase_="[MASK]" , UpperCamelCase_=True , UpperCamelCase_=None , **UpperCamelCase_ , ) -> List[Any]: super().__init__( UpperCamelCase_ , tokenizer_file=UpperCamelCase_ , do_lower_case=UpperCamelCase_ , unk_token=UpperCamelCase_ , sep_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , cls_token=UpperCamelCase_ , mask_token=UpperCamelCase_ , tokenize_chinese_chars=UpperCamelCase_ , strip_accents=UpperCamelCase_ , **UpperCamelCase_ , ) __lowercase : Dict = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , UpperCamelCase_ ) != do_lower_case or normalizer_state.get('''strip_accents''' , UpperCamelCase_ ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , UpperCamelCase_ ) != tokenize_chinese_chars ): __lowercase : Dict = getattr(UpperCamelCase_ , normalizer_state.pop('''type''' ) ) __lowercase : List[str] = do_lower_case __lowercase : List[str] = strip_accents __lowercase : Tuple = tokenize_chinese_chars __lowercase : Optional[int] = normalizer_class(**UpperCamelCase_ ) __lowercase : Union[str, Any] = do_lower_case def _lowerCamelCase ( self , UpperCamelCase_ , UpperCamelCase_=None ) -> List[str]: __lowercase : Union[str, Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _lowerCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ = None ) -> List[int]: __lowercase : List[Any] = [self.sep_token_id] __lowercase : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowerCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ = None ) -> Tuple[str]: __lowercase : str = self._tokenizer.model.save(UpperCamelCase_ , name=UpperCamelCase_ ) return tuple(UpperCamelCase_ )
76
import warnings from ...utils import logging from .image_processing_deit import DeiTImageProcessor _snake_case : List[str] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[str] , *lowerCamelCase : str , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class DeiTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use DeiTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A = logging.get_logger(__name__) A = { """distilbert-base-uncased""": """https://huggingface.co/distilbert-base-uncased/resolve/main/config.json""", """distilbert-base-uncased-distilled-squad""": ( """https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json""" ), """distilbert-base-cased""": """https://huggingface.co/distilbert-base-cased/resolve/main/config.json""", """distilbert-base-cased-distilled-squad""": ( """https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json""" ), """distilbert-base-german-cased""": """https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json""", """distilbert-base-multilingual-cased""": ( """https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json""" ), """distilbert-base-uncased-finetuned-sst-2-english""": ( """https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json""" ), } class a__ ( __magic_name__ ): lowercase_ = "distilbert" lowercase_ = { "hidden_size": "dim", "num_attention_heads": "n_heads", "num_hidden_layers": "n_layers", } def __init__( self : Optional[int] , UpperCamelCase_ : str=30522 , UpperCamelCase_ : Tuple=512 , UpperCamelCase_ : Optional[int]=False , UpperCamelCase_ : List[str]=6 , UpperCamelCase_ : str=12 , UpperCamelCase_ : List[Any]=768 , UpperCamelCase_ : Optional[Any]=4 * 768 , UpperCamelCase_ : int=0.1 , UpperCamelCase_ : Any=0.1 , UpperCamelCase_ : Tuple="gelu" , UpperCamelCase_ : Any=0.02 , UpperCamelCase_ : Optional[int]=0.1 , UpperCamelCase_ : Tuple=0.2 , UpperCamelCase_ : Optional[int]=0 , **UpperCamelCase_ : List[str] , ): """simple docstring""" __UpperCAmelCase : int = vocab_size __UpperCAmelCase : Tuple = max_position_embeddings __UpperCAmelCase : str = sinusoidal_pos_embds __UpperCAmelCase : Union[str, Any] = n_layers __UpperCAmelCase : str = n_heads __UpperCAmelCase : Union[str, Any] = dim __UpperCAmelCase : str = hidden_dim __UpperCAmelCase : Optional[int] = dropout __UpperCAmelCase : Optional[int] = attention_dropout __UpperCAmelCase : Optional[int] = activation __UpperCAmelCase : int = initializer_range __UpperCAmelCase : int = qa_dropout __UpperCAmelCase : Optional[Any] = seq_classif_dropout super().__init__(**UpperCamelCase_ , pad_token_id=UpperCamelCase_) class a__ ( __magic_name__ ): @property def a_ ( self : Union[str, Any]): """simple docstring""" if self.task == "multiple-choice": __UpperCAmelCase : Dict = {0: "batch", 1: "choice", 2: "sequence"} else: __UpperCAmelCase : Optional[int] = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ])
77
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case : Union[str, Any] = { "configuration_owlvit": [ "OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "OwlViTConfig", "OwlViTOnnxConfig", "OwlViTTextConfig", "OwlViTVisionConfig", ], "processing_owlvit": ["OwlViTProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[Any] = ["OwlViTFeatureExtractor"] _snake_case : Optional[int] = ["OwlViTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "OwlViTModel", "OwlViTPreTrainedModel", "OwlViTTextModel", "OwlViTVisionModel", "OwlViTForObjectDetection", ] if TYPE_CHECKING: from .configuration_owlvit import ( OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, OwlViTConfig, OwlViTOnnxConfig, OwlViTTextConfig, OwlViTVisionConfig, ) from .processing_owlvit import OwlViTProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_owlvit import OwlViTFeatureExtractor from .image_processing_owlvit import OwlViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_owlvit import ( OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST, OwlViTForObjectDetection, OwlViTModel, OwlViTPreTrainedModel, OwlViTTextModel, OwlViTVisionModel, ) else: import sys _snake_case : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
0
'''simple docstring''' from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import tensorflow as tf from transformers import AutoTokenizer, TFAutoModelForSeqaSeqLM @require_tf @require_sentencepiece @require_tokenizers class __A ( unittest.TestCase ): @slow def _lowercase (self : List[Any] ): UpperCAmelCase_ = TFAutoModelForSeqaSeqLM.from_pretrained("google/mt5-small" ) UpperCAmelCase_ = AutoTokenizer.from_pretrained("google/mt5-small" ) UpperCAmelCase_ = tokenizer("Hello there" , return_tensors="tf" ).input_ids UpperCAmelCase_ = tokenizer("Hi I am" , return_tensors="tf" ).input_ids UpperCAmelCase_ = model(__a , labels=__a ).loss UpperCAmelCase_ = -tf.math.reduce_mean(__a ).numpy() UpperCAmelCase_ = -21.22_81_68 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 2E-4 )
78
from ....configuration_utils import PretrainedConfig from ....utils import logging _snake_case : int = logging.get_logger(__name__) _snake_case : str = { "Visual-Attention-Network/van-base": ( "https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json" ), } class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[Any] = "van" def __init__( self : Optional[int] , lowerCamelCase : Any=224 , lowerCamelCase : str=3 , lowerCamelCase : Any=[7, 3, 3, 3] , lowerCamelCase : Dict=[4, 2, 2, 2] , lowerCamelCase : List[Any]=[64, 128, 320, 512] , lowerCamelCase : str=[3, 3, 12, 3] , lowerCamelCase : Dict=[8, 8, 4, 4] , lowerCamelCase : Any="gelu" , lowerCamelCase : Optional[int]=0.02 , lowerCamelCase : Tuple=1E-6 , lowerCamelCase : Optional[int]=1E-2 , lowerCamelCase : int=0.0 , lowerCamelCase : Optional[Any]=0.0 , **lowerCamelCase : Optional[int] , ) -> int: super().__init__(**lowerCamelCase ) __snake_case : Union[str, Any] = image_size __snake_case : Any = num_channels __snake_case : Any = patch_sizes __snake_case : List[Any] = strides __snake_case : str = hidden_sizes __snake_case : Dict = depths __snake_case : Optional[int] = mlp_ratios __snake_case : Dict = hidden_act __snake_case : Union[str, Any] = initializer_range __snake_case : List[str] = layer_norm_eps __snake_case : Optional[int] = layer_scale_init_value __snake_case : List[Any] = drop_path_rate __snake_case : int = dropout_rate
81
0
import math class UpperCAmelCase_ : def __init__( self , _lowerCAmelCase=0 ): # a graph with Node 0,1,...,N-1 UpperCAmelCase__ : Union[str, Any] = n UpperCAmelCase__ : Union[str, Any] = [ [math.inf for j in range(0 , _lowerCAmelCase )] for i in range(0 , _lowerCAmelCase ) ] # adjacency matrix for weight UpperCAmelCase__ : Optional[Any] = [ [math.inf for j in range(0 , _lowerCAmelCase )] for i in range(0 , _lowerCAmelCase ) ] # dp[i][j] stores minimum distance from i to j def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ : Dict = w def __UpperCAmelCase ( self ): for k in range(0 , self.n ): for i in range(0 , self.n ): for j in range(0 , self.n ): UpperCAmelCase__ : List[str] = min(self.dp[i][j] , self.dp[i][k] + self.dp[k][j] ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase ): return self.dp[u][v] if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : str = Graph(5) graph.add_edge(0, 2, 9) graph.add_edge(0, 4, 10) graph.add_edge(1, 3, 5) graph.add_edge(2, 3, 7) graph.add_edge(3, 0, 10) graph.add_edge(3, 1, 2) graph.add_edge(3, 2, 1) graph.add_edge(3, 4, 6) graph.add_edge(4, 1, 3) graph.add_edge(4, 2, 4) graph.add_edge(4, 3, 9) graph.floyd_warshall() graph.show_min(1, 4) graph.show_min(0, 3)
79
import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer _snake_case : Union[str, Any] = logging.getLogger(__name__) def lowerCAmelCase_ ( ): __snake_case : int = argparse.ArgumentParser( description="Prepare TFRecord shards from pre-tokenized samples of the wikitext dataset." ) parser.add_argument( "--dataset_name" , type=__lowerCamelCase , default="wikitext" , help="Name of the training. Explore datasets at: hf.co/datasets." , ) parser.add_argument( "--dataset_config" , type=__lowerCamelCase , default="wikitext-103-raw-v1" , help="Configuration name of the dataset." ) parser.add_argument( "--tokenizer_name_or_path" , type=__lowerCamelCase , default="sayakpaul/unigram-tokenizer-wikitext" , help="Tokenizer identifier. Can be a local filepath or a Hub identifier." , ) parser.add_argument( "--shard_size" , type=__lowerCamelCase , default=1_0_0_0 , help="Number of entries to go in a single shard." , ) parser.add_argument("--split" , type=__lowerCamelCase , default="train" , choices=["train", "test", "validation"] ) parser.add_argument( "--limit" , default=__lowerCamelCase , type=__lowerCamelCase , help="Limit the number of shards (used for debugging)." , ) parser.add_argument( "--max_length" , type=__lowerCamelCase , default=5_1_2 , help="Maximum sequence length. For training on TPUs, it helps to have a maximum" " sequence length that is a multiple of 8." , ) parser.add_argument( "--output_dir" , default="tf-tpu" , type=__lowerCamelCase , help="Output directory where the TFRecord shards will be saved. If the" " path is appended with `gs://` ('gs://tf-tpu', for example) then the TFRecord" " shards will be directly saved to a Google Cloud Storage bucket." , ) __snake_case : List[str] = parser.parse_args() return args def lowerCAmelCase_ ( __lowerCamelCase ): def fn(__lowerCamelCase ): return tokenizer(examples["text"] ) return fn def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = [] for i in range(len(tokenized_data["input_ids"] ) ): __snake_case : Tuple = { "input_ids": tf.train.Feature(intaa_list=tf.train.IntaaList(value=tokenized_data["input_ids"][i] ) ), "attention_mask": tf.train.Feature( intaa_list=tf.train.IntaaList(value=tokenized_data["attention_mask"][i] ) ), } __snake_case : List[Any] = tf.train.Features(feature=__lowerCamelCase ) __snake_case : str = tf.train.Example(features=__lowerCamelCase ) __snake_case : List[str] = example.SerializeToString() records.append(__lowerCamelCase ) return records def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Optional[int] = datasets.load_dataset(args.dataset_name , args.dataset_config , split=args.split ) if args.limit is not None: __snake_case : Optional[Any] = min(len(__lowerCamelCase ) , args.limit ) __snake_case : Dict = dataset.select(range(__lowerCamelCase ) ) print(F'Limiting the dataset to {args.limit} entries.' ) __snake_case : Dict = AutoTokenizer.from_pretrained(args.tokenizer_name_or_path ) # Handle output directory creation. # For serializing into a Google Cloud Storage Bucket, one needs to first # create a bucket. if "gs" not in args.output_dir: if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) __snake_case : Dict = os.path.join(args.output_dir , args.split ) if not os.path.exists(__lowerCamelCase ): os.makedirs(__lowerCamelCase ) else: __snake_case : str = os.path.join(args.output_dir , args.split ) # Tokenize the whole dataset at once. __snake_case : Any = tokenize_function(__lowerCamelCase ) __snake_case : Optional[Any] = dataset.map(__lowerCamelCase , batched=__lowerCamelCase , num_proc=4 , remove_columns=["text"] ) # We need to concatenate all our texts together, and then split the result # into chunks of a fixed size, which we will call block_size. To do this, we # will use the map method again, with the option batched=True. When we use batched=True, # the function we pass to map() will be passed multiple inputs at once, allowing us # to group them into more or fewer examples than we had in the input. # This allows us to create our new fixed-length samples. The advantage of this # method is that we don't lose a whole lot of content from the dataset compared to the # case where we simply tokenize with a pre-defined max_length. def group_texts(__lowerCamelCase ): # Concatenate all texts. __snake_case : List[str] = {k: sum(examples[k] , [] ) for k in examples.keys()} __snake_case : List[Any] = len(concatenated_examples[list(examples.keys() )[0]] ) # We drop the small remainder, though you could add padding instead if the model supports it # In this, as in all things, we advise you to follow your heart 🫀 __snake_case : Any = (total_length // args.max_length) * args.max_length # Split by chunks of max_len. __snake_case : int = { k: [t[i : i + args.max_length] for i in range(0 , __lowerCamelCase , args.max_length )] for k, t in concatenated_examples.items() } return result __snake_case : Any = dataset_tokenized.map(__lowerCamelCase , batched=__lowerCamelCase , batch_size=1_0_0_0 , num_proc=4 ) __snake_case : Optional[Any] = 0 __snake_case : Optional[Any] = 0 for shard in range(0 , len(__lowerCamelCase ) , args.shard_size ): __snake_case : List[str] = grouped_dataset[shard : shard + args.shard_size] __snake_case : Any = len(dataset_snapshot["input_ids"] ) __snake_case : List[Any] = os.path.join(__lowerCamelCase , F'dataset-{shard_count}-{records_containing}.tfrecord' ) __snake_case : Optional[Any] = get_serialized_examples(__lowerCamelCase ) with tf.io.TFRecordWriter(__lowerCamelCase ) as out_file: for i in range(len(__lowerCamelCase ) ): __snake_case : Union[str, Any] = serialized_examples[i] out_file.write(__lowerCamelCase ) print("Wrote file {} containing {} records".format(__lowerCamelCase , __lowerCamelCase ) ) shard_count += 1 total_records += records_containing with open(F'split-{args.split}-records-count.txt' , "w" ) as f: print(F'Total {args.split} records: {total_records}' , file=__lowerCamelCase ) if __name__ == "__main__": _snake_case : List[Any] = parse_args() main(args)
81
0
import unittest from diffusers import FlaxAutoencoderKL from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax from .test_modeling_common_flax import FlaxModelTesterMixin if is_flax_available(): import jax @require_flax class __UpperCamelCase ( _lowerCAmelCase , unittest.TestCase ): __snake_case :Any = FlaxAutoencoderKL @property def _a ( self : Tuple ) -> Optional[int]: """simple docstring""" __lowercase = 4 __lowercase = 3 __lowercase = (32, 32) __lowercase = jax.random.PRNGKey(0 ) __lowercase = jax.random.uniform(_lowerCAmelCase , ((batch_size, num_channels) + sizes) ) return {"sample": image, "prng_key": prng_key} def _a ( self : str ) -> Optional[int]: """simple docstring""" __lowercase = { """block_out_channels""": [32, 64], """in_channels""": 3, """out_channels""": 3, """down_block_types""": ["""DownEncoderBlock2D""", """DownEncoderBlock2D"""], """up_block_types""": ["""UpDecoderBlock2D""", """UpDecoderBlock2D"""], """latent_channels""": 4, } __lowercase = self.dummy_input return init_dict, inputs_dict
80
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) _snake_case : List[Any] = "\\n Text data.\n Second line of data." _snake_case : Tuple = "file" @pytest.fixture(scope="session" ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / (FILE_PATH + ".zstd") __snake_case : Optional[Any] = bytes(__lowerCamelCase , "utf-8" ) with zstd.open(__lowerCamelCase , "wb" ) as f: f.write(__lowerCamelCase ) return path @pytest.fixture def lowerCAmelCase_ ( __lowerCamelCase ): with open(os.path.join(tmpfs.local_root_dir , __lowerCamelCase ) , "w" ) as f: f.write(__lowerCamelCase ) return FILE_PATH @pytest.mark.parametrize("compression_format" , ["gzip", "xz", "zstd"] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Optional[int] = {"gzip": gz_file, "xz": xz_file, "zstd": zstd_path} __snake_case : str = input_paths[compression_format] __snake_case : Optional[Any] = tmp_path / "cache" __snake_case : Optional[int] = DownloadConfig(cache_dir=__lowerCamelCase , extract_compressed_file=__lowerCamelCase ) __snake_case : Union[str, Any] = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) with open(__lowerCamelCase ) as f: __snake_case : Dict = f.read() with open(__lowerCamelCase ) as f: __snake_case : Tuple = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("default_extracted" , [True, False] ) @pytest.mark.parametrize("default_cache_dir" , [True, False] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Tuple = "custom_cache" __snake_case : List[str] = "custom_extracted_dir" __snake_case : Any = tmp_path / "custom_extracted_path" if default_extracted: __snake_case : List[Any] = ("downloads" if default_cache_dir else custom_cache_dir, "extracted") else: monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_DIR" , __lowerCamelCase ) monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_PATH" , str(__lowerCamelCase ) ) __snake_case : Optional[Any] = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) __snake_case : Optional[int] = xz_file __snake_case : Optional[int] = ( DownloadConfig(extract_compressed_file=__lowerCamelCase ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=__lowerCamelCase ) ) __snake_case : str = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) assert Path(__lowerCamelCase ).parent.parts[-2:] == expected def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : Optional[Any] = str(Path(__lowerCamelCase ).resolve() ) assert cached_path(__lowerCamelCase ) == text_file # relative path __snake_case : Any = str(Path(__lowerCamelCase ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(__lowerCamelCase ) == text_file def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : List[Any] = str(tmp_path.resolve() / "__missing_file__.txt" ) with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) # relative path __snake_case : Optional[int] = "./__missing_file__.txt" with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : str = get_from_cache(F'tmp://{tmpfs_file}' ) with open(__lowerCamelCase ) as f: __snake_case : Union[str, Any] = f.read() assert output_file_content == FILE_CONTENT @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( ): with pytest.raises(__lowerCamelCase ): cached_path("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[Any] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): http_get("https://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): http_head("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[str] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): ftp_get("ftp://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): ftp_head("ftp://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): fsspec_get("s3://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): fsspec_head("s3://huggingface.co" )
81
0
"""simple docstring""" import importlib import inspect import json import os import re import shutil import sys from pathlib import Path from typing import Dict, Optional, Union from urllib import request from huggingface_hub import HfFolder, cached_download, hf_hub_download, model_info from packaging import version from .. import __version__ from . import DIFFUSERS_DYNAMIC_MODULE_NAME, HF_MODULES_CACHE, logging lowerCamelCase = ( """https://raw.githubusercontent.com/huggingface/diffusers/{revision}/examples/community/{pipeline}.py""" ) lowerCamelCase = logging.get_logger(__name__) # pylint: disable=invalid-name def a__ ( ): UpperCAmelCase_ = "https://pypi.org/pypi/diffusers/json" UpperCAmelCase_ = json.loads(request.urlopen(lowerCAmelCase__ ).read() )["releases"].keys() return sorted(lowerCAmelCase__ , key=lambda lowerCAmelCase__ : version.Version(lowerCAmelCase__ ) ) def a__ ( ): # This function has already been executed if HF_MODULES_CACHE already is in the Python path. if HF_MODULES_CACHE in sys.path: return sys.path.append(lowerCAmelCase__ ) os.makedirs(lowerCAmelCase__ , exist_ok=lowerCAmelCase__ ) UpperCAmelCase_ = Path(lowerCAmelCase__ ) / "__init__.py" if not init_path.exists(): init_path.touch() def a__ ( lowerCAmelCase__ ): init_hf_modules() UpperCAmelCase_ = Path(lowerCAmelCase__ ) / name # If the parent module does not exist yet, recursively create it. if not dynamic_module_path.parent.exists(): create_dynamic_module(dynamic_module_path.parent ) os.makedirs(lowerCAmelCase__ , exist_ok=lowerCAmelCase__ ) UpperCAmelCase_ = dynamic_module_path / "__init__.py" if not init_path.exists(): init_path.touch() def a__ ( lowerCAmelCase__ ): with open(lowerCAmelCase__ , "r" , encoding="utf-8" ) as f: UpperCAmelCase_ = f.read() # Imports of the form `import .xxx` UpperCAmelCase_ = re.findall("^\s*import\s+\.(\S+)\s*$" , lowerCAmelCase__ , flags=re.MULTILINE ) # Imports of the form `from .xxx import yyy` relative_imports += re.findall("^\s*from\s+\.(\S+)\s+import" , lowerCAmelCase__ , flags=re.MULTILINE ) # Unique-ify return list(set(lowerCAmelCase__ ) ) def a__ ( lowerCAmelCase__ ): UpperCAmelCase_ = False UpperCAmelCase_ = [module_file] UpperCAmelCase_ = [] # Let's recurse through all relative imports while not no_change: UpperCAmelCase_ = [] for f in files_to_check: new_imports.extend(get_relative_imports(lowerCAmelCase__ ) ) UpperCAmelCase_ = Path(lowerCAmelCase__ ).parent UpperCAmelCase_ = [str(module_path / m ) for m in new_imports] UpperCAmelCase_ = [f for f in new_import_files if f not in all_relative_imports] UpperCAmelCase_ = [f"""{f}.py""" for f in new_import_files] UpperCAmelCase_ = len(lowerCAmelCase__ ) == 0 all_relative_imports.extend(lowerCAmelCase__ ) return all_relative_imports def a__ ( lowerCAmelCase__ ): with open(lowerCAmelCase__ , "r" , encoding="utf-8" ) as f: UpperCAmelCase_ = f.read() # Imports of the form `import xxx` UpperCAmelCase_ = re.findall("^\s*import\s+(\S+)\s*$" , lowerCAmelCase__ , flags=re.MULTILINE ) # Imports of the form `from xxx import yyy` imports += re.findall("^\s*from\s+(\S+)\s+import" , lowerCAmelCase__ , flags=re.MULTILINE ) # Only keep the top-level module UpperCAmelCase_ = [imp.split("." )[0] for imp in imports if not imp.startswith("." )] # Unique-ify and test we got them all UpperCAmelCase_ = list(set(lowerCAmelCase__ ) ) UpperCAmelCase_ = [] for imp in imports: try: importlib.import_module(lowerCAmelCase__ ) except ImportError: missing_packages.append(lowerCAmelCase__ ) if len(lowerCAmelCase__ ) > 0: raise ImportError( "This modeling file requires the following packages that were not found in your environment: " f"""{', '.join(lowerCAmelCase__ )}. Run `pip install {' '.join(lowerCAmelCase__ )}`""" ) return get_relative_imports(lowerCAmelCase__ ) def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = module_path.replace(os.path.sep , "." ) UpperCAmelCase_ = importlib.import_module(lowerCAmelCase__ ) if class_name is None: return find_pipeline_class(lowerCAmelCase__ ) return getattr(lowerCAmelCase__ , lowerCAmelCase__ ) def a__ ( lowerCAmelCase__ ): from ..pipelines import DiffusionPipeline UpperCAmelCase_ = dict(inspect.getmembers(lowerCAmelCase__ , inspect.isclass ) ) UpperCAmelCase_ = None for cls_name, cls in cls_members.items(): if ( cls_name != DiffusionPipeline.__name__ and issubclass(cls , lowerCAmelCase__ ) and cls.__module__.split("." )[0] != "diffusers" ): if pipeline_class is not None: raise ValueError( f"""Multiple classes that inherit from {DiffusionPipeline.__name__} have been found:""" f""" {pipeline_class.__name__}, and {cls_name}. Please make sure to define only one in""" f""" {loaded_module}.""" ) UpperCAmelCase_ = cls return pipeline_class def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = False , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = False , ): UpperCAmelCase_ = str(lowerCAmelCase__ ) UpperCAmelCase_ = os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) if os.path.isfile(lowerCAmelCase__ ): UpperCAmelCase_ = module_file_or_url UpperCAmelCase_ = "local" elif pretrained_model_name_or_path.count("/" ) == 0: UpperCAmelCase_ = get_diffusers_versions() # cut ".dev0" UpperCAmelCase_ = "v" + ".".join(__version__.split("." )[:3] ) # retrieve github version that matches if revision is None: UpperCAmelCase_ = latest_version if latest_version[1:] in available_versions else "main" logger.info(f"""Defaulting to latest_version: {revision}.""" ) elif revision in available_versions: UpperCAmelCase_ = f"""v{revision}""" elif revision == "main": UpperCAmelCase_ = revision else: raise ValueError( f"""`custom_revision`: {revision} does not exist. Please make sure to choose one of""" f""" {', '.join(available_versions + ['main'] )}.""" ) # community pipeline on GitHub UpperCAmelCase_ = COMMUNITY_PIPELINES_URL.format(revision=lowerCAmelCase__ , pipeline=lowerCAmelCase__ ) try: UpperCAmelCase_ = cached_download( lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , proxies=lowerCAmelCase__ , resume_download=lowerCAmelCase__ , local_files_only=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , ) UpperCAmelCase_ = "git" UpperCAmelCase_ = pretrained_model_name_or_path + ".py" except EnvironmentError: logger.error(f"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise else: try: # Load from URL or cache if already cached UpperCAmelCase_ = hf_hub_download( lowerCAmelCase__ , lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , proxies=lowerCAmelCase__ , resume_download=lowerCAmelCase__ , local_files_only=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , ) UpperCAmelCase_ = os.path.join("local" , "--".join(pretrained_model_name_or_path.split("/" ) ) ) except EnvironmentError: logger.error(f"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise # Check we have all the requirements in our environment UpperCAmelCase_ = check_imports(lowerCAmelCase__ ) # Now we move the module inside our cached dynamic modules. UpperCAmelCase_ = DIFFUSERS_DYNAMIC_MODULE_NAME + os.path.sep + submodule create_dynamic_module(lowerCAmelCase__ ) UpperCAmelCase_ = Path(lowerCAmelCase__ ) / full_submodule if submodule == "local" or submodule == "git": # We always copy local files (we could hash the file to see if there was a change, and give them the name of # that hash, to only copy when there is a modification but it seems overkill for now). # The only reason we do the copy is to avoid putting too many folders in sys.path. shutil.copy(lowerCAmelCase__ , submodule_path / module_file ) for module_needed in modules_needed: UpperCAmelCase_ = f"""{module_needed}.py""" shutil.copy(os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) , submodule_path / module_needed ) else: # Get the commit hash # TODO: we will get this info in the etag soon, so retrieve it from there and not here. if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = use_auth_token elif use_auth_token is True: UpperCAmelCase_ = HfFolder.get_token() else: UpperCAmelCase_ = None UpperCAmelCase_ = model_info(lowerCAmelCase__ , revision=lowerCAmelCase__ , token=lowerCAmelCase__ ).sha # The module file will end up being placed in a subfolder with the git hash of the repo. This way we get the # benefit of versioning. UpperCAmelCase_ = submodule_path / commit_hash UpperCAmelCase_ = full_submodule + os.path.sep + commit_hash create_dynamic_module(lowerCAmelCase__ ) if not (submodule_path / module_file).exists(): shutil.copy(lowerCAmelCase__ , submodule_path / module_file ) # Make sure we also have every file with relative for module_needed in modules_needed: if not (submodule_path / module_needed).exists(): get_cached_module_file( lowerCAmelCase__ , f"""{module_needed}.py""" , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , resume_download=lowerCAmelCase__ , proxies=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , revision=lowerCAmelCase__ , local_files_only=lowerCAmelCase__ , ) return os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = False , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = False , **lowerCAmelCase__ , ): UpperCAmelCase_ = get_cached_module_file( lowerCAmelCase__ , lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , force_download=lowerCAmelCase__ , resume_download=lowerCAmelCase__ , proxies=lowerCAmelCase__ , use_auth_token=lowerCAmelCase__ , revision=lowerCAmelCase__ , local_files_only=lowerCAmelCase__ , ) return get_class_in_module(lowerCAmelCase__ , final_module.replace(".py" , "" ) )
82
_snake_case : Optional[int] = {"a": ["c", "b"], "b": ["d", "e"], "c": [], "d": [], "e": []} _snake_case : Dict = ["a", "b", "c", "d", "e"] def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : List[str] = start # add current to visited visited.append(__lowerCamelCase ) __snake_case : List[Any] = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: __snake_case : Tuple = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # if all neighbors visited add current to sort sort.append(__lowerCamelCase ) # if all vertices haven't been visited select a new one to visit if len(__lowerCamelCase ) != len(__lowerCamelCase ): for vertice in vertices: if vertice not in visited: __snake_case : int = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # return sort return sort if __name__ == "__main__": _snake_case : List[Any] = topological_sort("a", [], []) print(sort)
81
0
"""simple docstring""" import contextlib import copy import random from typing import Any, Dict, Iterable, Optional, Union import numpy as np import torch from .utils import deprecate, is_transformers_available if is_transformers_available(): import transformers def snake_case_ ( A_ : int ): '''simple docstring''' random.seed(A_ ) np.random.seed(A_ ) torch.manual_seed(A_ ) torch.cuda.manual_seed_all(A_ ) # ^^ safe to call this function even if cuda is not available class __snake_case : def __init__( self : int , __lowerCAmelCase : Iterable[torch.nn.Parameter] , __lowerCAmelCase : float = 0.99_99 , __lowerCAmelCase : float = 0.0 , __lowerCAmelCase : int = 0 , __lowerCAmelCase : bool = False , __lowerCAmelCase : Union[float, int] = 1.0 , __lowerCAmelCase : Union[float, int] = 2 / 3 , __lowerCAmelCase : Optional[Any] = None , __lowerCAmelCase : Dict[str, Any] = None , **__lowerCAmelCase : Optional[Any] , ): """simple docstring""" if isinstance(__lowerCAmelCase , torch.nn.Module ): _lowerCamelCase : Dict = ( '''Passing a `torch.nn.Module` to `ExponentialMovingAverage` is deprecated. ''' '''Please pass the parameters of the module instead.''' ) deprecate( '''passing a `torch.nn.Module` to `ExponentialMovingAverage`''' , '''1.0.0''' , __lowerCAmelCase , standard_warn=__lowerCAmelCase , ) _lowerCamelCase : int = parameters.parameters() # set use_ema_warmup to True if a torch.nn.Module is passed for backwards compatibility _lowerCamelCase : Optional[int] = True if kwargs.get('''max_value''' , __lowerCAmelCase ) is not None: _lowerCamelCase : str = '''The `max_value` argument is deprecated. Please use `decay` instead.''' deprecate('''max_value''' , '''1.0.0''' , __lowerCAmelCase , standard_warn=__lowerCAmelCase ) _lowerCamelCase : Any = kwargs['''max_value'''] if kwargs.get('''min_value''' , __lowerCAmelCase ) is not None: _lowerCamelCase : Optional[int] = '''The `min_value` argument is deprecated. Please use `min_decay` instead.''' deprecate('''min_value''' , '''1.0.0''' , __lowerCAmelCase , standard_warn=__lowerCAmelCase ) _lowerCamelCase : Optional[int] = kwargs['''min_value'''] _lowerCamelCase : int = list(__lowerCAmelCase ) _lowerCamelCase : int = [p.clone().detach() for p in parameters] if kwargs.get('''device''' , __lowerCAmelCase ) is not None: _lowerCamelCase : Tuple = '''The `device` argument is deprecated. Please use `to` instead.''' deprecate('''device''' , '''1.0.0''' , __lowerCAmelCase , standard_warn=__lowerCAmelCase ) self.to(device=kwargs['''device'''] ) _lowerCamelCase : Union[str, Any] = None _lowerCamelCase : Tuple = decay _lowerCamelCase : Any = min_decay _lowerCamelCase : str = update_after_step _lowerCamelCase : Any = use_ema_warmup _lowerCamelCase : Optional[Any] = inv_gamma _lowerCamelCase : Tuple = power _lowerCamelCase : Any = 0 _lowerCamelCase : Any = None # set in `step()` _lowerCamelCase : int = model_cls _lowerCamelCase : Dict = model_config @classmethod def SCREAMING_SNAKE_CASE ( cls : int , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : Optional[int] ): """simple docstring""" _lowerCamelCase , _lowerCamelCase : Optional[int] = model_cls.load_config(__lowerCAmelCase , return_unused_kwargs=__lowerCAmelCase ) _lowerCamelCase : int = model_cls.from_pretrained(__lowerCAmelCase ) _lowerCamelCase : Union[str, Any] = cls(model.parameters() , model_cls=__lowerCAmelCase , model_config=model.config ) ema_model.load_state_dict(__lowerCAmelCase ) return ema_model def SCREAMING_SNAKE_CASE ( self : Optional[int] , __lowerCAmelCase : Optional[int] ): """simple docstring""" if self.model_cls is None: raise ValueError('''`save_pretrained` can only be used if `model_cls` was defined at __init__.''' ) if self.model_config is None: raise ValueError('''`save_pretrained` can only be used if `model_config` was defined at __init__.''' ) _lowerCamelCase : Dict = self.model_cls.from_config(self.model_config ) _lowerCamelCase : int = self.state_dict() state_dict.pop('''shadow_params''' , __lowerCAmelCase ) model.register_to_config(**__lowerCAmelCase ) self.copy_to(model.parameters() ) model.save_pretrained(__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : str , __lowerCAmelCase : int ): """simple docstring""" _lowerCamelCase : List[str] = max(0 , optimization_step - self.update_after_step - 1 ) if step <= 0: return 0.0 if self.use_ema_warmup: _lowerCamelCase : List[Any] = 1 - (1 + step / self.inv_gamma) ** -self.power else: _lowerCamelCase : str = (1 + step) / (1_0 + step) _lowerCamelCase : List[Any] = min(__lowerCAmelCase , self.decay ) # make sure decay is not smaller than min_decay _lowerCamelCase : str = max(__lowerCAmelCase , self.min_decay ) return cur_decay_value @torch.no_grad() def SCREAMING_SNAKE_CASE ( self : List[str] , __lowerCAmelCase : Iterable[torch.nn.Parameter] ): """simple docstring""" if isinstance(__lowerCAmelCase , torch.nn.Module ): _lowerCamelCase : int = ( '''Passing a `torch.nn.Module` to `ExponentialMovingAverage.step` is deprecated. ''' '''Please pass the parameters of the module instead.''' ) deprecate( '''passing a `torch.nn.Module` to `ExponentialMovingAverage.step`''' , '''1.0.0''' , __lowerCAmelCase , standard_warn=__lowerCAmelCase , ) _lowerCamelCase : Dict = parameters.parameters() _lowerCamelCase : List[Any] = list(__lowerCAmelCase ) self.optimization_step += 1 # Compute the decay factor for the exponential moving average. _lowerCamelCase : Optional[int] = self.get_decay(self.optimization_step ) _lowerCamelCase : Optional[Any] = decay _lowerCamelCase : Union[str, Any] = 1 - decay _lowerCamelCase : Dict = contextlib.nullcontext if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): import deepspeed for s_param, param in zip(self.shadow_params , __lowerCAmelCase ): if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): _lowerCamelCase : Optional[int] = deepspeed.zero.GatheredParameters(__lowerCAmelCase , modifier_rank=__lowerCAmelCase ) with context_manager(): if param.requires_grad: s_param.sub_(one_minus_decay * (s_param - param) ) else: s_param.copy_(__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : str , __lowerCAmelCase : Iterable[torch.nn.Parameter] ): """simple docstring""" _lowerCamelCase : List[str] = list(__lowerCAmelCase ) for s_param, param in zip(self.shadow_params , __lowerCAmelCase ): param.data.copy_(s_param.to(param.device ).data ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] , __lowerCAmelCase : List[Any]=None , __lowerCAmelCase : List[str]=None ): """simple docstring""" _lowerCamelCase : Optional[Any] = [ p.to(device=__lowerCAmelCase , dtype=__lowerCAmelCase ) if p.is_floating_point() else p.to(device=__lowerCAmelCase ) for p in self.shadow_params ] def SCREAMING_SNAKE_CASE ( self : int ): """simple docstring""" return { "decay": self.decay, "min_decay": self.min_decay, "optimization_step": self.optimization_step, "update_after_step": self.update_after_step, "use_ema_warmup": self.use_ema_warmup, "inv_gamma": self.inv_gamma, "power": self.power, "shadow_params": self.shadow_params, } def SCREAMING_SNAKE_CASE ( self : Tuple , __lowerCAmelCase : Iterable[torch.nn.Parameter] ): """simple docstring""" _lowerCamelCase : Optional[Any] = [param.detach().cpu().clone() for param in parameters] def SCREAMING_SNAKE_CASE ( self : Optional[Any] , __lowerCAmelCase : Iterable[torch.nn.Parameter] ): """simple docstring""" if self.temp_stored_params is None: raise RuntimeError('''This ExponentialMovingAverage has no `store()`ed weights ''' '''to `restore()`''' ) for c_param, param in zip(self.temp_stored_params , __lowerCAmelCase ): param.data.copy_(c_param.data ) # Better memory-wise. _lowerCamelCase : str = None def SCREAMING_SNAKE_CASE ( self : List[str] , __lowerCAmelCase : dict ): """simple docstring""" _lowerCamelCase : Any = copy.deepcopy(__lowerCAmelCase ) _lowerCamelCase : str = state_dict.get('''decay''' , self.decay ) if self.decay < 0.0 or self.decay > 1.0: raise ValueError('''Decay must be between 0 and 1''' ) _lowerCamelCase : Any = state_dict.get('''min_decay''' , self.min_decay ) if not isinstance(self.min_decay , __lowerCAmelCase ): raise ValueError('''Invalid min_decay''' ) _lowerCamelCase : Union[str, Any] = state_dict.get('''optimization_step''' , self.optimization_step ) if not isinstance(self.optimization_step , __lowerCAmelCase ): raise ValueError('''Invalid optimization_step''' ) _lowerCamelCase : Optional[Any] = state_dict.get('''update_after_step''' , self.update_after_step ) if not isinstance(self.update_after_step , __lowerCAmelCase ): raise ValueError('''Invalid update_after_step''' ) _lowerCamelCase : Tuple = state_dict.get('''use_ema_warmup''' , self.use_ema_warmup ) if not isinstance(self.use_ema_warmup , __lowerCAmelCase ): raise ValueError('''Invalid use_ema_warmup''' ) _lowerCamelCase : int = state_dict.get('''inv_gamma''' , self.inv_gamma ) if not isinstance(self.inv_gamma , (float, int) ): raise ValueError('''Invalid inv_gamma''' ) _lowerCamelCase : Tuple = state_dict.get('''power''' , self.power ) if not isinstance(self.power , (float, int) ): raise ValueError('''Invalid power''' ) _lowerCamelCase : List[Any] = state_dict.get('''shadow_params''' , __lowerCAmelCase ) if shadow_params is not None: _lowerCamelCase : Optional[Any] = shadow_params if not isinstance(self.shadow_params , __lowerCAmelCase ): raise ValueError('''shadow_params must be a list''' ) if not all(isinstance(__lowerCAmelCase , torch.Tensor ) for p in self.shadow_params ): raise ValueError('''shadow_params must all be Tensors''' )
83
def lowerCAmelCase_ ( ): return [ a * b * (1_0_0_0 - a - b) for a in range(1 , 9_9_9 ) for b in range(__lowerCamelCase , 9_9_9 ) if (a * a + b * b == (1_0_0_0 - a - b) ** 2) ][0] if __name__ == "__main__": print(f'''{solution() = }''')
81
0
import argparse from pathlib import Path import torch from transformers import OPTConfig, OPTModel from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase = logging.get_logger(__name__) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): lowercase = torch.load(__SCREAMING_SNAKE_CASE , map_location='cpu' ) if "model" in sd.keys(): lowercase = torch.load(__SCREAMING_SNAKE_CASE , map_location='cpu' )['model'] # pop unnecessary weights lowercase = [ 'decoder.version', 'decoder.output_projection.weight', ] for key in keys_to_delete: if key in sd: sd.pop(__SCREAMING_SNAKE_CASE ) lowercase = { 'decoder.project_in_dim.weight': 'decoder.project_in.weight', 'decoder.project_out_dim.weight': 'decoder.project_out.weight', 'decoder.layer_norm.weight': 'decoder.final_layer_norm.weight', 'decoder.layer_norm.bias': 'decoder.final_layer_norm.bias', } for old_key, new_key in keys_to_rename.items(): if old_key in sd: lowercase = sd.pop(__SCREAMING_SNAKE_CASE ) lowercase = list(sd.keys() ) for key in keys: if ".qkv_proj." in key: lowercase = sd[key] # We split QKV in separate Q,K,V lowercase = key.replace('.qkv_proj.' , '.q_proj.' ) lowercase = key.replace('.qkv_proj.' , '.k_proj.' ) lowercase = key.replace('.qkv_proj.' , '.v_proj.' ) lowercase = value.shape[0] assert depth % 3 == 0 # `SequeuceParallelTransformerBlock` has QKV weight is separated in K,V,Q despite the naming: # https://cs.github.com/facebookresearch/metaseq/blob/51871bd73cd04c038f239ea2a26db1d7f6b37927/metaseq/modules/sequence_parallel_transformer_layer.py#L97 lowercase , lowercase , lowercase = torch.split(__SCREAMING_SNAKE_CASE , depth // 3 , dim=0 ) lowercase = q lowercase = k lowercase = v del sd[key] return sd @torch.no_grad() def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=None ): lowercase = load_checkpoint(__SCREAMING_SNAKE_CASE ) if config is not None: lowercase = OPTConfig.from_pretrained(__SCREAMING_SNAKE_CASE ) else: lowercase = OPTConfig() lowercase = OPTModel(__SCREAMING_SNAKE_CASE ).half().eval() model.load_state_dict(__SCREAMING_SNAKE_CASE ) # Check results Path(__SCREAMING_SNAKE_CASE ).mkdir(exist_ok=__SCREAMING_SNAKE_CASE ) model.save_pretrained(__SCREAMING_SNAKE_CASE ) if __name__ == "__main__": UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--fairseq_path''', type=str, help=( '''path to fairseq checkpoint in correct format. You can find all checkpoints in the correct format here:''' ''' https://huggingface.co/models?other=opt_metasq''' ), ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--hf_config''', default=None, type=str, help='''Define HF config.''') UpperCAmelCase = parser.parse_args() convert_opt_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, config=args.hf_config)
84
from math import sqrt def lowerCAmelCase_ ( __lowerCamelCase = 1_0_0_0_0_0_0 ): __snake_case : int = 0 __snake_case : int = 0 __snake_case : int while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(__lowerCamelCase , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(f'''{solution() = }''')
81
0
import copy import json import os import tempfile from transformers import is_torch_available from .test_configuration_utils import config_common_kwargs class snake_case ( UpperCamelCase_ ): def __init__( self : Dict , a_ : str , a_ : Optional[int]=None , a_ : str=True , a_ : Optional[Any]=None , **a_ : Optional[int] )-> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = parent SCREAMING_SNAKE_CASE__ : str = config_class SCREAMING_SNAKE_CASE__ : Union[str, Any] = has_text_modality SCREAMING_SNAKE_CASE__ : Optional[int] = kwargs SCREAMING_SNAKE_CASE__ : str = common_properties def __lowercase( self : List[str] )-> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : str = self.config_class(**self.inputs_dict ) SCREAMING_SNAKE_CASE__ : int = ( ['hidden_size', 'num_attention_heads', 'num_hidden_layers'] if self.common_properties is None else self.common_properties ) # Add common fields for text models if self.has_text_modality: common_properties.extend(['vocab_size'] ) # Test that config has the common properties as getters for prop in common_properties: self.parent.assertTrue(hasattr(a_ , a_ ) , msg=F'''`{prop}` does not exist''' ) # Test that config has the common properties as setter for idx, name in enumerate(a_ ): try: setattr(a_ , a_ , a_ ) self.parent.assertEqual( getattr(a_ , a_ ) , a_ , msg=F'''`{name} value {idx} expected, but was {getattr(a_ , a_ )}''' ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass # Test if config class can be called with Config(prop_name=..) for idx, name in enumerate(a_ ): try: SCREAMING_SNAKE_CASE__ : Tuple = self.config_class(**{name: idx} ) self.parent.assertEqual( getattr(a_ , a_ ) , a_ , msg=F'''`{name} value {idx} expected, but was {getattr(a_ , a_ )}''' ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass def __lowercase( self : Tuple )-> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : str = self.config_class(**self.inputs_dict ) SCREAMING_SNAKE_CASE__ : Tuple = json.loads(config.to_json_string() ) for key, value in self.inputs_dict.items(): self.parent.assertEqual(obj[key] , a_ ) def __lowercase( self : int )-> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: SCREAMING_SNAKE_CASE__ : Optional[Any] = os.path.join(a_ , 'config.json' ) config_first.to_json_file(a_ ) SCREAMING_SNAKE_CASE__ : int = self.config_class.from_json_file(a_ ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def __lowercase( self : int )-> int: """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: config_first.save_pretrained(a_ ) SCREAMING_SNAKE_CASE__ : Tuple = self.config_class.from_pretrained(a_ ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def __lowercase( self : Any )-> int: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[Any] = self.config_class(**self.inputs_dict ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = 'test' with tempfile.TemporaryDirectory() as tmpdirname: SCREAMING_SNAKE_CASE__ : List[str] = os.path.join(a_ , a_ ) config_first.save_pretrained(a_ ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.config_class.from_pretrained(a_ , subfolder=a_ ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def __lowercase( self : Union[str, Any] )-> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : Tuple = self.config_class(**self.inputs_dict , num_labels=5 ) self.parent.assertEqual(len(config.idalabel ) , 5 ) self.parent.assertEqual(len(config.labelaid ) , 5 ) SCREAMING_SNAKE_CASE__ : int = 3 self.parent.assertEqual(len(config.idalabel ) , 3 ) self.parent.assertEqual(len(config.labelaid ) , 3 ) def __lowercase( self : Any )-> Dict: """simple docstring""" if self.config_class.is_composition: return SCREAMING_SNAKE_CASE__ : Any = self.config_class() self.parent.assertIsNotNone(a_ ) def __lowercase( self : int )-> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ : Any = copy.deepcopy(a_ ) SCREAMING_SNAKE_CASE__ : Tuple = self.config_class(**a_ ) SCREAMING_SNAKE_CASE__ : Any = [] for key, value in config_common_kwargs.items(): if key == "torch_dtype": if not is_torch_available(): continue else: import torch if config.torch_dtype != torch.floataa: wrong_values.append(('torch_dtype', config.torch_dtype, torch.floataa) ) elif getattr(a_ , a_ ) != value: wrong_values.append((key, getattr(a_ , a_ ), value) ) if len(a_ ) > 0: SCREAMING_SNAKE_CASE__ : str = '\n'.join([F'''- {v[0]}: got {v[1]} instead of {v[2]}''' for v in wrong_values] ) raise ValueError(F'''The following keys were not properly set in the config:\n{errors}''' ) def __lowercase( self : Tuple )-> Dict: """simple docstring""" self.create_and_test_config_common_properties() self.create_and_test_config_to_json_string() self.create_and_test_config_to_json_file() self.create_and_test_config_from_and_save_pretrained() self.create_and_test_config_from_and_save_pretrained_subfolder() self.create_and_test_config_with_num_labels() self.check_config_can_be_init_without_params() self.check_config_arguments_init()
85
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class a (_lowerCAmelCase ): """simple docstring""" def __snake_case ( self : str ) -> str: __snake_case : Tuple = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCamelCase , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "neck_hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "num_attention_heads" ) ) class a : """simple docstring""" def __init__( self : Optional[int] , lowerCamelCase : List[str] , lowerCamelCase : Tuple=13 , lowerCamelCase : str=32 , lowerCamelCase : Dict=2 , lowerCamelCase : List[str]=3 , lowerCamelCase : Any=640 , lowerCamelCase : Optional[Any]=4 , lowerCamelCase : Tuple="silu" , lowerCamelCase : int=3 , lowerCamelCase : Dict=32 , lowerCamelCase : str=0.1 , lowerCamelCase : Optional[int]=0.1 , lowerCamelCase : Optional[Any]=0.1 , lowerCamelCase : Dict=0.02 , lowerCamelCase : Union[str, Any]=True , lowerCamelCase : Optional[int]=True , lowerCamelCase : Union[str, Any]=10 , lowerCamelCase : int=None , ) -> str: __snake_case : Optional[Any] = parent __snake_case : Optional[Any] = batch_size __snake_case : Any = image_size __snake_case : List[Any] = patch_size __snake_case : Any = num_channels __snake_case : Union[str, Any] = last_hidden_size __snake_case : Any = num_attention_heads __snake_case : Any = hidden_act __snake_case : Tuple = conv_kernel_size __snake_case : Any = output_stride __snake_case : Any = hidden_dropout_prob __snake_case : List[Any] = attention_probs_dropout_prob __snake_case : Optional[Any] = classifier_dropout_prob __snake_case : Union[str, Any] = use_labels __snake_case : Optional[int] = is_training __snake_case : Dict = num_labels __snake_case : Any = initializer_range __snake_case : Optional[int] = scope def __snake_case ( self : str ) -> Union[str, Any]: __snake_case : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : List[Any] = None __snake_case : Optional[int] = None if self.use_labels: __snake_case : Optional[int] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : Any = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __snake_case : Optional[Any] = self.get_config() return config, pixel_values, labels, pixel_labels def __snake_case ( self : Any ) -> Union[str, Any]: return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __snake_case ( self : Any , lowerCamelCase : Tuple , lowerCamelCase : int , lowerCamelCase : Tuple , lowerCamelCase : Optional[int] ) -> Dict: __snake_case : List[Any] = MobileViTModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[str] = model(lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[Any] , lowerCamelCase : List[str] , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] , lowerCamelCase : Tuple ) -> List[str]: __snake_case : str = self.num_labels __snake_case : List[Any] = MobileViTForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : Dict , lowerCamelCase : Any , lowerCamelCase : Any , lowerCamelCase : Optional[Any] , lowerCamelCase : Dict ) -> Dict: __snake_case : Union[str, Any] = self.num_labels __snake_case : Optional[int] = MobileViTForSemanticSegmentation(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Tuple = model(lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[int] ) -> List[Any]: __snake_case : Optional[Any] = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case , __snake_case : Union[str, Any] = config_and_inputs __snake_case : Dict = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) __UpperCAmelCase : Optional[Any] = ( { "feature-extraction": MobileViTModel, "image-classification": MobileViTForImageClassification, "image-segmentation": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) __UpperCAmelCase : List[str] = False __UpperCAmelCase : int = False __UpperCAmelCase : Optional[int] = False __UpperCAmelCase : Optional[int] = False def __snake_case ( self : Optional[int] ) -> Dict: __snake_case : Tuple = MobileViTModelTester(self ) __snake_case : Any = MobileViTConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase ) def __snake_case ( self : Optional[int] ) -> Dict: self.config_tester.run_common_tests() @unittest.skip(reason="MobileViT does not use inputs_embeds" ) def __snake_case ( self : Dict ) -> Any: pass @unittest.skip(reason="MobileViT does not support input and output embeddings" ) def __snake_case ( self : Dict ) -> List[Any]: pass @unittest.skip(reason="MobileViT does not output attentions" ) def __snake_case ( self : int ) -> Dict: pass def __snake_case ( self : int ) -> Union[str, Any]: __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Optional[Any] = model_class(lowerCamelCase ) __snake_case : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : int = [*signature.parameters.keys()] __snake_case : List[str] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __snake_case ( self : int ) -> Tuple: pass def __snake_case ( self : Any ) -> Tuple: __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def __snake_case ( self : Any ) -> str: def check_hidden_states_output(lowerCamelCase : Union[str, Any] , lowerCamelCase : Dict , lowerCamelCase : Any ): __snake_case : int = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : int = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : Union[str, Any] = outputs.hidden_states __snake_case : int = 5 self.assertEqual(len(lowerCamelCase ) , lowerCamelCase ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. __snake_case : List[Any] = 2 for i in range(len(lowerCamelCase ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) __snake_case , __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : str = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : List[Any] = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Any ) -> Any: __snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) def __snake_case ( self : List[str] ) -> List[str]: __snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowerCamelCase ) @slow def __snake_case ( self : List[str] ) -> Any: for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : List[str] = MobileViTModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def lowerCAmelCase_ ( ): __snake_case : Optional[int] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a (unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self : str ) -> Dict: return MobileViTImageProcessor.from_pretrained("apple/mobilevit-xx-small" ) if is_vision_available() else None @slow def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : Tuple = MobileViTForImageClassification.from_pretrained("apple/mobilevit-xx-small" ).to(lowerCamelCase ) __snake_case : Optional[Any] = self.default_image_processor __snake_case : Union[str, Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Dict = model(**lowerCamelCase ) # verify the logits __snake_case : Union[str, Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([-1.93_64, -1.23_27, -0.46_53] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : str ) -> Optional[int]: __snake_case : Optional[int] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : str = model.to(lowerCamelCase ) __snake_case : int = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Optional[int] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : List[str] = model(**lowerCamelCase ) __snake_case : Union[str, Any] = outputs.logits # verify the logits __snake_case : Tuple = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , lowerCamelCase ) __snake_case : Union[str, Any] = torch.tensor( [ [[6.97_13, 6.97_86, 7.24_22], [7.28_93, 7.28_25, 7.44_46], [7.65_80, 7.87_97, 7.94_20]], [[-10.68_69, -10.32_50, -10.34_71], [-10.42_28, -9.98_68, -9.71_32], [-11.04_05, -11.02_21, -10.73_18]], [[-3.30_89, -2.85_39, -2.67_40], [-3.27_06, -2.56_21, -2.51_08], [-3.25_34, -2.66_15, -2.66_51]], ] , device=lowerCamelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : Union[str, Any] ) -> Optional[int]: __snake_case : Optional[Any] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Tuple = model.to(lowerCamelCase ) __snake_case : Dict = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : List[Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Any = model(**lowerCamelCase ) __snake_case : Dict = outputs.logits.detach().cpu() __snake_case : Any = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase , target_sizes=[(50, 60)] ) __snake_case : int = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , lowerCamelCase ) __snake_case : List[str] = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase ) __snake_case : Optional[int] = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , lowerCamelCase )
81
0
import os # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_doctest_list.py __a :Optional[int] = '.' if __name__ == "__main__": __a :str = os.path.join(REPO_PATH, 'utils/documentation_tests.txt') __a :Any = [] __a :Optional[Any] = [] with open(doctest_file_path) as fp: for line in fp: __a :List[str] = line.strip() __a :Optional[Any] = os.path.join(REPO_PATH, line) if not (os.path.isfile(path) or os.path.isdir(path)): non_existent_paths.append(line) all_paths.append(path) if len(non_existent_paths) > 0: __a :Optional[Any] = '\n'.join(non_existent_paths) raise ValueError(F"`utils/documentation_tests.txt` contains non-existent paths:\n{non_existent_paths}") if all_paths != sorted(all_paths): raise ValueError('Files in `utils/documentation_tests.txt` are not in alphabetical order.')
86
import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor _snake_case : int = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[int] , *lowerCamelCase : List[Any] , **lowerCamelCase : int ) -> None: warnings.warn( "The class ImageGPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use ImageGPTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _lowerCamelCase : Tuple = { """configuration_nezha""": ["""NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """NezhaConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : int = [ """NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST""", """NezhaForNextSentencePrediction""", """NezhaForMaskedLM""", """NezhaForPreTraining""", """NezhaForMultipleChoice""", """NezhaForQuestionAnswering""", """NezhaForSequenceClassification""", """NezhaForTokenClassification""", """NezhaModel""", """NezhaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_nezha import NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP, NezhaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nezha import ( NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, NezhaPreTrainedModel, ) else: import sys _lowerCamelCase : Any = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
87
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _snake_case : Optional[Any] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[str] = ["pixel_values"] def __init__( self : List[Any] , lowerCamelCase : bool = True , lowerCamelCase : Union[int, float] = 1 / 255 , lowerCamelCase : bool = True , lowerCamelCase : int = 8 , **lowerCamelCase : Tuple , ) -> None: super().__init__(**lowerCamelCase ) __snake_case : Dict = do_rescale __snake_case : Dict = rescale_factor __snake_case : Optional[Any] = do_pad __snake_case : Tuple = pad_size def __snake_case ( self : Dict , lowerCamelCase : np.ndarray , lowerCamelCase : float , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase : Optional[int] ) -> np.ndarray: return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def __snake_case ( self : Optional[Any] , lowerCamelCase : np.ndarray , lowerCamelCase : int , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None ) -> Tuple: __snake_case , __snake_case : List[str] = get_image_size(lowerCamelCase ) __snake_case : Optional[Any] = (old_height // size + 1) * size - old_height __snake_case : List[Any] = (old_width // size + 1) * size - old_width return pad(lowerCamelCase , ((0, pad_height), (0, pad_width)) , mode="symmetric" , data_format=lowerCamelCase ) def __snake_case ( self : Tuple , lowerCamelCase : ImageInput , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[float] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[int] = None , lowerCamelCase : Optional[Union[str, TensorType]] = None , lowerCamelCase : Union[str, ChannelDimension] = ChannelDimension.FIRST , **lowerCamelCase : Union[str, Any] , ) -> List[str]: __snake_case : int = do_rescale if do_rescale is not None else self.do_rescale __snake_case : Dict = rescale_factor if rescale_factor is not None else self.rescale_factor __snake_case : str = do_pad if do_pad is not None else self.do_pad __snake_case : Any = pad_size if pad_size is not None else self.pad_size __snake_case : int = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) # All transformations expect numpy arrays. __snake_case : str = [to_numpy_array(lowerCamelCase ) for image in images] if do_rescale: __snake_case : Optional[int] = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_pad: __snake_case : Optional[Any] = [self.pad(lowerCamelCase , size=lowerCamelCase ) for image in images] __snake_case : int = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __snake_case : Union[str, Any] = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
81
0
"""simple docstring""" import argparse import copy def _snake_case ( __snake_case : Tuple ): """simple docstring""" _lowerCamelCase : str = {} with open(__snake_case ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: _lowerCamelCase : Optional[Any] = [] _list.append([line.split()[1], line.split()[2]] ) _lowerCamelCase : int = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: _lowerCamelCase : str = [] _list.append([line.split()[0], line.split()[2]] ) _lowerCamelCase : Any = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def _snake_case ( __snake_case : List[Any] , __snake_case : int ): """simple docstring""" with open(__snake_case ) as f: _lowerCamelCase : str = f.read(1 ) _lowerCamelCase : Dict = start_node _lowerCamelCase : Tuple = [] _lowerCamelCase : int = start_node _lowerCamelCase : Tuple = 0 while visiting not in first_solution: _lowerCamelCase : int = 10000 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(__snake_case ) and k[0] not in first_solution: _lowerCamelCase : List[Any] = k[1] _lowerCamelCase : Any = k[0] first_solution.append(__snake_case ) _lowerCamelCase : Optional[int] = distance_of_first_solution + int(__snake_case ) _lowerCamelCase : Any = best_node first_solution.append(__snake_case ) _lowerCamelCase : List[Any] = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 _lowerCamelCase : str = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 10000 ) return first_solution, distance_of_first_solution def _snake_case ( __snake_case : List[Any] , __snake_case : Optional[Any] ): """simple docstring""" _lowerCamelCase : str = [] for n in solution[1:-1]: _lowerCamelCase : Tuple = solution.index(__snake_case ) for kn in solution[1:-1]: _lowerCamelCase : Union[str, Any] = solution.index(__snake_case ) if n == kn: continue _lowerCamelCase : Optional[int] = copy.deepcopy(__snake_case ) _lowerCamelCase : Any = kn _lowerCamelCase : str = n _lowerCamelCase : List[str] = 0 for k in _tmp[:-1]: _lowerCamelCase : List[Any] = _tmp[_tmp.index(__snake_case ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: _lowerCamelCase : str = distance + int(i[1] ) _tmp.append(__snake_case ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) _lowerCamelCase : List[Any] = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda __snake_case : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def _snake_case ( __snake_case : List[str] , __snake_case : Optional[Any] , __snake_case : Optional[Any] , __snake_case : Dict , __snake_case : List[Any] ): """simple docstring""" _lowerCamelCase : Any = 1 _lowerCamelCase : str = first_solution _lowerCamelCase : int = [] _lowerCamelCase : str = distance_of_first_solution _lowerCamelCase : int = solution while count <= iters: _lowerCamelCase : Optional[Any] = find_neighborhood(__snake_case , __snake_case ) _lowerCamelCase : Tuple = 0 _lowerCamelCase : List[Any] = neighborhood[index_of_best_solution] _lowerCamelCase : List[str] = len(__snake_case ) - 1 _lowerCamelCase : Optional[int] = False while not found: _lowerCamelCase : str = 0 while i < len(__snake_case ): if best_solution[i] != solution[i]: _lowerCamelCase : Optional[Any] = best_solution[i] _lowerCamelCase : List[str] = solution[i] break _lowerCamelCase : int = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) _lowerCamelCase : Optional[int] = True _lowerCamelCase : Any = best_solution[:-1] _lowerCamelCase : Dict = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: _lowerCamelCase : Tuple = cost _lowerCamelCase : Optional[Any] = solution else: _lowerCamelCase : Dict = index_of_best_solution + 1 _lowerCamelCase : Union[str, Any] = neighborhood[index_of_best_solution] if len(__snake_case ) >= size: tabu_list.pop(0 ) _lowerCamelCase : Any = count + 1 return best_solution_ever, best_cost def _snake_case ( __snake_case : Any=None ): """simple docstring""" _lowerCamelCase : Dict = generate_neighbours(args.File ) _lowerCamelCase , _lowerCamelCase : Any = generate_first_solution( args.File , __snake_case ) _lowerCamelCase , _lowerCamelCase : List[str] = tabu_search( __snake_case , __snake_case , __snake_case , args.Iterations , args.Size , ) print(F'Best solution: {best_sol}, with total distance: {best_cost}.' ) if __name__ == "__main__": UpperCAmelCase = argparse.ArgumentParser(description="""Tabu Search""") parser.add_argument( """-f""", """--File""", type=str, help="""Path to the file containing the data""", required=True, ) parser.add_argument( """-i""", """--Iterations""", type=int, help="""How many iterations the algorithm should perform""", required=True, ) parser.add_argument( """-s""", """--Size""", type=int, help="""Size of the tabu list""", required=True ) # Pass the arguments to main method main(parser.parse_args())
88
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case : Any = { "configuration_informer": [ "INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "InformerConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "InformerForPrediction", "InformerModel", "InformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys _snake_case : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
0
from __future__ import annotations import typing from collections import Counter def UpperCamelCase_( lowerCamelCase_ ) -> typing.Counter[int]: _lowercase : typing.Counter[int] = Counter() for base in range(1 , max_perimeter + 1 ): for perpendicular in range(lowerCamelCase_ , max_perimeter + 1 ): _lowercase : str = (base * base + perpendicular * perpendicular) ** 0.5 if hypotenuse == int(lowerCamelCase_ ): _lowercase : Any = int(base + perpendicular + hypotenuse ) if perimeter > max_perimeter: continue triplets[perimeter] += 1 return triplets def UpperCamelCase_( lowerCamelCase_ = 1000 ) -> int: _lowercase : List[str] = pythagorean_triple(lowerCamelCase_ ) return triplets.most_common(1 )[0][0] if __name__ == "__main__": print(F"Perimeter {solution()} has maximum solutions")
89
from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging _snake_case : Any = logging.get_logger(__name__) # pylint: disable=invalid-name class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[Any] , lowerCamelCase : CLIPSegForImageSegmentation , lowerCamelCase : CLIPSegProcessor , lowerCamelCase : AutoencoderKL , lowerCamelCase : CLIPTextModel , lowerCamelCase : CLIPTokenizer , lowerCamelCase : UNetaDConditionModel , lowerCamelCase : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , lowerCamelCase : StableDiffusionSafetyChecker , lowerCamelCase : CLIPImageProcessor , ) -> Tuple: super().__init__() if hasattr(scheduler.config , "steps_offset" ) and scheduler.config.steps_offset != 1: __snake_case : Tuple = ( F'The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`' F' should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure ' "to update the config accordingly as leaving `steps_offset` might led to incorrect results" " in future versions. If you have downloaded this checkpoint from the Hugging Face Hub," " it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`" " file" ) deprecate("steps_offset!=1" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : Any = dict(scheduler.config ) __snake_case : List[Any] = 1 __snake_case : Tuple = FrozenDict(lowerCamelCase ) if hasattr(scheduler.config , "skip_prk_steps" ) and scheduler.config.skip_prk_steps is False: __snake_case : List[str] = ( F'The configuration file of this scheduler: {scheduler} has not set the configuration' " `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make" " sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to" " incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face" " Hub, it would be very nice if you could open a Pull request for the" " `scheduler/scheduler_config.json` file" ) deprecate("skip_prk_steps not set" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : List[str] = dict(scheduler.config ) __snake_case : List[str] = True __snake_case : Any = FrozenDict(lowerCamelCase ) if safety_checker is None: logger.warning( F'You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure' " that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered" " results in services or applications open to the public. Both the diffusers team and Hugging Face" " strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling" " it only for use-cases that involve analyzing network behavior or auditing its results. For more" " information, please have a look at https://github.com/huggingface/diffusers/pull/254 ." ) self.register_modules( segmentation_model=lowerCamelCase , segmentation_processor=lowerCamelCase , vae=lowerCamelCase , text_encoder=lowerCamelCase , tokenizer=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase , safety_checker=lowerCamelCase , feature_extractor=lowerCamelCase , ) def __snake_case ( self : Dict , lowerCamelCase : Optional[Union[str, int]] = "auto" ) -> Tuple: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __snake_case : Any = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowerCamelCase ) def __snake_case ( self : List[Any] ) -> Any: self.enable_attention_slicing(lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> str: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) __snake_case : Optional[int] = torch.device("cuda" ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(lowerCamelCase , lowerCamelCase ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __snake_case ( self : int ) -> Any: if self.device != torch.device("meta" ) or not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowerCamelCase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self : List[Any] , lowerCamelCase : Union[str, List[str]] , lowerCamelCase : Union[torch.FloatTensor, PIL.Image.Image] , lowerCamelCase : str , lowerCamelCase : int = 512 , lowerCamelCase : int = 512 , lowerCamelCase : int = 50 , lowerCamelCase : float = 7.5 , lowerCamelCase : Optional[Union[str, List[str]]] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : float = 0.0 , lowerCamelCase : Optional[torch.Generator] = None , lowerCamelCase : Optional[torch.FloatTensor] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , lowerCamelCase : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , lowerCamelCase : int = 1 , **lowerCamelCase : Dict , ) -> List[str]: __snake_case : Tuple = self.segmentation_processor( text=[text] , images=[image] , padding="max_length" , return_tensors="pt" ).to(self.device ) __snake_case : str = self.segmentation_model(**lowerCamelCase ) __snake_case : Dict = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() __snake_case : List[Any] = self.numpy_to_pil(lowerCamelCase )[0].resize(image.size ) # Run inpainting pipeline with the generated mask __snake_case : Tuple = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=lowerCamelCase , image=lowerCamelCase , mask_image=lowerCamelCase , height=lowerCamelCase , width=lowerCamelCase , num_inference_steps=lowerCamelCase , guidance_scale=lowerCamelCase , negative_prompt=lowerCamelCase , num_images_per_prompt=lowerCamelCase , eta=lowerCamelCase , generator=lowerCamelCase , latents=lowerCamelCase , output_type=lowerCamelCase , return_dict=lowerCamelCase , callback=lowerCamelCase , callback_steps=lowerCamelCase , )
81
0
'''simple docstring''' import random import unittest from torch.utils.data import BatchSampler, DataLoader, IterableDataset from accelerate import Accelerator from accelerate.data_loader import ( BatchSamplerShard, DataLoaderDispatcher, DataLoaderShard, IterableDatasetShard, SkipBatchSampler, SkipDataLoader, skip_first_batches, ) class a__ ( a__ ): '''simple docstring''' def __init__( self , lowerCamelCase_=0.01 , lowerCamelCase_=10_00 ) -> Union[str, Any]: lowerCAmelCase__ = p_stop lowerCAmelCase__ = max_length def __iter__( self ) -> Any: lowerCAmelCase__ = 0 lowerCAmelCase__ = False while not stop and count < self.max_length: yield count count += 1 lowerCAmelCase__ = random.random() < self.p_stop class a__ ( unittest.TestCase ): '''simple docstring''' def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_=False , lowerCamelCase_=True ) -> Optional[Any]: lowerCAmelCase__ = [ BatchSamplerShard(lowerCamelCase_ , 2 , lowerCamelCase_ , split_batches=lowerCamelCase_ , even_batches=lowerCamelCase_ ) for i in range(2 ) ] lowerCAmelCase__ = [list(lowerCamelCase_ ) for batch_sampler_shard in batch_sampler_shards] if not split_batches: self.assertListEqual([len(lowerCamelCase_ ) for shard in batch_sampler_shards] , [len(lowerCamelCase_ ) for e in expected] ) self.assertListEqual(lowerCamelCase_ , lowerCamelCase_ ) def __SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: # Check the shards when the dataset is a round multiple of total batch size. lowerCAmelCase__ = BatchSampler(range(24 ) , batch_size=3 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 22, 23]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ ) lowerCAmelCase__ = BatchSampler(range(24 ) , batch_size=3 , drop_last=lowerCamelCase_ ) # Expected shouldn't change self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. lowerCAmelCase__ = BatchSampler(range(21 ) , batch_size=3 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [0, 1, 2]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ ) lowerCAmelCase__ = BatchSampler(range(21 ) , batch_size=3 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. lowerCAmelCase__ = BatchSampler(range(22 ) , batch_size=3 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 0, 1]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ ) lowerCAmelCase__ = BatchSampler(range(22 ) , batch_size=3 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. lowerCAmelCase__ = BatchSampler(range(20 ) , batch_size=3 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 0]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [1, 2, 3]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ ) lowerCAmelCase__ = BatchSampler(range(20 ) , batch_size=3 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ ) # Check the shards when the dataset is very small. lowerCAmelCase__ = BatchSampler(range(2 ) , batch_size=3 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [[[0, 1, 0]], [[1, 0, 1]]] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ ) lowerCAmelCase__ = BatchSampler(range(2 ) , batch_size=3 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [[], []] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ ) def __SCREAMING_SNAKE_CASE ( self ) -> str: # Check the shards when the dataset is a round multiple of batch size. lowerCAmelCase__ = BatchSampler(range(24 ) , batch_size=4 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [22, 23]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , split_batches=lowerCamelCase_ ) lowerCAmelCase__ = BatchSampler(range(24 ) , batch_size=4 , drop_last=lowerCamelCase_ ) # Expected shouldn't change self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , split_batches=lowerCamelCase_ ) # Check the shards when the dataset is not a round multiple of batch size. lowerCAmelCase__ = BatchSampler(range(22 ) , batch_size=4 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [0, 1]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , split_batches=lowerCamelCase_ ) lowerCAmelCase__ = BatchSampler(range(22 ) , batch_size=4 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , split_batches=lowerCamelCase_ ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. lowerCAmelCase__ = BatchSampler(range(21 ) , batch_size=4 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 0]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [1, 2]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , split_batches=lowerCamelCase_ ) lowerCAmelCase__ = BatchSampler(range(21 ) , batch_size=4 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , split_batches=lowerCamelCase_ ) # Check the shards when the dataset is very small. lowerCAmelCase__ = BatchSampler(range(2 ) , batch_size=4 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [[[0, 1]], [[0, 1]]] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , split_batches=lowerCamelCase_ ) lowerCAmelCase__ = BatchSampler(range(2 ) , batch_size=4 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [[], []] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , split_batches=lowerCamelCase_ ) def __SCREAMING_SNAKE_CASE ( self ) -> Any: # Check the shards when the dataset is a round multiple of total batch size. lowerCAmelCase__ = BatchSampler(range(24 ) , batch_size=3 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 22, 23]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , even_batches=lowerCamelCase_ ) lowerCAmelCase__ = BatchSampler(range(24 ) , batch_size=3 , drop_last=lowerCamelCase_ ) # Expected shouldn't change self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , even_batches=lowerCamelCase_ ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. lowerCAmelCase__ = BatchSampler(range(21 ) , batch_size=3 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , even_batches=lowerCamelCase_ ) lowerCAmelCase__ = BatchSampler(range(21 ) , batch_size=3 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , even_batches=lowerCamelCase_ ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. lowerCAmelCase__ = BatchSampler(range(22 ) , batch_size=3 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , even_batches=lowerCamelCase_ ) lowerCAmelCase__ = BatchSampler(range(22 ) , batch_size=3 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , even_batches=lowerCamelCase_ ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. lowerCAmelCase__ = BatchSampler(range(20 ) , batch_size=3 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , even_batches=lowerCamelCase_ ) lowerCAmelCase__ = BatchSampler(range(20 ) , batch_size=3 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , even_batches=lowerCamelCase_ ) # Check the shards when the dataset is very small. lowerCAmelCase__ = BatchSampler(range(2 ) , batch_size=3 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [[[0, 1]], []] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , even_batches=lowerCamelCase_ ) lowerCAmelCase__ = BatchSampler(range(2 ) , batch_size=3 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [[], []] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , even_batches=lowerCamelCase_ ) def __SCREAMING_SNAKE_CASE ( self ) -> Any: # Check the shards when the dataset is a round multiple of batch size. lowerCAmelCase__ = BatchSampler(range(24 ) , batch_size=4 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [22, 23]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , split_batches=lowerCamelCase_ , even_batches=lowerCamelCase_ ) lowerCAmelCase__ = BatchSampler(range(24 ) , batch_size=4 , drop_last=lowerCamelCase_ ) # Expected shouldn't change self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , split_batches=lowerCamelCase_ , even_batches=lowerCamelCase_ ) # Check the shards when the dataset is not a round multiple of batch size. lowerCAmelCase__ = BatchSampler(range(22 ) , batch_size=4 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , split_batches=lowerCamelCase_ , even_batches=lowerCamelCase_ ) lowerCAmelCase__ = BatchSampler(range(22 ) , batch_size=4 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , split_batches=lowerCamelCase_ , even_batches=lowerCamelCase_ ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. lowerCAmelCase__ = BatchSampler(range(21 ) , batch_size=4 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , split_batches=lowerCamelCase_ , even_batches=lowerCamelCase_ ) lowerCAmelCase__ = BatchSampler(range(21 ) , batch_size=4 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , split_batches=lowerCamelCase_ , even_batches=lowerCamelCase_ ) # Check the shards when the dataset is very small. lowerCAmelCase__ = BatchSampler(range(2 ) , batch_size=4 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [[[0, 1]], []] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , split_batches=lowerCamelCase_ , even_batches=lowerCamelCase_ ) lowerCAmelCase__ = BatchSampler(range(2 ) , batch_size=4 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = [[], []] self.check_batch_sampler_shards(lowerCamelCase_ , lowerCamelCase_ , split_batches=lowerCamelCase_ , even_batches=lowerCamelCase_ ) def __SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: lowerCAmelCase__ = [[0, 1, 2], [3, 4], [5, 6, 7, 8], [9, 10, 11], [12, 13]] lowerCAmelCase__ = [BatchSamplerShard(lowerCamelCase_ , 2 , lowerCamelCase_ , even_batches=lowerCamelCase_ ) for i in range(2 )] self.assertEqual(len(batch_sampler_shards[0] ) , 3 ) self.assertEqual(len(batch_sampler_shards[1] ) , 2 ) self.assertListEqual(list(batch_sampler_shards[0] ) , [[0, 1, 2], [5, 6, 7, 8], [12, 13]] ) self.assertListEqual(list(batch_sampler_shards[1] ) , [[3, 4], [9, 10, 11]] ) def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_=False , lowerCamelCase_=2 , lowerCamelCase_=False ) -> str: random.seed(lowerCamelCase_ ) lowerCAmelCase__ = list(lowerCamelCase_ ) lowerCAmelCase__ = [ IterableDatasetShard( lowerCamelCase_ , batch_size=lowerCamelCase_ , drop_last=lowerCamelCase_ , num_processes=lowerCamelCase_ , process_index=lowerCamelCase_ , split_batches=lowerCamelCase_ , ) for i in range(lowerCamelCase_ ) ] lowerCAmelCase__ = [] for iterable_dataset_shard in iterable_dataset_shards: # Since our random iterable dataset will be... random... we need to use a seed to get reproducible results. random.seed(lowerCamelCase_ ) iterable_dataset_lists.append(list(lowerCamelCase_ ) ) lowerCAmelCase__ = batch_size // num_processes if split_batches else batch_size # All iterable dataset shard should have the same length, a round multiple of shard_batch_size lowerCAmelCase__ = iterable_dataset_lists[0] for l in iterable_dataset_lists[1:]: self.assertEqual(len(lowerCamelCase_ ) , len(lowerCamelCase_ ) ) self.assertTrue(len(lowerCamelCase_ ) % shard_batch_size == 0 ) lowerCAmelCase__ = [] for idx in range(0 , len(lowerCamelCase_ ) , lowerCamelCase_ ): for l in iterable_dataset_lists: observed += l[idx : idx + shard_batch_size] if not drop_last: while len(lowerCamelCase_ ) < len(lowerCamelCase_ ): reference += reference self.assertListEqual(lowerCamelCase_ , reference[: len(lowerCamelCase_ )] ) def __SCREAMING_SNAKE_CASE ( self ) -> Tuple: lowerCAmelCase__ = 42 lowerCAmelCase__ = RandomIterableDataset() self.check_iterable_dataset_shards(lowerCamelCase_ , lowerCamelCase_ , batch_size=4 , drop_last=lowerCamelCase_ , split_batches=lowerCamelCase_ ) self.check_iterable_dataset_shards(lowerCamelCase_ , lowerCamelCase_ , batch_size=4 , drop_last=lowerCamelCase_ , split_batches=lowerCamelCase_ ) self.check_iterable_dataset_shards(lowerCamelCase_ , lowerCamelCase_ , batch_size=4 , drop_last=lowerCamelCase_ , split_batches=lowerCamelCase_ ) self.check_iterable_dataset_shards(lowerCamelCase_ , lowerCamelCase_ , batch_size=4 , drop_last=lowerCamelCase_ , split_batches=lowerCamelCase_ ) # Edge case with a very small dataset lowerCAmelCase__ = RandomIterableDataset(max_length=2 ) self.check_iterable_dataset_shards(lowerCamelCase_ , lowerCamelCase_ , batch_size=4 , drop_last=lowerCamelCase_ , split_batches=lowerCamelCase_ ) self.check_iterable_dataset_shards(lowerCamelCase_ , lowerCamelCase_ , batch_size=4 , drop_last=lowerCamelCase_ , split_batches=lowerCamelCase_ ) self.check_iterable_dataset_shards(lowerCamelCase_ , lowerCamelCase_ , batch_size=4 , drop_last=lowerCamelCase_ , split_batches=lowerCamelCase_ ) self.check_iterable_dataset_shards(lowerCamelCase_ , lowerCamelCase_ , batch_size=4 , drop_last=lowerCamelCase_ , split_batches=lowerCamelCase_ ) def __SCREAMING_SNAKE_CASE ( self ) -> Any: lowerCAmelCase__ = BatchSampler(range(16 ) , batch_size=4 , drop_last=lowerCamelCase_ ) lowerCAmelCase__ = SkipBatchSampler(lowerCamelCase_ , 2 ) self.assertListEqual(list(lowerCamelCase_ ) , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def __SCREAMING_SNAKE_CASE ( self ) -> List[str]: lowerCAmelCase__ = SkipDataLoader(list(range(16 ) ) , batch_size=4 , skip_batches=2 ) self.assertListEqual([t.tolist() for t in dataloader] , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def __SCREAMING_SNAKE_CASE ( self ) -> Tuple: lowerCAmelCase__ = DataLoader(list(range(16 ) ) , batch_size=4 ) lowerCAmelCase__ = skip_first_batches(lowerCamelCase_ , num_batches=2 ) self.assertListEqual([t.tolist() for t in new_dataloader] , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def __SCREAMING_SNAKE_CASE ( self ) -> List[Any]: lowerCAmelCase__ = DataLoaderShard(list(range(16 ) ) , batch_size=4 ) for idx, _ in enumerate(lowerCamelCase_ ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(lowerCamelCase_ ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) def __SCREAMING_SNAKE_CASE ( self ) -> str: Accelerator() lowerCAmelCase__ = DataLoaderDispatcher(range(16 ) , batch_size=4 ) for idx, _ in enumerate(lowerCamelCase_ ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(lowerCamelCase_ ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 )
90
class a : """simple docstring""" def __init__( self : Tuple , lowerCamelCase : list ) -> None: __snake_case : str = set_counts __snake_case : Union[str, Any] = max(lowerCamelCase ) __snake_case : List[Any] = len(lowerCamelCase ) __snake_case : Tuple = [1] * num_sets __snake_case : Dict = list(range(lowerCamelCase ) ) def __snake_case ( self : str , lowerCamelCase : int , lowerCamelCase : int ) -> bool: __snake_case : List[Any] = self.get_parent(lowerCamelCase ) __snake_case : Tuple = self.get_parent(lowerCamelCase ) if src_parent == dst_parent: return False if self.ranks[dst_parent] >= self.ranks[src_parent]: self.set_counts[dst_parent] += self.set_counts[src_parent] __snake_case : List[str] = 0 __snake_case : List[Any] = dst_parent if self.ranks[dst_parent] == self.ranks[src_parent]: self.ranks[dst_parent] += 1 __snake_case : Dict = self.set_counts[dst_parent] else: self.set_counts[src_parent] += self.set_counts[dst_parent] __snake_case : Union[str, Any] = 0 __snake_case : Optional[int] = src_parent __snake_case : Tuple = self.set_counts[src_parent] __snake_case : str = max(self.max_set , lowerCamelCase ) return True def __snake_case ( self : int , lowerCamelCase : int ) -> int: if self.parents[disj_set] == disj_set: return disj_set __snake_case : Optional[int] = self.get_parent(self.parents[disj_set] ) return self.parents[disj_set]
81
0
"""simple docstring""" import unittest from transformers import AutoTokenizer, is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, slow if is_flax_available(): import jax.numpy as jnp from transformers import FlaxXLMRobertaModel @require_sentencepiece @require_tokenizers @require_flax class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' @slow def _SCREAMING_SNAKE_CASE ( self : Any ) -> Optional[Any]: A = FlaxXLMRobertaModel.from_pretrained('xlm-roberta-base' ) A = AutoTokenizer.from_pretrained('xlm-roberta-base' ) A = 'The dog is cute and lives in the garden house' A = jnp.array([tokenizer.encode(A_ )] ) A = (1, 12, 768) # batch_size, sequence_length, embedding_vector_dim A = jnp.array( [[-0.01_01, 0.12_18, -0.08_03, 0.08_01, 0.13_27, 0.07_76, -0.12_15, 0.23_83, 0.33_38, 0.31_06, 0.03_00, 0.02_52]] ) A = model(A_ )['last_hidden_state'] self.assertEqual(output.shape ,A_ ) # compare the actual values for a slice of last dim self.assertTrue(jnp.allclose(output[:, :, -1] ,A_ ,atol=1e-3 ) )
91
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class a (unittest.TestCase ): """simple docstring""" def __init__( self : int , lowerCamelCase : Any , lowerCamelCase : int=7 , lowerCamelCase : str=3 , lowerCamelCase : Optional[int]=30 , lowerCamelCase : Dict=400 , lowerCamelCase : str=True , lowerCamelCase : str=None , lowerCamelCase : Any=True , lowerCamelCase : Union[str, Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[str]=True , lowerCamelCase : Optional[int]=1 / 255 , lowerCamelCase : Any=True , ) -> str: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __snake_case : Optional[Any] = size if size is not None else {"shortest_edge": 18, "longest_edge": 1333} __snake_case : Optional[Any] = parent __snake_case : List[Any] = batch_size __snake_case : Optional[int] = num_channels __snake_case : str = min_resolution __snake_case : int = max_resolution __snake_case : int = do_resize __snake_case : Tuple = size __snake_case : Any = do_normalize __snake_case : int = image_mean __snake_case : Tuple = image_std __snake_case : Dict = do_rescale __snake_case : Optional[Any] = rescale_factor __snake_case : str = do_pad def __snake_case ( self : Any ) -> int: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def __snake_case ( self : Dict , lowerCamelCase : List[Any] , lowerCamelCase : Optional[Any]=False ) -> List[str]: if not batched: __snake_case : Dict = image_inputs[0] if isinstance(lowerCamelCase , Image.Image ): __snake_case , __snake_case : Dict = image.size else: __snake_case , __snake_case : List[str] = image.shape[1], image.shape[2] if w < h: __snake_case : Optional[int] = int(self.size["shortest_edge"] * h / w ) __snake_case : int = self.size["shortest_edge"] elif w > h: __snake_case : List[str] = self.size["shortest_edge"] __snake_case : Optional[Any] = int(self.size["shortest_edge"] * w / h ) else: __snake_case : List[Any] = self.size["shortest_edge"] __snake_case : Any = self.size["shortest_edge"] else: __snake_case : int = [] for image in image_inputs: __snake_case , __snake_case : List[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[0] )[0] __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class a (_lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ConditionalDetrImageProcessor if is_vision_available() else None def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : Optional[Any] = ConditionalDetrImageProcessingTester(self ) @property def __snake_case ( self : Any ) -> str: return self.image_processor_tester.prepare_image_processor_dict() def __snake_case ( self : Optional[Any] ) -> Optional[int]: __snake_case : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase , "image_mean" ) ) self.assertTrue(hasattr(lowerCamelCase , "image_std" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_normalize" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase , "size" ) ) def __snake_case ( self : Any ) -> Dict: __snake_case : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 1333} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) __snake_case : str = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=lowerCamelCase ) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> Dict: pass def __snake_case ( self : Tuple ) -> str: # Initialize image_processing __snake_case : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , Image.Image ) # Test not batched input __snake_case : int = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) __snake_case : Dict = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> str: # Initialize image_processing __snake_case : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __snake_case : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , numpify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , np.ndarray ) # Test not batched input __snake_case : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : List[Any] = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> List[str]: # Initialize image_processing __snake_case : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , torchify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , torch.Tensor ) # Test not batched input __snake_case : Optional[Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : List[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : int = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def __snake_case ( self : Any ) -> Optional[int]: # prepare image and target __snake_case : List[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : List[Any] = {"image_id": 39769, "annotations": target} # encode them __snake_case : List[str] = ConditionalDetrImageProcessor.from_pretrained("microsoft/conditional-detr-resnet-50" ) __snake_case : List[str] = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : Tuple = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Optional[int] = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : List[Any] = torch.tensor([58_87.96_00, 1_12_50.20_61, 48_93_53.84_38, 83_71_22.75_00, 14_79_67.51_56, 16_57_32.34_38] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : Optional[int] = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([0.55_03, 0.27_65, 0.06_04, 0.22_15] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Optional[Any] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : List[Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : Dict = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify orig_size __snake_case : int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Tuple = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) ) @slow def __snake_case ( self : str ) -> Tuple: # prepare image, target and masks_path __snake_case : Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : str = {"file_name": "000000039769.png", "image_id": 39769, "segments_info": target} __snake_case : Optional[int] = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them __snake_case : int = ConditionalDetrImageProcessor(format="coco_panoptic" ) __snake_case : str = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , masks_path=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : List[str] = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Dict = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : Any = torch.tensor([14_79_79.68_75, 16_55_27.04_69, 48_46_38.59_38, 1_12_92.93_75, 58_79.65_62, 76_34.11_47] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : str = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : Optional[Any] = torch.tensor([0.26_25, 0.54_37, 0.46_88, 0.86_25] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Tuple = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : Dict = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : int = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify masks __snake_case : List[Any] = 822873 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , lowerCamelCase ) # verify orig_size __snake_case : List[str] = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Any = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) )
81
0
'''simple docstring''' import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFAutoModel, is_tensorflow_text_available, is_tf_available from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.testing_utils import require_tensorflow_text, require_tf, slow if is_tf_available(): import tensorflow as tf if is_tensorflow_text_available(): from transformers.models.bert import TFBertTokenizer UpperCamelCase_ = ["""bert-base-uncased""", """bert-base-cased"""] UpperCamelCase_ = """hf-internal-testing/tiny-bert-tf-only""" if is_tf_available(): class __SCREAMING_SNAKE_CASE ( tf.keras.Model ): def __init__( self : Union[str, Any] , UpperCAmelCase__ : List[Any] ): '''simple docstring''' super().__init__() lowercase : Dict =tokenizer lowercase : str =AutoConfig.from_pretrained(UpperCAmelCase__ ) lowercase : str =TFAutoModel.from_config(UpperCAmelCase__ ) def lowerCamelCase_ ( self : Any , UpperCAmelCase__ : List[Any] ): '''simple docstring''' lowercase : List[Any] =self.tokenizer(UpperCAmelCase__ ) lowercase : int =self.bert(**UpperCAmelCase__ ) return out["pooler_output"] @require_tf @require_tensorflow_text class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): def lowerCamelCase_ ( self : int ): '''simple docstring''' super().setUp() lowercase : str =[ BertTokenizer.from_pretrained(UpperCAmelCase__ ) for checkpoint in (TOKENIZER_CHECKPOINTS * 2) ] # repeat for when fast_bert_tokenizer=false lowercase : Tuple =[TFBertTokenizer.from_pretrained(UpperCAmelCase__ ) for checkpoint in TOKENIZER_CHECKPOINTS] + [ TFBertTokenizer.from_pretrained(UpperCAmelCase__ , use_fast_bert_tokenizer=UpperCAmelCase__ ) for checkpoint in TOKENIZER_CHECKPOINTS ] assert len(self.tokenizers ) == len(self.tf_tokenizers ) lowercase : Any =[ '''This is a straightforward English test sentence.''', '''This one has some weird characters\rto\nsee\r\nif those\u00E9break things.''', '''Now we\'re going to add some Chinese: 一 二 三 一二三''', '''And some much more rare Chinese: 齉 堃 齉堃''', '''Je vais aussi écrire en français pour tester les accents''', '''Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ''', ] lowercase : Optional[Any] =list(zip(self.test_sentences , self.test_sentences[::-1] ) ) def lowerCamelCase_ ( self : str ): '''simple docstring''' for tokenizer, tf_tokenizer in zip(self.tokenizers , self.tf_tokenizers ): for test_inputs in (self.test_sentences, self.paired_sentences): lowercase : Tuple =tokenizer(UpperCAmelCase__ , return_tensors='''tf''' , padding='''longest''' ) lowercase : Optional[int] =tf_tokenizer(UpperCAmelCase__ ) for key in python_outputs.keys(): self.assertTrue(tf.reduce_all(python_outputs[key].shape == tf_outputs[key].shape ) ) self.assertTrue(tf.reduce_all(tf.cast(python_outputs[key] , tf.intaa ) == tf_outputs[key] ) ) @slow def lowerCamelCase_ ( self : Dict ): '''simple docstring''' for tf_tokenizer in self.tf_tokenizers: lowercase : int =tf_tokenizer(self.paired_sentences ) lowercase : Dict =tf_tokenizer( text=[sentence[0] for sentence in self.paired_sentences] , text_pair=[sentence[1] for sentence in self.paired_sentences] , ) for key in merged_outputs.keys(): self.assertTrue(tf.reduce_all(tf.cast(merged_outputs[key] , tf.intaa ) == separated_outputs[key] ) ) @slow def lowerCamelCase_ ( self : List[Any] ): '''simple docstring''' for tf_tokenizer in self.tf_tokenizers: lowercase : Any =tf.function(UpperCAmelCase__ ) for test_inputs in (self.test_sentences, self.paired_sentences): lowercase : int =tf.constant(UpperCAmelCase__ ) lowercase : List[str] =compiled_tokenizer(UpperCAmelCase__ ) lowercase : Tuple =tf_tokenizer(UpperCAmelCase__ ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def lowerCamelCase_ ( self : List[Any] ): '''simple docstring''' for tf_tokenizer in self.tf_tokenizers: lowercase : str =ModelToSave(tokenizer=UpperCAmelCase__ ) lowercase : List[str] =tf.convert_to_tensor(self.test_sentences ) lowercase : List[Any] =model(UpperCAmelCase__ ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: lowercase : List[str] =Path(UpperCAmelCase__ ) / '''saved.model''' model.save(UpperCAmelCase__ ) lowercase : Dict =tf.keras.models.load_model(UpperCAmelCase__ ) lowercase : Dict =loaded_model(UpperCAmelCase__ ) # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertLessEqual(tf.reduce_max(tf.abs(out - loaded_output ) ) , 1E-5 )
92
from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from ... import AutoBackbone from ...modeling_outputs import SemanticSegmenterOutput from ...modeling_utils import PreTrainedModel from ...utils import add_start_docstrings, add_start_docstrings_to_model_forward, replace_return_docstrings from ...utils.backbone_utils import BackboneMixin from .configuration_upernet import UperNetConfig _snake_case : Optional[Any] = [ "openmmlab/upernet-convnext-tiny", # See all UperNet models at https://huggingface.co/models?filter=upernet ] # General docstring _snake_case : Dict = "UperNetConfig" class a (nn.Module ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : Union[int, Tuple[int, int]] , lowerCamelCase : Union[int, Tuple[int, int], str] = 0 , lowerCamelCase : bool = False , lowerCamelCase : Union[int, Tuple[int, int]] = 1 , ) -> None: super().__init__() __snake_case : Union[str, Any] = nn.Convad( in_channels=lowerCamelCase , out_channels=lowerCamelCase , kernel_size=lowerCamelCase , padding=lowerCamelCase , bias=lowerCamelCase , dilation=lowerCamelCase , ) __snake_case : Dict = nn.BatchNormad(lowerCamelCase ) __snake_case : List[Any] = nn.ReLU() def __snake_case ( self : List[Any] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : Dict = self.conv(lowerCamelCase ) __snake_case : int = self.batch_norm(lowerCamelCase ) __snake_case : Optional[Any] = self.activation(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : str , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : int ) -> None: super().__init__() __snake_case : Tuple = [ nn.AdaptiveAvgPoolad(lowerCamelCase ), UperNetConvModule(lowerCamelCase , lowerCamelCase , kernel_size=1 ), ] for i, layer in enumerate(self.layers ): self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : Dict , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : List[str] = input for layer in self.layers: __snake_case : Tuple = layer(lowerCamelCase ) return hidden_state class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Tuple[int, ...] , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : bool ) -> None: super().__init__() __snake_case : Dict = pool_scales __snake_case : List[str] = align_corners __snake_case : List[Any] = in_channels __snake_case : str = channels __snake_case : Optional[Any] = [] for i, pool_scale in enumerate(lowerCamelCase ): __snake_case : Union[str, Any] = UperNetPyramidPoolingBlock(pool_scale=lowerCamelCase , in_channels=lowerCamelCase , channels=lowerCamelCase ) self.blocks.append(lowerCamelCase ) self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> List[torch.Tensor]: __snake_case : Tuple = [] for ppm in self.blocks: __snake_case : Any = ppm(lowerCamelCase ) __snake_case : List[Any] = nn.functional.interpolate( lowerCamelCase , size=x.size()[2:] , mode="bilinear" , align_corners=self.align_corners ) ppm_outs.append(lowerCamelCase ) return ppm_outs class a (nn.Module ): """simple docstring""" def __init__( self : int , lowerCamelCase : str , lowerCamelCase : str ) -> List[str]: super().__init__() __snake_case : Dict = config __snake_case : List[str] = config.pool_scales # e.g. (1, 2, 3, 6) __snake_case : Tuple = in_channels __snake_case : str = config.hidden_size __snake_case : List[str] = False __snake_case : Any = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) # PSP Module __snake_case : Tuple = UperNetPyramidPoolingModule( self.pool_scales , self.in_channels[-1] , self.channels , align_corners=self.align_corners , ) __snake_case : List[str] = UperNetConvModule( self.in_channels[-1] + len(self.pool_scales ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) # FPN Module __snake_case : List[Any] = nn.ModuleList() __snake_case : Dict = nn.ModuleList() for in_channels in self.in_channels[:-1]: # skip the top layer __snake_case : Union[str, Any] = UperNetConvModule(lowerCamelCase , self.channels , kernel_size=1 ) __snake_case : Optional[int] = UperNetConvModule(self.channels , self.channels , kernel_size=3 , padding=1 ) self.lateral_convs.append(lowerCamelCase ) self.fpn_convs.append(lowerCamelCase ) __snake_case : int = UperNetConvModule( len(self.in_channels ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) def __snake_case ( self : List[str] ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> str: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : List[Any] , lowerCamelCase : Tuple ) -> Optional[int]: __snake_case : str = inputs[-1] __snake_case : int = [x] psp_outs.extend(self.psp_modules(lowerCamelCase ) ) __snake_case : Tuple = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Union[str, Any] = self.bottleneck(lowerCamelCase ) return output def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # build laterals __snake_case : Any = [lateral_conv(encoder_hidden_states[i] ) for i, lateral_conv in enumerate(self.lateral_convs )] laterals.append(self.psp_forward(lowerCamelCase ) ) # build top-down path __snake_case : Dict = len(lowerCamelCase ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Union[str, Any] = laterals[i - 1].shape[2:] __snake_case : Optional[int] = laterals[i - 1] + nn.functional.interpolate( laterals[i] , size=lowerCamelCase , mode="bilinear" , align_corners=self.align_corners ) # build outputs __snake_case : str = [self.fpn_convs[i](laterals[i] ) for i in range(used_backbone_levels - 1 )] # append psp feature fpn_outs.append(laterals[-1] ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Tuple = nn.functional.interpolate( fpn_outs[i] , size=fpn_outs[0].shape[2:] , mode="bilinear" , align_corners=self.align_corners ) __snake_case : str = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Optional[Any] = self.fpn_bottleneck(lowerCamelCase ) __snake_case : Tuple = self.classifier(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Optional[int] , lowerCamelCase : int = 2 , lowerCamelCase : int = 3 , lowerCamelCase : Union[int, Tuple[int, int]] = 1 ) -> None: super().__init__() __snake_case : List[Any] = config __snake_case : List[str] = config.auxiliary_in_channels __snake_case : List[Any] = config.auxiliary_channels __snake_case : Tuple = config.auxiliary_num_convs __snake_case : int = config.auxiliary_concat_input __snake_case : Optional[int] = in_index __snake_case : Tuple = (kernel_size // 2) * dilation __snake_case : Optional[int] = [] convs.append( UperNetConvModule( self.in_channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) for i in range(self.num_convs - 1 ): convs.append( UperNetConvModule( self.channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) if self.num_convs == 0: __snake_case : Union[str, Any] = nn.Identity() else: __snake_case : Any = nn.Sequential(*lowerCamelCase ) if self.concat_input: __snake_case : int = UperNetConvModule( self.in_channels + self.channels , self.channels , kernel_size=lowerCamelCase , padding=kernel_size // 2 ) __snake_case : Optional[Any] = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) def __snake_case ( self : Dict ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : Tuple ) -> Optional[int]: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : Optional[int] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # just take the relevant feature maps __snake_case : List[str] = encoder_hidden_states[self.in_index] __snake_case : Optional[Any] = self.convs(lowerCamelCase ) if self.concat_input: __snake_case : Optional[Any] = self.conv_cat(torch.cat([hidden_states, output] , dim=1 ) ) __snake_case : Union[str, Any] = self.classifier(lowerCamelCase ) return output class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = UperNetConfig __UpperCAmelCase : int = "pixel_values" __UpperCAmelCase : str = True def __snake_case ( self : Union[str, Any] , lowerCamelCase : List[Any] ) -> List[Any]: if isinstance(lowerCamelCase , lowerCamelCase ): module.backbone.init_weights() module.decode_head.init_weights() module.auxiliary_head.init_weights() def __snake_case ( self : Optional[Any] ) -> List[str]: self.backbone.init_weights() self.decode_head.init_weights() self.auxiliary_head.init_weights() def __snake_case ( self : List[Any] , lowerCamelCase : Any , lowerCamelCase : Optional[Any]=False ) -> Dict: if isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : Union[str, Any] = value _snake_case : Dict = R"\n Parameters:\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n config ([`UperNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" _snake_case : Tuple = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using\n [`AutoImageProcessor`]. See [`SegformerImageProcessor.__call__`] for details.\n output_attentions (`bool`, *optional*):\n Whether or not to return the attentions tensors of all attention layers in case the backbone has them. See\n `attentions` under returned tensors for more detail.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers of the backbone. See `hidden_states` under\n returned tensors for more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n" @add_start_docstrings( "UperNet framework leveraging any vision backbone e.g. for ADE20k, CityScapes." , _lowerCAmelCase , ) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Dict , lowerCamelCase : int ) -> Optional[int]: super().__init__(lowerCamelCase ) __snake_case : Any = AutoBackbone.from_config(config.backbone_config ) # Semantic segmentation head(s) __snake_case : Union[str, Any] = UperNetHead(lowerCamelCase , in_channels=self.backbone.channels ) __snake_case : Any = UperNetFCNHead(lowerCamelCase ) if config.use_auxiliary_head else None # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UPERNET_INPUTS_DOCSTRING.format("batch_size, sequence_length" ) ) @replace_return_docstrings(output_type=lowerCamelCase , config_class=_CONFIG_FOR_DOC ) def __snake_case ( self : Union[str, Any] , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , ) -> Union[tuple, SemanticSegmenterOutput]: __snake_case : Any = return_dict if return_dict is not None else self.config.use_return_dict __snake_case : int = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __snake_case : str = output_attentions if output_attentions is not None else self.config.output_attentions __snake_case : Tuple = self.backbone.forward_with_filtered_kwargs( lowerCamelCase , output_hidden_states=lowerCamelCase , output_attentions=lowerCamelCase ) __snake_case : List[Any] = outputs.feature_maps __snake_case : List[Any] = self.decode_head(lowerCamelCase ) __snake_case : List[str] = nn.functional.interpolate(lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : Optional[int] = None if self.auxiliary_head is not None: __snake_case : Dict = self.auxiliary_head(lowerCamelCase ) __snake_case : Dict = nn.functional.interpolate( lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : int = None if labels is not None: if self.config.num_labels == 1: raise ValueError("The number of labels should be greater than one" ) else: # compute weighted loss __snake_case : Any = CrossEntropyLoss(ignore_index=self.config.loss_ignore_index ) __snake_case : Union[str, Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = main_loss + self.config.auxiliary_loss_weight * auxiliary_loss if not return_dict: if output_hidden_states: __snake_case : Any = (logits,) + outputs[1:] else: __snake_case : str = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return SemanticSegmenterOutput( loss=lowerCamelCase , logits=lowerCamelCase , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
81
0
"""simple docstring""" __A = tuple[float, float, float] __A = tuple[float, float, float] def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Vectorad: """simple docstring""" lowerCAmelCase__ :Tuple = end_pointa[0] - end_pointa[0] lowerCAmelCase__ :List[str] = end_pointa[1] - end_pointa[1] lowerCAmelCase__ :Any = end_pointa[2] - end_pointa[2] return (x, y, z) def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Vectorad: """simple docstring""" lowerCAmelCase__ :str = ab[1] * ac[2] - ab[2] * ac[1] # *i lowerCAmelCase__ :int = (ab[0] * ac[2] - ab[2] * ac[0]) * -1 # *j lowerCAmelCase__ :Optional[int] = ab[0] * ac[1] - ab[1] * ac[0] # *k return (x, y, z) def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->bool: """simple docstring""" return tuple(round(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for x in vector ) == (0, 0, 0) def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 10 ) ->bool: """simple docstring""" lowerCAmelCase__ :List[str] = create_vector(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :str = create_vector(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) return is_zero_vector(get_ad_vectors_cross(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE )
93
from __future__ import annotations def lowerCAmelCase_ ( __lowerCamelCase ): if not nums: return 0 __snake_case : Any = nums[0] __snake_case : str = 0 for num in nums[1:]: __snake_case , __snake_case : List[str] = ( max_excluding + num, max(__lowerCamelCase , __lowerCamelCase ), ) return max(__lowerCamelCase , __lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
81
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available SCREAMING_SNAKE_CASE = {'configuration_glpn': ['GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GLPNConfig']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ['GLPNFeatureExtractor'] SCREAMING_SNAKE_CASE = ['GLPNImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ 'GLPN_PRETRAINED_MODEL_ARCHIVE_LIST', 'GLPNForDepthEstimation', 'GLPNLayer', 'GLPNModel', 'GLPNPreTrainedModel', ] if TYPE_CHECKING: from .configuration_glpn import GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP, GLPNConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_glpn import GLPNFeatureExtractor from .image_processing_glpn import GLPNImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_glpn import ( GLPN_PRETRAINED_MODEL_ARCHIVE_LIST, GLPNForDepthEstimation, GLPNLayer, GLPNModel, GLPNPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
94
from __future__ import annotations from typing import Any def lowerCAmelCase_ ( __lowerCamelCase ): create_state_space_tree(__lowerCamelCase , [] , 0 ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): if index == len(__lowerCamelCase ): print(__lowerCamelCase ) return create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.pop() if __name__ == "__main__": _snake_case : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["A", "B", "C"]) generate_all_subsequences(seq)
81
0
"""simple docstring""" import random def snake_case ( A__ ,A__ ): UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : List[str] = [], [], [] for element in data: if element < pivot: less.append(A__ ) elif element > pivot: greater.append(A__ ) else: equal.append(A__ ) return less, equal, greater def snake_case ( A__ ,A__ ): # index = len(items) // 2 when trying to find the median # (value of index when items is sorted) # invalid input if index >= len(A__ ) or index < 0: return None UpperCAmelCase_ : str = items[random.randint(0 ,len(A__ ) - 1 )] UpperCAmelCase_ : Dict = 0 UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : List[Any] = _partition(A__ ,A__ ) UpperCAmelCase_ : Tuple = len(A__ ) UpperCAmelCase_ : List[str] = len(A__ ) # index is the pivot if m <= index < m + count: return pivot # must be in smaller elif m > index: return quick_select(A__ ,A__ ) # must be in larger else: return quick_select(A__ ,index - (m + count) )
95
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case , __snake_case : List[Any] = image.size __snake_case , __snake_case : Tuple = (x - x % 3_2 for x in (w, h)) # resize to integer multiple of 32 __snake_case : str = image.resize((w, h) , resample=PIL_INTERPOLATION["lanczos"] ) __snake_case : int = np.array(__lowerCamelCase ).astype(np.floataa ) / 2_5_5.0 __snake_case : Union[str, Any] = image[None].transpose(0 , 3 , 1 , 2 ) __snake_case : Union[str, Any] = torch.from_numpy(__lowerCamelCase ) return 2.0 * image - 1.0 class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : VQModel , lowerCamelCase : UNetaDModel , lowerCamelCase : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ) -> Union[str, Any]: super().__init__() self.register_modules(vqvae=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase ) @torch.no_grad() def __call__( self : List[str] , lowerCamelCase : Union[torch.Tensor, PIL.Image.Image] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : Optional[int] = 100 , lowerCamelCase : Optional[float] = 0.0 , lowerCamelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , ) -> Union[Tuple, ImagePipelineOutput]: if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : Any = 1 elif isinstance(lowerCamelCase , torch.Tensor ): __snake_case : Any = image.shape[0] else: raise ValueError(F'`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(lowerCamelCase )}' ) if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : List[Any] = preprocess(lowerCamelCase ) __snake_case , __snake_case : int = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image __snake_case : str = (batch_size, self.unet.config.in_channels // 2, height, width) __snake_case : str = next(self.unet.parameters() ).dtype __snake_case : Tuple = randn_tensor(lowerCamelCase , generator=lowerCamelCase , device=self.device , dtype=lowerCamelCase ) __snake_case : List[Any] = image.to(device=self.device , dtype=lowerCamelCase ) # set timesteps and move to the correct device self.scheduler.set_timesteps(lowerCamelCase , device=self.device ) __snake_case : str = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler __snake_case : Dict = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] __snake_case : Union[str, Any] = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __snake_case : int = {} if accepts_eta: __snake_case : List[str] = eta for t in self.progress_bar(lowerCamelCase ): # concat latents and low resolution image in the channel dimension. __snake_case : Union[str, Any] = torch.cat([latents, image] , dim=1 ) __snake_case : Optional[Any] = self.scheduler.scale_model_input(lowerCamelCase , lowerCamelCase ) # predict the noise residual __snake_case : int = self.unet(lowerCamelCase , lowerCamelCase ).sample # compute the previous noisy sample x_t -> x_t-1 __snake_case : Union[str, Any] = self.scheduler.step(lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase ).prev_sample # decode the image latents with the VQVAE __snake_case : List[Any] = self.vqvae.decode(lowerCamelCase ).sample __snake_case : Dict = torch.clamp(lowerCamelCase , -1.0 , 1.0 ) __snake_case : Any = image / 2 + 0.5 __snake_case : Tuple = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __snake_case : Tuple = self.numpy_to_pil(lowerCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCamelCase )
81
0
"""simple docstring""" from typing import List, Optional, Union import torch from transformers import ( XLMRobertaTokenizer, ) from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) from .text_encoder import MultilingualCLIP __lowerCamelCase = logging.get_logger(__name__) # pylint: disable=invalid-name __lowerCamelCase = '\n Examples:\n ```py\n >>> from diffusers import KandinskyPipeline, KandinskyPriorPipeline\n >>> import torch\n\n >>> pipe_prior = KandinskyPriorPipeline.from_pretrained("kandinsky-community/Kandinsky-2-1-prior")\n >>> pipe_prior.to("cuda")\n\n >>> prompt = "red cat, 4k photo"\n >>> out = pipe_prior(prompt)\n >>> image_emb = out.image_embeds\n >>> negative_image_emb = out.negative_image_embeds\n\n >>> pipe = KandinskyPipeline.from_pretrained("kandinsky-community/kandinsky-2-1")\n >>> pipe.to("cuda")\n\n >>> image = pipe(\n ... prompt,\n ... image_embeds=image_emb,\n ... negative_image_embeds=negative_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=100,\n ... ).images\n\n >>> image[0].save("cat.png")\n ```\n' def a ( __UpperCAmelCase : Dict , __UpperCAmelCase : Tuple , __UpperCAmelCase : Any=8 ) -> Optional[Any]: __magic_name__: Optional[Any] = h // scale_factor**2 if h % scale_factor**2 != 0: new_h += 1 __magic_name__: Dict = w // scale_factor**2 if w % scale_factor**2 != 0: new_w += 1 return new_h * scale_factor, new_w * scale_factor class __A ( SCREAMING_SNAKE_CASE_ ): def __init__( self : Dict , __snake_case : MultilingualCLIP , __snake_case : XLMRobertaTokenizer , __snake_case : UNetaDConditionModel , __snake_case : Union[DDIMScheduler, DDPMScheduler] , __snake_case : VQModel , ) -> Optional[Any]: super().__init__() self.register_modules( text_encoder=__snake_case , tokenizer=__snake_case , unet=__snake_case , scheduler=__snake_case , movq=__snake_case , ) __magic_name__: Union[str, Any] = 2 ** (len(self.movq.config.block_out_channels ) - 1) def lowerCamelCase__ ( self : Dict , __snake_case : Union[str, Any] , __snake_case : List[str] , __snake_case : List[str] , __snake_case : str , __snake_case : Any , __snake_case : List[str] ) -> Tuple: if latents is None: __magic_name__: int = randn_tensor(__snake_case , generator=__snake_case , device=__snake_case , dtype=__snake_case ) else: if latents.shape != shape: raise ValueError(F'Unexpected latents shape, got {latents.shape}, expected {shape}' ) __magic_name__: List[Any] = latents.to(__snake_case ) __magic_name__: Any = latents * scheduler.init_noise_sigma return latents def lowerCamelCase__ ( self : Optional[Any] , __snake_case : int , __snake_case : Optional[Any] , __snake_case : str , __snake_case : Union[str, Any] , __snake_case : Dict=None , ) -> Optional[Any]: __magic_name__: Any = len(__snake_case ) if isinstance(__snake_case , __snake_case ) else 1 # get prompt text embeddings __magic_name__: Any = self.tokenizer( __snake_case , padding="""max_length""" , truncation=__snake_case , max_length=7_7 , return_attention_mask=__snake_case , add_special_tokens=__snake_case , return_tensors="""pt""" , ) __magic_name__: Tuple = text_inputs.input_ids __magic_name__: Optional[Any] = self.tokenizer(__snake_case , padding="""longest""" , return_tensors="""pt""" ).input_ids if untruncated_ids.shape[-1] >= text_input_ids.shape[-1] and not torch.equal(__snake_case , __snake_case ): __magic_name__: Dict = self.tokenizer.batch_decode(untruncated_ids[:, self.tokenizer.model_max_length - 1 : -1] ) logger.warning( """The following part of your input was truncated because CLIP can only handle sequences up to""" F' {self.tokenizer.model_max_length} tokens: {removed_text}' ) __magic_name__: Optional[int] = text_input_ids.to(__snake_case ) __magic_name__: Optional[Any] = text_inputs.attention_mask.to(__snake_case ) __magic_name__, __magic_name__: List[str] = self.text_encoder( input_ids=__snake_case , attention_mask=__snake_case ) __magic_name__: List[Any] = prompt_embeds.repeat_interleave(__snake_case , dim=0 ) __magic_name__: Any = text_encoder_hidden_states.repeat_interleave(__snake_case , dim=0 ) __magic_name__: Optional[int] = text_mask.repeat_interleave(__snake_case , dim=0 ) if do_classifier_free_guidance: __magic_name__: List[str] if negative_prompt is None: __magic_name__: List[Any] = [""""""] * batch_size elif type(__snake_case ) is not type(__snake_case ): raise TypeError( F'`negative_prompt` should be the same type to `prompt`, but got {type(__snake_case )} !=' F' {type(__snake_case )}.' ) elif isinstance(__snake_case , __snake_case ): __magic_name__: str = [negative_prompt] elif batch_size != len(__snake_case ): raise ValueError( F'`negative_prompt`: {negative_prompt} has batch size {len(__snake_case )}, but `prompt`:' F' {prompt} has batch size {batch_size}. Please make sure that passed `negative_prompt` matches' """ the batch size of `prompt`.""" ) else: __magic_name__: Any = negative_prompt __magic_name__: Union[str, Any] = self.tokenizer( __snake_case , padding="""max_length""" , max_length=7_7 , truncation=__snake_case , return_attention_mask=__snake_case , add_special_tokens=__snake_case , return_tensors="""pt""" , ) __magic_name__: Optional[Any] = uncond_input.input_ids.to(__snake_case ) __magic_name__: Union[str, Any] = uncond_input.attention_mask.to(__snake_case ) __magic_name__, __magic_name__: Any = self.text_encoder( input_ids=__snake_case , attention_mask=__snake_case ) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method __magic_name__: List[Any] = negative_prompt_embeds.shape[1] __magic_name__: Union[str, Any] = negative_prompt_embeds.repeat(1 , __snake_case ) __magic_name__: int = negative_prompt_embeds.view(batch_size * num_images_per_prompt , __snake_case ) __magic_name__: int = uncond_text_encoder_hidden_states.shape[1] __magic_name__: str = uncond_text_encoder_hidden_states.repeat(1 , __snake_case , 1 ) __magic_name__: Optional[Any] = uncond_text_encoder_hidden_states.view( batch_size * num_images_per_prompt , __snake_case , -1 ) __magic_name__: int = uncond_text_mask.repeat_interleave(__snake_case , dim=0 ) # done duplicates # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes __magic_name__: Any = torch.cat([negative_prompt_embeds, prompt_embeds] ) __magic_name__: List[Any] = torch.cat([uncond_text_encoder_hidden_states, text_encoder_hidden_states] ) __magic_name__: Union[str, Any] = torch.cat([uncond_text_mask, text_mask] ) return prompt_embeds, text_encoder_hidden_states, text_mask def lowerCamelCase__ ( self : Optional[int] , __snake_case : List[Any]=0 ) -> Dict: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("""Please install accelerate via `pip install accelerate`""" ) __magic_name__: Any = torch.device(F'cuda:{gpu_id}' ) __magic_name__: List[str] = [ self.unet, self.text_encoder, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(__snake_case , __snake_case ) def lowerCamelCase__ ( self : Tuple , __snake_case : int=0 ) -> Optional[Any]: if is_accelerate_available() and is_accelerate_version(""">=""" , """0.17.0.dev0""" ): from accelerate import cpu_offload_with_hook else: raise ImportError("""`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.""" ) __magic_name__: str = torch.device(F'cuda:{gpu_id}' ) if self.device.type != "cpu": self.to("""cpu""" , silence_dtype_warnings=__snake_case ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) __magic_name__: Dict = None for cpu_offloaded_model in [self.text_encoder, self.unet, self.movq]: __magic_name__, __magic_name__: int = cpu_offload_with_hook(__snake_case , __snake_case , prev_module_hook=__snake_case ) if self.safety_checker is not None: __magic_name__, __magic_name__: List[str] = cpu_offload_with_hook(self.safety_checker , __snake_case , prev_module_hook=__snake_case ) # We'll offload the last model manually. __magic_name__: Dict = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def lowerCamelCase__ ( self : Dict ) -> Optional[Any]: if not hasattr(self.unet , """_hf_hook""" ): return self.device for module in self.unet.modules(): if ( hasattr(__snake_case , """_hf_hook""" ) and hasattr(module._hf_hook , """execution_device""" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(__snake_case ) def __call__( self : Optional[int] , __snake_case : Union[str, List[str]] , __snake_case : Union[torch.FloatTensor, List[torch.FloatTensor]] , __snake_case : Union[torch.FloatTensor, List[torch.FloatTensor]] , __snake_case : Optional[Union[str, List[str]]] = None , __snake_case : int = 5_1_2 , __snake_case : int = 5_1_2 , __snake_case : int = 1_0_0 , __snake_case : float = 4.0 , __snake_case : int = 1 , __snake_case : Optional[Union[torch.Generator, List[torch.Generator]]] = None , __snake_case : Optional[torch.FloatTensor] = None , __snake_case : Optional[str] = "pil" , __snake_case : bool = True , ) -> Any: if isinstance(__snake_case , __snake_case ): __magic_name__: Optional[int] = 1 elif isinstance(__snake_case , __snake_case ): __magic_name__: List[Any] = len(__snake_case ) else: raise ValueError(F'`prompt` has to be of type `str` or `list` but is {type(__snake_case )}' ) __magic_name__: List[Any] = self._execution_device __magic_name__: Dict = batch_size * num_images_per_prompt __magic_name__: Union[str, Any] = guidance_scale > 1.0 __magic_name__, __magic_name__, __magic_name__: str = self._encode_prompt( __snake_case , __snake_case , __snake_case , __snake_case , __snake_case ) if isinstance(__snake_case , __snake_case ): __magic_name__: List[str] = torch.cat(__snake_case , dim=0 ) if isinstance(__snake_case , __snake_case ): __magic_name__: int = torch.cat(__snake_case , dim=0 ) if do_classifier_free_guidance: __magic_name__: Optional[Any] = image_embeds.repeat_interleave(__snake_case , dim=0 ) __magic_name__: Dict = negative_image_embeds.repeat_interleave(__snake_case , dim=0 ) __magic_name__: List[Any] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to( dtype=prompt_embeds.dtype , device=__snake_case ) self.scheduler.set_timesteps(__snake_case , device=__snake_case ) __magic_name__: Optional[Any] = self.scheduler.timesteps __magic_name__: Dict = self.unet.config.in_channels __magic_name__, __magic_name__: Union[str, Any] = get_new_h_w(__snake_case , __snake_case , self.movq_scale_factor ) # create initial latent __magic_name__: Any = self.prepare_latents( (batch_size, num_channels_latents, height, width) , text_encoder_hidden_states.dtype , __snake_case , __snake_case , __snake_case , self.scheduler , ) for i, t in enumerate(self.progress_bar(__snake_case ) ): # expand the latents if we are doing classifier free guidance __magic_name__: List[Any] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents __magic_name__: Union[str, Any] = {"""text_embeds""": prompt_embeds, """image_embeds""": image_embeds} __magic_name__: Union[str, Any] = self.unet( sample=__snake_case , timestep=__snake_case , encoder_hidden_states=__snake_case , added_cond_kwargs=__snake_case , return_dict=__snake_case , )[0] if do_classifier_free_guidance: __magic_name__, __magic_name__: Union[str, Any] = noise_pred.split(latents.shape[1] , dim=1 ) __magic_name__, __magic_name__: Any = noise_pred.chunk(2 ) __magic_name__, __magic_name__: List[Any] = variance_pred.chunk(2 ) __magic_name__: Dict = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) __magic_name__: str = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , """variance_type""" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): __magic_name__, __magic_name__: Optional[Any] = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 __magic_name__: List[str] = self.scheduler.step( __snake_case , __snake_case , __snake_case , generator=__snake_case , ).prev_sample # post-processing __magic_name__: Dict = self.movq.decode(__snake_case , force_not_quantize=__snake_case )["""sample"""] if output_type not in ["pt", "np", "pil"]: raise ValueError(F'Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}' ) if output_type in ["np", "pil"]: __magic_name__: Optional[int] = image * 0.5 + 0.5 __magic_name__: Dict = image.clamp(0 , 1 ) __magic_name__: List[str] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": __magic_name__: List[Any] = self.numpy_to_pil(__snake_case ) if not return_dict: return (image,) return ImagePipelineOutput(images=__snake_case )
96
import gc import unittest import torch from parameterized import parameterized from diffusers import AutoencoderKL from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = AutoencoderKL __UpperCAmelCase : Optional[Any] = "sample" __UpperCAmelCase : Optional[int] = 1e-2 @property def __snake_case ( self : Dict ) -> Optional[Any]: __snake_case : Optional[Any] = 4 __snake_case : Tuple = 3 __snake_case : List[str] = (32, 32) __snake_case : str = floats_tensor((batch_size, num_channels) + sizes ).to(lowerCamelCase ) return {"sample": image} @property def __snake_case ( self : Union[str, Any] ) -> Tuple: return (3, 32, 32) @property def __snake_case ( self : int ) -> int: return (3, 32, 32) def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Optional[Any] = { "block_out_channels": [32, 64], "in_channels": 3, "out_channels": 3, "down_block_types": ["DownEncoderBlock2D", "DownEncoderBlock2D"], "up_block_types": ["UpDecoderBlock2D", "UpDecoderBlock2D"], "latent_channels": 4, } __snake_case : Any = self.dummy_input return init_dict, inputs_dict def __snake_case ( self : str ) -> Dict: pass def __snake_case ( self : Tuple ) -> List[str]: pass @unittest.skipIf(torch_device == "mps" , "Gradient checkpointing skipped on MPS" ) def __snake_case ( self : Any ) -> Optional[Any]: # enable deterministic behavior for gradient checkpointing __snake_case , __snake_case : int = self.prepare_init_args_and_inputs_for_common() __snake_case : str = self.model_class(**lowerCamelCase ) model.to(lowerCamelCase ) assert not model.is_gradient_checkpointing and model.training __snake_case : str = model(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model.zero_grad() __snake_case : Any = torch.randn_like(lowerCamelCase ) __snake_case : str = (out - labels).mean() loss.backward() # re-instantiate the model now enabling gradient checkpointing __snake_case : Optional[int] = self.model_class(**lowerCamelCase ) # clone model model_a.load_state_dict(model.state_dict() ) model_a.to(lowerCamelCase ) model_a.enable_gradient_checkpointing() assert model_a.is_gradient_checkpointing and model_a.training __snake_case : int = model_a(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model_a.zero_grad() __snake_case : Union[str, Any] = (out_a - labels).mean() loss_a.backward() # compare the output and parameters gradients self.assertTrue((loss - loss_a).abs() < 1E-5 ) __snake_case : Optional[int] = dict(model.named_parameters() ) __snake_case : List[Any] = dict(model_a.named_parameters() ) for name, param in named_params.items(): self.assertTrue(torch_all_close(param.grad.data , named_params_a[name].grad.data , atol=5E-5 ) ) def __snake_case ( self : List[Any] ) -> Optional[int]: __snake_case , __snake_case : Optional[Any] = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" , output_loading_info=lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(lowerCamelCase ) __snake_case : Optional[Any] = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def __snake_case ( self : Optional[Any] ) -> Union[str, Any]: __snake_case : Tuple = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" ) __snake_case : Dict = model.to(lowerCamelCase ) model.eval() if torch_device == "mps": __snake_case : int = torch.manual_seed(0 ) else: __snake_case : str = torch.Generator(device=lowerCamelCase ).manual_seed(0 ) __snake_case : List[str] = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) __snake_case : Union[str, Any] = image.to(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , sample_posterior=lowerCamelCase , generator=lowerCamelCase ).sample __snake_case : List[Any] = output[0, -1, -3:, -3:].flatten().cpu() # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. if torch_device == "mps": __snake_case : Union[str, Any] = torch.tensor( [ -4.0078E-01, -3.8323E-04, -1.2681E-01, -1.1462E-01, 2.0095E-01, 1.0893E-01, -8.8247E-02, -3.0361E-01, -9.8644E-03, ] ) elif torch_device == "cpu": __snake_case : Tuple = torch.tensor( [-0.13_52, 0.08_78, 0.04_19, -0.08_18, -0.10_69, 0.06_88, -0.14_58, -0.44_46, -0.00_26] ) else: __snake_case : List[str] = torch.tensor( [-0.24_21, 0.46_42, 0.25_07, -0.04_38, 0.06_82, 0.31_60, -0.20_18, -0.07_27, 0.24_85] ) self.assertTrue(torch_all_close(lowerCamelCase , lowerCamelCase , rtol=1E-2 ) ) @slow class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : int , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] ) -> List[str]: return F'gaussian_noise_s={seed}_shape={"_".join([str(lowerCamelCase ) for s in shape] )}.npy' def __snake_case ( self : List[Any] ) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __snake_case ( self : Tuple , lowerCamelCase : List[Any]=0 , lowerCamelCase : Tuple=(4, 3, 512, 512) , lowerCamelCase : Optional[int]=False ) -> str: __snake_case : List[Any] = torch.floataa if fpaa else torch.floataa __snake_case : Tuple = torch.from_numpy(load_hf_numpy(self.get_file_format(lowerCamelCase , lowerCamelCase ) ) ).to(lowerCamelCase ).to(lowerCamelCase ) return image def __snake_case ( self : Optional[Any] , lowerCamelCase : int="CompVis/stable-diffusion-v1-4" , lowerCamelCase : int=False ) -> int: __snake_case : str = "fp16" if fpaa else None __snake_case : int = torch.floataa if fpaa else torch.floataa __snake_case : int = AutoencoderKL.from_pretrained( lowerCamelCase , subfolder="vae" , torch_dtype=lowerCamelCase , revision=lowerCamelCase , ) model.to(lowerCamelCase ).eval() return model def __snake_case ( self : str , lowerCamelCase : int=0 ) -> Optional[Any]: if torch_device == "mps": return torch.manual_seed(lowerCamelCase ) return torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) @parameterized.expand( [ # fmt: off [33, [-0.16_03, 0.98_78, -0.04_95, -0.07_90, -0.27_09, 0.83_75, -0.20_60, -0.08_24], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_76, 0.11_68, 0.13_32, -0.48_40, -0.25_08, -0.07_91, -0.04_93, -0.40_89], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[str] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] ) -> List[Any]: __snake_case : Optional[Any] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) __snake_case : Tuple = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : List[Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : int = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [33, [-0.05_13, 0.02_89, 1.37_99, 0.21_66, -0.25_73, -0.08_71, 0.51_03, -0.09_99]], [47, [-0.41_28, -0.13_20, -0.37_04, 0.19_65, -0.41_16, -0.23_32, -0.33_40, 0.22_47]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : Any , lowerCamelCase : List[str] , lowerCamelCase : List[str] ) -> Tuple: __snake_case : Any = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Any = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.16_09, 0.98_66, -0.04_87, -0.07_77, -0.27_16, 0.83_68, -0.20_55, -0.08_14], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_77, 0.11_47, 0.13_33, -0.48_41, -0.25_06, -0.08_05, -0.04_91, -0.40_85], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Any , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) with torch.no_grad(): __snake_case : int = model(lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Union[str, Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : List[str] = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [13, [-0.20_51, -0.18_03, -0.23_11, -0.21_14, -0.32_92, -0.35_74, -0.29_53, -0.33_23]], [37, [-0.26_32, -0.26_25, -0.21_99, -0.27_41, -0.45_39, -0.49_90, -0.37_20, -0.49_25]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : List[str] , lowerCamelCase : Tuple , lowerCamelCase : Any ) -> Optional[Any]: __snake_case : List[str] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : str = sample[-1, -2:, :2, -2:].flatten().cpu() __snake_case : Optional[int] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-3 ) @parameterized.expand( [ # fmt: off [27, [-0.03_69, 0.02_07, -0.07_76, -0.06_82, -0.17_47, -0.19_30, -0.14_65, -0.20_39]], [16, [-0.16_28, -0.21_34, -0.27_47, -0.26_42, -0.37_74, -0.44_04, -0.36_87, -0.42_77]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : str , lowerCamelCase : Optional[int] , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[str] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : Union[str, Any] = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Optional[Any] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=5E-3 ) @parameterized.expand([(13,), (16,), (27,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> Tuple: __snake_case : Dict = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : Any = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Any = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-1 ) @parameterized.expand([(13,), (16,), (37,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : List[Any] , lowerCamelCase : Any ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : Union[str, Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : List[Any] = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Dict = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.30_01, 0.09_18, -2.69_84, -3.97_20, -3.20_99, -5.03_53, 1.73_38, -0.20_65, 3.42_67]], [47, [-1.50_30, -4.38_71, -6.03_55, -9.11_57, -1.66_61, -2.78_53, 2.16_07, -5.08_23, 2.56_33]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Dict ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : int = self.get_sd_image(lowerCamelCase ) __snake_case : int = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model.encode(lowerCamelCase ).latent_dist __snake_case : Dict = dist.sample(generator=lowerCamelCase ) assert list(sample.shape ) == [image.shape[0], 4] + [i // 8 for i in image.shape[2:]] __snake_case : List[str] = sample[0, -1, -3:, -3:].flatten().cpu() __snake_case : Dict = torch.tensor(lowerCamelCase ) __snake_case : Dict = 3E-3 if torch_device != "mps" else 1E-2 assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=lowerCamelCase )
81
0
import copy import os import cva import numpy as np from matplotlib import pyplot as plt class lowercase__: """simple docstring""" def __init__( self : Optional[int] ) -> int: lowercase_ = '''''' lowercase_ = '''''' lowercase_ = [] lowercase_ = 0 lowercase_ = 2_5_6 lowercase_ = 0 lowercase_ = 0 lowercase_ = 0 lowercase_ = 0 def _lowercase ( self : Tuple , SCREAMING_SNAKE_CASE_ : Tuple ) -> List[str]: lowercase_ = cva.imread(SCREAMING_SNAKE_CASE_ , 0 ) lowercase_ = copy.deepcopy(self.img ) lowercase_ , lowercase_ , lowercase_ = plt.hist(self.img.ravel() , 2_5_6 , [0, 2_5_6] , label='''x''' ) lowercase_ = np.sum(SCREAMING_SNAKE_CASE_ ) for i in range(len(SCREAMING_SNAKE_CASE_ ) ): lowercase_ = x[i] / self.k self.sk += prk lowercase_ = (self.L - 1) * self.sk if self.rem != 0: lowercase_ = int(last % last ) lowercase_ = int(last + 1 if self.rem >= 0.5 else last ) self.last_list.append(SCREAMING_SNAKE_CASE_ ) lowercase_ = int(np.ma.count(self.img ) / self.img[1].size ) lowercase_ = self.img[1].size for i in range(self.number_of_cols ): for j in range(self.number_of_rows ): lowercase_ = self.img[j][i] if num != self.last_list[num]: lowercase_ = self.last_list[num] cva.imwrite('''output_data/output.jpg''' , self.img ) def _lowercase ( self : Dict ) -> Union[str, Any]: plt.hist(self.img.ravel() , 2_5_6 , [0, 2_5_6] ) def _lowercase ( self : int ) -> List[Any]: cva.imshow('''Output-Image''' , self.img ) cva.imshow('''Input-Image''' , self.original_image ) cva.waitKey(5_0_0_0 ) cva.destroyAllWindows() if __name__ == "__main__": __a = os.path.join(os.path.basename(__file__), 'image_data/input.jpg') __a = ConstantStretch() stretcher.stretch(file_path) stretcher.plot_histogram() stretcher.show_image()
97
import warnings from ...utils import logging from .image_processing_flava import FlavaImageProcessor _snake_case : Optional[int] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[Any] , *lowerCamelCase : Any , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class FlavaFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use FlavaImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
'''simple docstring''' def a__ ( lowercase : int, lowercase : int ) -> float: """simple docstring""" return base * power(lowercase, (exponent - 1) ) if exponent else 1 if __name__ == "__main__": print('Raise base to the power of exponent using recursion...') lowercase__ : Dict = int(input('Enter the base: ').strip()) lowercase__ : Optional[Any] = int(input('Enter the exponent: ').strip()) lowercase__ : Dict = power(base, abs(exponent)) if exponent < 0: # power() does not properly deal w/ negative exponents lowercase__ : Dict = 1 / result print(F"""{base} to the power of {exponent} is {result}""")
98
import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : Tuple ) -> Optional[Any]: __snake_case : Dict = tempfile.mkdtemp() __snake_case : Any = SamImageProcessor() __snake_case : Optional[int] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : Optional[Any] , **lowerCamelCase : Optional[int] ) -> Optional[Any]: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[Any] ) -> Dict: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : int ) -> List[Any]: __snake_case : List[str] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : int = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : List[Any] ) -> Dict: __snake_case : int = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Dict = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Optional[Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : List[str] ) -> Tuple: __snake_case : int = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Optional[int] = self.prepare_image_inputs() __snake_case : List[str] = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : Dict = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_torch def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Tuple = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[str] = [torch.ones((1, 3, 5, 5) )] __snake_case : Tuple = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : int = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , torch.tensor(lowerCamelCase ) , torch.tensor(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : List[str] = [np.ones((1, 3, 5, 5) )] __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : str = [[1, 0], [0, 1]] with self.assertRaises(lowerCamelCase ): __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) @require_vision @require_tf class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[Any] ) -> Union[str, Any]: __snake_case : int = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : str , **lowerCamelCase : Any ) -> Tuple: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> Any: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : str ) -> List[Any]: __snake_case : int = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : Dict = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : int ) -> List[str]: __snake_case : List[Any] = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Optional[Any] = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Tuple = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : Union[str, Any] ) -> List[Any]: __snake_case : str = self.get_image_processor() __snake_case : Union[str, Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : int = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : List[str] = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_tf def __snake_case ( self : Any ) -> Optional[int]: __snake_case : List[str] = self.get_image_processor() __snake_case : Dict = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Union[str, Any] = [tf.ones((1, 3, 5, 5) )] __snake_case : List[Any] = [[1764, 2646]] __snake_case : Dict = [[683, 1024]] __snake_case : List[str] = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , tf.convert_to_tensor(lowerCamelCase ) , tf.convert_to_tensor(lowerCamelCase ) , return_tensors="tf" , ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : Union[str, Any] = [np.ones((1, 3, 5, 5) )] __snake_case : List[str] = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Tuple = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): __snake_case : Dict = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[str] ) -> str: __snake_case : Optional[int] = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : List[str] , **lowerCamelCase : Any ) -> Dict: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> List[Any]: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : str = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : List[Any] = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : str = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[Any] = np.random.randint(0 , 2 , size=(1, 3, 5, 5) ).astype(np.floataa ) __snake_case : Dict = [tf.convert_to_tensor(lowerCamelCase )] __snake_case : List[Any] = [torch.tensor(lowerCamelCase )] __snake_case : Optional[Any] = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : Union[str, Any] = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) __snake_case : Dict = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="pt" ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def __snake_case ( self : List[Any] ) -> List[str]: __snake_case : Any = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : Any = image_processor(lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Optional[Any] = processor(images=lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Tuple = image_processor(lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() __snake_case : List[Any] = processor(images=lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) )
81
0
import logging import re import pytorch_quantization import pytorch_quantization.nn as quant_nn import torch from pytorch_quantization import calib from pytorch_quantization.tensor_quant import QuantDescriptor SCREAMING_SNAKE_CASE = logging.getLogger(__name__) SCREAMING_SNAKE_CASE = 5_0 # max width of layer names SCREAMING_SNAKE_CASE = 7_0 # max width of quantizer names def a (lowerCAmelCase__ ): __a = parser.add_argument_group("""quant_trainer arguments""" ) group.add_argument("""--wprec""" , type=lowerCAmelCase__ , default=8 , help="""weight precision""" ) group.add_argument("""--aprec""" , type=lowerCAmelCase__ , default=8 , help="""activation precision""" ) group.add_argument("""--quant-per-tensor""" , action="""store_true""" , help="""per tensor weight scaling""" ) group.add_argument("""--quant-disable""" , action="""store_true""" , help="""disable all quantizers""" ) group.add_argument("""--quant-disable-embeddings""" , action="""store_true""" , help="""disable all embeddings quantizers""" ) group.add_argument("""--quant-disable-keyword""" , type=lowerCAmelCase__ , nargs="""+""" , help="""disable quantizers by keyword""" ) group.add_argument("""--quant-disable-layer-module""" , type=lowerCAmelCase__ , help="""disable quantizers by keyword under layer.""" ) group.add_argument("""--quant-enable-layer-module""" , type=lowerCAmelCase__ , help="""enable quantizers by keyword under layer""" ) group.add_argument("""--calibrator""" , default="""max""" , help="""which quantization range calibrator to use""" ) group.add_argument("""--percentile""" , default=lowerCAmelCase__ , type=lowerCAmelCase__ , help="""percentile for PercentileCalibrator""" ) group.add_argument("""--fuse-qkv""" , action="""store_true""" , help="""use the same scale factor for qkv""" ) group.add_argument("""--clip-gelu""" , metavar="""N""" , type=lowerCAmelCase__ , help="""clip gelu output maximum value to N""" ) group.add_argument( """--recalibrate-weights""" , action="""store_true""" , help=( """recalibrate weight amaxes by taking the max of the weights.""" """ amaxes will be computed with the current quantization granularity (axis).""" ) , ) def a (lowerCAmelCase__ ): if args.calibrator == "max": __a = """max""" elif args.calibrator == "percentile": if args.percentile is None: raise ValueError("""Specify --percentile when using percentile calibrator""" ) __a = """histogram""" elif args.calibrator == "mse": __a = """histogram""" else: raise ValueError(f'''Invalid calibrator {args.calibrator}''' ) __a = QuantDescriptor(num_bits=args.aprec , calib_method=lowerCAmelCase__ ) __a = QuantDescriptor(num_bits=args.wprec , axis=(None if args.quant_per_tensor else (0,)) ) quant_nn.QuantLinear.set_default_quant_desc_input(lowerCAmelCase__ ) quant_nn.QuantLinear.set_default_quant_desc_weight(lowerCAmelCase__ ) def a (lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=False , lowerCAmelCase__=False ): logger.info("""Configuring Model for Quantization""" ) logger.info(f'''using quantization package {pytorch_quantization.__file__}''' ) if not calib: if args.quant_disable_embeddings: set_quantizer_by_name(lowerCAmelCase__ , ["""embeddings"""] , which="""weight""" , _disabled=lowerCAmelCase__ ) if args.quant_disable: set_quantizer_by_name(lowerCAmelCase__ , [""""""] , _disabled=lowerCAmelCase__ ) if args.quant_disable_keyword: set_quantizer_by_name(lowerCAmelCase__ , args.quant_disable_keyword , _disabled=lowerCAmelCase__ ) if args.quant_disable_layer_module: set_quantizer_by_name(lowerCAmelCase__ , [r"""layer.\d+.""" + args.quant_disable_layer_module] , _disabled=lowerCAmelCase__ ) if args.quant_enable_layer_module: set_quantizer_by_name(lowerCAmelCase__ , [r"""layer.\d+.""" + args.quant_enable_layer_module] , _disabled=lowerCAmelCase__ ) if args.recalibrate_weights: recalibrate_weights(lowerCAmelCase__ ) if args.fuse_qkv: fuse_qkv(lowerCAmelCase__ , lowerCAmelCase__ ) if args.clip_gelu: clip_gelu(lowerCAmelCase__ , args.clip_gelu ) # if args.local_rank in [-1, 0] and not calib: print_quant_summary(lowerCAmelCase__ ) def a (lowerCAmelCase__ ): logger.info("""Enabling Calibration""" ) for name, module in model.named_modules(): if name.endswith("""_quantizer""" ): if module._calibrator is not None: module.disable_quant() module.enable_calib() else: module.disable() logger.info(f'''{name:80}: {module}''' ) def a (lowerCAmelCase__ , lowerCAmelCase__ ): logger.info("""Loading calibrated amax""" ) for name, module in model.named_modules(): if name.endswith("""_quantizer""" ): if module._calibrator is not None: if isinstance(module._calibrator , calib.MaxCalibrator ): module.load_calib_amax() else: module.load_calib_amax("""percentile""" , percentile=args.percentile ) module.enable_quant() module.disable_calib() else: module.enable() model.cuda() print_quant_summary(lowerCAmelCase__ ) def a (lowerCAmelCase__ , lowerCAmelCase__ ): def fusea(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): for mod in [qq, qk, qv]: if not hasattr(lowerCAmelCase__ , """_amax""" ): print(""" WARNING: NO AMAX BUFFER""" ) return __a = qq._amax.detach().item() __a = qk._amax.detach().item() __a = qv._amax.detach().item() __a = max(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) qq._amax.fill_(lowerCAmelCase__ ) qk._amax.fill_(lowerCAmelCase__ ) qv._amax.fill_(lowerCAmelCase__ ) logger.info(f''' q={q:5.2f} k={k:5.2f} v={v:5.2f} -> {amax:5.2f}''' ) for name, mod in model.named_modules(): if name.endswith(""".attention.self""" ): logger.info(f'''FUSE_QKV: {name:{name_width}}''' ) fusea(mod.matmul_q_input_quantizer , mod.matmul_k_input_quantizer , mod.matmul_v_input_quantizer ) if args.quant_per_tensor: fusea(mod.query._weight_quantizer , mod.key._weight_quantizer , mod.value._weight_quantizer ) def a (lowerCAmelCase__ , lowerCAmelCase__ ): for name, mod in model.named_modules(): if name.endswith(""".output.dense""" ) and not name.endswith("""attention.output.dense""" ): __a = mod._input_quantizer._amax.data.detach().item() mod._input_quantizer._amax.data.detach().clamp_(max=lowerCAmelCase__ ) __a = mod._input_quantizer._amax.data.detach().item() logger.info(f'''CLIP_GELU: {name:{name_width}} amax: {amax_init:5.2f} -> {amax:5.2f}''' ) def a (lowerCAmelCase__ ): for name, mod in model.named_modules(): if hasattr(lowerCAmelCase__ , """_weight_quantizer""" ) and mod._weight_quantizer.axis is not None: __a = mod.weight.shape[0] __a = mod._weight_quantizer._amax.detach() __a = torch.ones(lowerCAmelCase__ , dtype=amax.dtype , device=amax.device ) * amax print(f'''expanding {name} {amax} -> {mod._weight_quantizer._amax}''' ) def a (lowerCAmelCase__ ): for name, mod in model.named_modules(): if hasattr(lowerCAmelCase__ , """_weight_quantizer""" ): if not hasattr(mod.weight_quantizer , """_amax""" ): print("""RECALIB: {name:{name_width}} WARNING: NO AMAX BUFFER""" ) continue # determine which axes to reduce across # e.g. a 4D tensor quantized per axis 0 should reduce over (1,2,3) __a = set() if mod._weight_quantizer.axis is None else set(mod._weight_quantizer.axis ) __a = set(range(len(mod.weight.size() ) ) ) - axis_set __a = pytorch_quantization.utils.reduce_amax(mod.weight , axis=lowerCAmelCase__ , keepdims=lowerCAmelCase__ ).detach() logger.info(f'''RECALIB: {name:{name_width}} {mod._weight_quantizer._amax.flatten()} -> {amax.flatten()}''' ) __a = amax def a (lowerCAmelCase__ , lowerCAmelCase__=25 , lowerCAmelCase__=180 , lowerCAmelCase__=None ): if ignore is None: __a = [] elif not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): __a = [ignore] __a = 0 for name, mod in model.named_modules(): if not hasattr(lowerCAmelCase__ , """weight""" ): continue __a = max(lowerCAmelCase__ , len(lowerCAmelCase__ ) ) for name, mod in model.named_modules(): __a = getattr(lowerCAmelCase__ , """_input_quantizer""" , lowerCAmelCase__ ) __a = getattr(lowerCAmelCase__ , """_weight_quantizer""" , lowerCAmelCase__ ) if not hasattr(lowerCAmelCase__ , """weight""" ): continue if type(lowerCAmelCase__ ) in ignore: continue if [True for s in ignore if type(lowerCAmelCase__ ) is str and s in name]: continue __a = f'''Act:{input_q.extra_repr()}''' __a = f'''Wgt:{weight_q.extra_repr()}''' __a = f'''{name:{name_width}} {act_str} {wgt_str}''' if len(lowerCAmelCase__ ) <= line_width: logger.info(lowerCAmelCase__ ) else: logger.info(f'''{name:{name_width}} {act_str}''' ) logger.info(f'''{' ':{name_width}} {wgt_str}''' ) def a (lowerCAmelCase__ ): __a = 0 for name, mod in model.named_modules(): if isinstance(lowerCAmelCase__ , pytorch_quantization.nn.TensorQuantizer ): print(f'''{name:80} {mod}''' ) count += 1 print(f'''{count} TensorQuantizers found in model''' ) def a (lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): __a = getattr(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) if quantizer_mod is not None: assert hasattr(lowerCAmelCase__ , lowerCAmelCase__ ) setattr(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) else: logger.warning(f'''{name} has no {quantizer}''' ) def a (lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__="both" , **lowerCAmelCase__ ): __a = f'''Warning: changing {which} quantizers of {name:{qname_width}}''' for k, v in kwargs.items(): s += f''' {k}={v}''' if which in ["input", "both"]: set_quantizer(lowerCAmelCase__ , lowerCAmelCase__ , """_input_quantizer""" , lowerCAmelCase__ , lowerCAmelCase__ ) if which in ["weight", "both"]: set_quantizer(lowerCAmelCase__ , lowerCAmelCase__ , """_weight_quantizer""" , lowerCAmelCase__ , lowerCAmelCase__ ) logger.info(lowerCAmelCase__ ) def a (lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ): for name, mod in model.named_modules(): if hasattr(lowerCAmelCase__ , """_input_quantizer""" ) or hasattr(lowerCAmelCase__ , """_weight_quantizer""" ): for n in names: if re.search(lowerCAmelCase__ , lowerCAmelCase__ ): set_quantizers(lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) elif name.endswith("""_quantizer""" ): for n in names: if re.search(lowerCAmelCase__ , lowerCAmelCase__ ): __a = f'''Warning: changing {name:{name_width}}''' for k, v in kwargs.items(): s += f''' {k}={v}''' setattr(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) logger.info(lowerCAmelCase__ )
99
import warnings from ...utils import logging from .image_processing_deit import DeiTImageProcessor _snake_case : List[str] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[str] , *lowerCamelCase : str , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class DeiTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use DeiTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
import baseaa import io import json import os from copy import deepcopy from ..optimizer import AcceleratedOptimizer from ..scheduler import AcceleratedScheduler class __snake_case : '''simple docstring''' def __init__( self , A_ ): '''simple docstring''' if isinstance(A_ , A_ ): # Don't modify user's data should they want to reuse it (e.g. in tests), because once we # modified it, it will not be accepted here again, since `auto` values would have been overridden SCREAMING_SNAKE_CASE__ = deepcopy(A_ ) elif os.path.exists(A_ ): with io.open(A_ , '''r''' , encoding='''utf-8''' ) as f: SCREAMING_SNAKE_CASE__ = json.load(A_ ) else: try: SCREAMING_SNAKE_CASE__ = baseaa.urlsafe_baadecode(A_ ).decode('''utf-8''' ) SCREAMING_SNAKE_CASE__ = json.loads(A_ ) except (UnicodeDecodeError, AttributeError, ValueError): raise ValueError( f'''Expected a string path to an existing deepspeed config, or a dictionary, or a base64 encoded string. Received: {config_file_or_dict}''' ) SCREAMING_SNAKE_CASE__ = config self.set_stage_and_offload() def lowercase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = self.get_value('''zero_optimization.stage''' , -1 ) # offload SCREAMING_SNAKE_CASE__ = False if self.is_zeroa() or self.is_zeroa(): SCREAMING_SNAKE_CASE__ = set(['''cpu''', '''nvme'''] ) SCREAMING_SNAKE_CASE__ = set( [ self.get_value('''zero_optimization.offload_optimizer.device''' ), self.get_value('''zero_optimization.offload_param.device''' ), ] ) if len(offload_devices & offload_devices_valid ) > 0: SCREAMING_SNAKE_CASE__ = True def lowercase_ ( self , A_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = self.config # find the config node of interest if it exists SCREAMING_SNAKE_CASE__ = ds_key_long.split('''.''' ) SCREAMING_SNAKE_CASE__ = nodes.pop() for node in nodes: SCREAMING_SNAKE_CASE__ = config.get(A_ ) if config is None: return None, ds_key return config, ds_key def lowercase_ ( self , A_ , A_=None ): '''simple docstring''' SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.find_config_node(A_ ) if config is None: return default return config.get(A_ , A_ ) def lowercase_ ( self , A_ , A_=False ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = self.config # find the config node of interest if it exists SCREAMING_SNAKE_CASE__ = ds_key_long.split('''.''' ) for node in nodes: SCREAMING_SNAKE_CASE__ = config SCREAMING_SNAKE_CASE__ = config.get(A_ ) if config is None: if must_exist: raise ValueError(f'''Can\'t find {ds_key_long} entry in the config: {self.config}''' ) else: return # if found remove it if parent_config is not None: parent_config.pop(A_ ) def lowercase_ ( self , A_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = self.get_value(A_ ) return False if value is None else bool(A_ ) def lowercase_ ( self , A_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = self.get_value(A_ ) return False if value is None else not bool(A_ ) def lowercase_ ( self ): '''simple docstring''' return self._stage == 2 def lowercase_ ( self ): '''simple docstring''' return self._stage == 3 def lowercase_ ( self ): '''simple docstring''' return self._offload class __snake_case : '''simple docstring''' def __init__( self , A_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = engine def lowercase_ ( self , A_ , **A_ ): '''simple docstring''' self.engine.backward(A_ , **A_ ) # Deepspeed's `engine.step` performs the following operations: # - gradient accumulation check # - gradient clipping # - optimizer step # - zero grad # - checking overflow # - lr_scheduler step (only if engine.lr_scheduler is not None) self.engine.step() # and this plugin overrides the above calls with no-ops when Accelerate runs under # Deepspeed, but allows normal functionality for non-Deepspeed cases thus enabling a simple # training loop that works transparently under many training regimes. class __snake_case ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self , A_ ): '''simple docstring''' super().__init__(A_ , device_placement=A_ , scaler=A_ ) SCREAMING_SNAKE_CASE__ = hasattr(self.optimizer , '''overflow''' ) def lowercase_ ( self , A_=None ): '''simple docstring''' pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed def lowercase_ ( self ): '''simple docstring''' pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed @property def lowercase_ ( self ): '''simple docstring''' if self.__has_overflow__: return self.optimizer.overflow return False class __snake_case ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self , A_ , A_ ): '''simple docstring''' super().__init__(A_ , A_ ) def lowercase_ ( self ): '''simple docstring''' pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed class __snake_case : '''simple docstring''' def __init__( self , A_ , A_=0.001 , A_=0 , **A_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = params SCREAMING_SNAKE_CASE__ = lr SCREAMING_SNAKE_CASE__ = weight_decay SCREAMING_SNAKE_CASE__ = kwargs class __snake_case : '''simple docstring''' def __init__( self , A_ , A_=None , A_=0 , **A_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = optimizer SCREAMING_SNAKE_CASE__ = total_num_steps SCREAMING_SNAKE_CASE__ = warmup_num_steps SCREAMING_SNAKE_CASE__ = kwargs
100
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case : Union[str, Any] = { "configuration_owlvit": [ "OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "OwlViTConfig", "OwlViTOnnxConfig", "OwlViTTextConfig", "OwlViTVisionConfig", ], "processing_owlvit": ["OwlViTProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[Any] = ["OwlViTFeatureExtractor"] _snake_case : Optional[int] = ["OwlViTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "OwlViTModel", "OwlViTPreTrainedModel", "OwlViTTextModel", "OwlViTVisionModel", "OwlViTForObjectDetection", ] if TYPE_CHECKING: from .configuration_owlvit import ( OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, OwlViTConfig, OwlViTOnnxConfig, OwlViTTextConfig, OwlViTVisionConfig, ) from .processing_owlvit import OwlViTProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_owlvit import OwlViTFeatureExtractor from .image_processing_owlvit import OwlViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_owlvit import ( OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST, OwlViTForObjectDetection, OwlViTModel, OwlViTPreTrainedModel, OwlViTTextModel, OwlViTVisionModel, ) else: import sys _snake_case : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
0
import torch from torch import nn from transformers import CLIPPreTrainedModel, CLIPVisionModel from ...models.attention import BasicTransformerBlock from ...utils import logging lowerCAmelCase__ : Optional[int] =logging.get_logger(__name__) # pylint: disable=invalid-name class __lowercase (__SCREAMING_SNAKE_CASE ): """simple docstring""" def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=7_6_8 ): """simple docstring""" super().__init__(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Tuple = proj_size SCREAMING_SNAKE_CASE_ : int = CLIPVisionModel(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : str = PaintByExampleMapper(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = nn.LayerNorm(config.hidden_size ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = nn.Linear(config.hidden_size , self.proj_size ) # uncondition for scaling SCREAMING_SNAKE_CASE_ : List[Any] = nn.Parameter(torch.randn((1, 1, self.proj_size) ) ) def UpperCamelCase__ ( self , lowerCAmelCase__ , lowerCAmelCase__=False ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str = self.model(pixel_values=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : List[str] = clip_output.pooler_output SCREAMING_SNAKE_CASE_ : Dict = self.mapper(latent_states[:, None] ) SCREAMING_SNAKE_CASE_ : Optional[Any] = self.final_layer_norm(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Tuple = self.proj_out(lowerCAmelCase__ ) if return_uncond_vector: return latent_states, self.uncond_vector return latent_states class __lowercase (nn.Module ): """simple docstring""" def __init__( self , lowerCAmelCase__ ): """simple docstring""" super().__init__() SCREAMING_SNAKE_CASE_ : str = (config.num_hidden_layers + 1) // 5 SCREAMING_SNAKE_CASE_ : Tuple = config.hidden_size SCREAMING_SNAKE_CASE_ : Optional[int] = 1 SCREAMING_SNAKE_CASE_ : Optional[int] = nn.ModuleList( [ BasicTransformerBlock(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , activation_fn='gelu' , attention_bias=lowerCAmelCase__ ) for _ in range(lowerCAmelCase__ ) ] ) def UpperCamelCase__ ( self , lowerCAmelCase__ ): """simple docstring""" for block in self.blocks: SCREAMING_SNAKE_CASE_ : Optional[Any] = block(lowerCAmelCase__ ) return hidden_states
101
from ....configuration_utils import PretrainedConfig from ....utils import logging _snake_case : int = logging.get_logger(__name__) _snake_case : str = { "Visual-Attention-Network/van-base": ( "https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json" ), } class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[Any] = "van" def __init__( self : Optional[int] , lowerCamelCase : Any=224 , lowerCamelCase : str=3 , lowerCamelCase : Any=[7, 3, 3, 3] , lowerCamelCase : Dict=[4, 2, 2, 2] , lowerCamelCase : List[Any]=[64, 128, 320, 512] , lowerCamelCase : str=[3, 3, 12, 3] , lowerCamelCase : Dict=[8, 8, 4, 4] , lowerCamelCase : Any="gelu" , lowerCamelCase : Optional[int]=0.02 , lowerCamelCase : Tuple=1E-6 , lowerCamelCase : Optional[int]=1E-2 , lowerCamelCase : int=0.0 , lowerCamelCase : Optional[Any]=0.0 , **lowerCamelCase : Optional[int] , ) -> int: super().__init__(**lowerCamelCase ) __snake_case : Union[str, Any] = image_size __snake_case : Any = num_channels __snake_case : Any = patch_sizes __snake_case : List[Any] = strides __snake_case : str = hidden_sizes __snake_case : Dict = depths __snake_case : Optional[int] = mlp_ratios __snake_case : Dict = hidden_act __snake_case : Union[str, Any] = initializer_range __snake_case : List[str] = layer_norm_eps __snake_case : Optional[int] = layer_scale_init_value __snake_case : List[Any] = drop_path_rate __snake_case : int = dropout_rate
81
0
"""simple docstring""" from math import ceil, sqrt def UpperCamelCase (SCREAMING_SNAKE_CASE = 100_0000 ): UpperCamelCase : int = 0 for outer_width in range(3 , (limit // 4) + 2 ): if outer_width**2 > limit: UpperCamelCase : Optional[Any] = max(ceil(sqrt(outer_width**2 - limit ) ) , 1 ) else: UpperCamelCase : str = 1 if (outer_width - hole_width_lower_bound) % 2: hole_width_lower_bound += 1 answer += (outer_width - hole_width_lower_bound - 2) // 2 + 1 return answer if __name__ == "__main__": print(f'''{solution() = }''')
102
import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer _snake_case : Union[str, Any] = logging.getLogger(__name__) def lowerCAmelCase_ ( ): __snake_case : int = argparse.ArgumentParser( description="Prepare TFRecord shards from pre-tokenized samples of the wikitext dataset." ) parser.add_argument( "--dataset_name" , type=__lowerCamelCase , default="wikitext" , help="Name of the training. Explore datasets at: hf.co/datasets." , ) parser.add_argument( "--dataset_config" , type=__lowerCamelCase , default="wikitext-103-raw-v1" , help="Configuration name of the dataset." ) parser.add_argument( "--tokenizer_name_or_path" , type=__lowerCamelCase , default="sayakpaul/unigram-tokenizer-wikitext" , help="Tokenizer identifier. Can be a local filepath or a Hub identifier." , ) parser.add_argument( "--shard_size" , type=__lowerCamelCase , default=1_0_0_0 , help="Number of entries to go in a single shard." , ) parser.add_argument("--split" , type=__lowerCamelCase , default="train" , choices=["train", "test", "validation"] ) parser.add_argument( "--limit" , default=__lowerCamelCase , type=__lowerCamelCase , help="Limit the number of shards (used for debugging)." , ) parser.add_argument( "--max_length" , type=__lowerCamelCase , default=5_1_2 , help="Maximum sequence length. For training on TPUs, it helps to have a maximum" " sequence length that is a multiple of 8." , ) parser.add_argument( "--output_dir" , default="tf-tpu" , type=__lowerCamelCase , help="Output directory where the TFRecord shards will be saved. If the" " path is appended with `gs://` ('gs://tf-tpu', for example) then the TFRecord" " shards will be directly saved to a Google Cloud Storage bucket." , ) __snake_case : List[str] = parser.parse_args() return args def lowerCAmelCase_ ( __lowerCamelCase ): def fn(__lowerCamelCase ): return tokenizer(examples["text"] ) return fn def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = [] for i in range(len(tokenized_data["input_ids"] ) ): __snake_case : Tuple = { "input_ids": tf.train.Feature(intaa_list=tf.train.IntaaList(value=tokenized_data["input_ids"][i] ) ), "attention_mask": tf.train.Feature( intaa_list=tf.train.IntaaList(value=tokenized_data["attention_mask"][i] ) ), } __snake_case : List[Any] = tf.train.Features(feature=__lowerCamelCase ) __snake_case : str = tf.train.Example(features=__lowerCamelCase ) __snake_case : List[str] = example.SerializeToString() records.append(__lowerCamelCase ) return records def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Optional[int] = datasets.load_dataset(args.dataset_name , args.dataset_config , split=args.split ) if args.limit is not None: __snake_case : Optional[Any] = min(len(__lowerCamelCase ) , args.limit ) __snake_case : Dict = dataset.select(range(__lowerCamelCase ) ) print(F'Limiting the dataset to {args.limit} entries.' ) __snake_case : Dict = AutoTokenizer.from_pretrained(args.tokenizer_name_or_path ) # Handle output directory creation. # For serializing into a Google Cloud Storage Bucket, one needs to first # create a bucket. if "gs" not in args.output_dir: if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) __snake_case : Dict = os.path.join(args.output_dir , args.split ) if not os.path.exists(__lowerCamelCase ): os.makedirs(__lowerCamelCase ) else: __snake_case : str = os.path.join(args.output_dir , args.split ) # Tokenize the whole dataset at once. __snake_case : Any = tokenize_function(__lowerCamelCase ) __snake_case : Optional[Any] = dataset.map(__lowerCamelCase , batched=__lowerCamelCase , num_proc=4 , remove_columns=["text"] ) # We need to concatenate all our texts together, and then split the result # into chunks of a fixed size, which we will call block_size. To do this, we # will use the map method again, with the option batched=True. When we use batched=True, # the function we pass to map() will be passed multiple inputs at once, allowing us # to group them into more or fewer examples than we had in the input. # This allows us to create our new fixed-length samples. The advantage of this # method is that we don't lose a whole lot of content from the dataset compared to the # case where we simply tokenize with a pre-defined max_length. def group_texts(__lowerCamelCase ): # Concatenate all texts. __snake_case : List[str] = {k: sum(examples[k] , [] ) for k in examples.keys()} __snake_case : List[Any] = len(concatenated_examples[list(examples.keys() )[0]] ) # We drop the small remainder, though you could add padding instead if the model supports it # In this, as in all things, we advise you to follow your heart 🫀 __snake_case : Any = (total_length // args.max_length) * args.max_length # Split by chunks of max_len. __snake_case : int = { k: [t[i : i + args.max_length] for i in range(0 , __lowerCamelCase , args.max_length )] for k, t in concatenated_examples.items() } return result __snake_case : Any = dataset_tokenized.map(__lowerCamelCase , batched=__lowerCamelCase , batch_size=1_0_0_0 , num_proc=4 ) __snake_case : Optional[Any] = 0 __snake_case : Optional[Any] = 0 for shard in range(0 , len(__lowerCamelCase ) , args.shard_size ): __snake_case : List[str] = grouped_dataset[shard : shard + args.shard_size] __snake_case : Any = len(dataset_snapshot["input_ids"] ) __snake_case : List[Any] = os.path.join(__lowerCamelCase , F'dataset-{shard_count}-{records_containing}.tfrecord' ) __snake_case : Optional[Any] = get_serialized_examples(__lowerCamelCase ) with tf.io.TFRecordWriter(__lowerCamelCase ) as out_file: for i in range(len(__lowerCamelCase ) ): __snake_case : Union[str, Any] = serialized_examples[i] out_file.write(__lowerCamelCase ) print("Wrote file {} containing {} records".format(__lowerCamelCase , __lowerCamelCase ) ) shard_count += 1 total_records += records_containing with open(F'split-{args.split}-records-count.txt' , "w" ) as f: print(F'Total {args.split} records: {total_records}' , file=__lowerCamelCase ) if __name__ == "__main__": _snake_case : List[Any] = parse_args() main(args)
81
0
"""simple docstring""" from __future__ import annotations def snake_case ( lowerCAmelCase_ , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = False , ) -> tuple[int, float, str]: _snake_case = cipher_alphabet or [chr(lowerCAmelCase_ ) for i in range(97 , 123 )] # If the argument is None or the user provided an empty dictionary if not frequencies_dict: # Frequencies of letters in the english language (how much they show up) _snake_case = { '''a''': 0.0_8497, '''b''': 0.0_1492, '''c''': 0.0_2202, '''d''': 0.0_4253, '''e''': 0.1_1162, '''f''': 0.0_2228, '''g''': 0.0_2015, '''h''': 0.0_6094, '''i''': 0.0_7546, '''j''': 0.0_0153, '''k''': 0.0_1292, '''l''': 0.0_4025, '''m''': 0.0_2406, '''n''': 0.0_6749, '''o''': 0.0_7507, '''p''': 0.0_1929, '''q''': 0.0_0095, '''r''': 0.0_7587, '''s''': 0.0_6327, '''t''': 0.0_9356, '''u''': 0.0_2758, '''v''': 0.0_0978, '''w''': 0.0_2560, '''x''': 0.0_0150, '''y''': 0.0_1994, '''z''': 0.0_0077, } else: # Custom frequencies dictionary _snake_case = frequencies_dict if not case_sensitive: _snake_case = ciphertext.lower() # Chi squared statistic values _snake_case = {} # cycle through all of the shifts for shift in range(len(lowerCAmelCase_ ) ): _snake_case = '''''' # decrypt the message with the shift for letter in ciphertext: try: # Try to index the letter in the alphabet _snake_case = (alphabet_letters.index(letter.lower() ) - shift) % len( lowerCAmelCase_ ) decrypted_with_shift += ( alphabet_letters[new_key].upper() if case_sensitive and letter.isupper() else alphabet_letters[new_key] ) except ValueError: # Append the character if it isn't in the alphabet decrypted_with_shift += letter _snake_case = 0.0 # Loop through each letter in the decoded message with the shift for letter in decrypted_with_shift: if case_sensitive: _snake_case = letter.lower() if letter in frequencies: # Get the amount of times the letter occurs in the message _snake_case = decrypted_with_shift.lower().count(lowerCAmelCase_ ) # Get the excepcted amount of times the letter should appear based # on letter frequencies _snake_case = frequencies[letter] * occurrences # Complete the chi squared statistic formula _snake_case = ((occurrences - expected) ** 2) / expected # Add the margin of error to the total chi squared statistic chi_squared_statistic += chi_letter_value else: if letter.lower() in frequencies: # Get the amount of times the letter occurs in the message _snake_case = decrypted_with_shift.count(lowerCAmelCase_ ) # Get the excepcted amount of times the letter should appear based # on letter frequencies _snake_case = frequencies[letter] * occurrences # Complete the chi squared statistic formula _snake_case = ((occurrences - expected) ** 2) / expected # Add the margin of error to the total chi squared statistic chi_squared_statistic += chi_letter_value # Add the data to the chi_squared_statistic_values dictionary _snake_case = ( chi_squared_statistic, decrypted_with_shift, ) # Get the most likely cipher by finding the cipher with the smallest chi squared # statistic def chi_squared_statistic_values_sorting_key(lowerCAmelCase_ ) -> tuple[float, str]: return chi_squared_statistic_values[key] _snake_case = min( lowerCAmelCase_ , key=lowerCAmelCase_ , ) # Get all the data from the most likely cipher (key, decoded message) ( ( _snake_case ) , ( _snake_case ) , ) = chi_squared_statistic_values[most_likely_cipher] # Return the data on the most likely shift return ( most_likely_cipher, most_likely_cipher_chi_squared_value, decoded_most_likely_cipher, )
103
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) _snake_case : List[Any] = "\\n Text data.\n Second line of data." _snake_case : Tuple = "file" @pytest.fixture(scope="session" ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / (FILE_PATH + ".zstd") __snake_case : Optional[Any] = bytes(__lowerCamelCase , "utf-8" ) with zstd.open(__lowerCamelCase , "wb" ) as f: f.write(__lowerCamelCase ) return path @pytest.fixture def lowerCAmelCase_ ( __lowerCamelCase ): with open(os.path.join(tmpfs.local_root_dir , __lowerCamelCase ) , "w" ) as f: f.write(__lowerCamelCase ) return FILE_PATH @pytest.mark.parametrize("compression_format" , ["gzip", "xz", "zstd"] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Optional[int] = {"gzip": gz_file, "xz": xz_file, "zstd": zstd_path} __snake_case : str = input_paths[compression_format] __snake_case : Optional[Any] = tmp_path / "cache" __snake_case : Optional[int] = DownloadConfig(cache_dir=__lowerCamelCase , extract_compressed_file=__lowerCamelCase ) __snake_case : Union[str, Any] = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) with open(__lowerCamelCase ) as f: __snake_case : Dict = f.read() with open(__lowerCamelCase ) as f: __snake_case : Tuple = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("default_extracted" , [True, False] ) @pytest.mark.parametrize("default_cache_dir" , [True, False] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Tuple = "custom_cache" __snake_case : List[str] = "custom_extracted_dir" __snake_case : Any = tmp_path / "custom_extracted_path" if default_extracted: __snake_case : List[Any] = ("downloads" if default_cache_dir else custom_cache_dir, "extracted") else: monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_DIR" , __lowerCamelCase ) monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_PATH" , str(__lowerCamelCase ) ) __snake_case : Optional[Any] = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) __snake_case : Optional[int] = xz_file __snake_case : Optional[int] = ( DownloadConfig(extract_compressed_file=__lowerCamelCase ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=__lowerCamelCase ) ) __snake_case : str = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) assert Path(__lowerCamelCase ).parent.parts[-2:] == expected def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : Optional[Any] = str(Path(__lowerCamelCase ).resolve() ) assert cached_path(__lowerCamelCase ) == text_file # relative path __snake_case : Any = str(Path(__lowerCamelCase ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(__lowerCamelCase ) == text_file def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : List[Any] = str(tmp_path.resolve() / "__missing_file__.txt" ) with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) # relative path __snake_case : Optional[int] = "./__missing_file__.txt" with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : str = get_from_cache(F'tmp://{tmpfs_file}' ) with open(__lowerCamelCase ) as f: __snake_case : Union[str, Any] = f.read() assert output_file_content == FILE_CONTENT @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( ): with pytest.raises(__lowerCamelCase ): cached_path("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[Any] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): http_get("https://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): http_head("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[str] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): ftp_get("ftp://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): ftp_head("ftp://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): fsspec_get("s3://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): fsspec_head("s3://huggingface.co" )
81
0
"""simple docstring""" def _lowerCamelCase ( UpperCAmelCase_ : int, UpperCAmelCase_ : int ) -> int: """simple docstring""" return int(input_a == input_a == 0 ) def _lowerCamelCase ( ) -> None: """simple docstring""" print("Truth Table of NOR Gate:" ) print("| Input 1 | Input 2 | Output |" ) print(F"""| 0 | 0 | {nor_gate(0, 0 )} |""" ) print(F"""| 0 | 1 | {nor_gate(0, 1 )} |""" ) print(F"""| 1 | 0 | {nor_gate(1, 0 )} |""" ) print(F"""| 1 | 1 | {nor_gate(1, 1 )} |""" ) if __name__ == "__main__": import doctest doctest.testmod() main()
104
_snake_case : Optional[int] = {"a": ["c", "b"], "b": ["d", "e"], "c": [], "d": [], "e": []} _snake_case : Dict = ["a", "b", "c", "d", "e"] def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : List[str] = start # add current to visited visited.append(__lowerCamelCase ) __snake_case : List[Any] = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: __snake_case : Tuple = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # if all neighbors visited add current to sort sort.append(__lowerCamelCase ) # if all vertices haven't been visited select a new one to visit if len(__lowerCamelCase ) != len(__lowerCamelCase ): for vertice in vertices: if vertice not in visited: __snake_case : int = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # return sort return sort if __name__ == "__main__": _snake_case : List[Any] = topological_sort("a", [], []) print(sort)
81
0
def __UpperCAmelCase ( lowerCamelCase_ : int ) -> bool: """simple docstring""" return sum(i for i in range(1 , number // 2 + 1 ) if number % i == 0 ) == number if __name__ == "__main__": print('''Program to check whether a number is a Perfect number or not...''') UpperCamelCase__ : Optional[int] = int(input('''Enter number: ''').strip()) print(F"""{number} is {"" if perfect(number) else "not "}a Perfect Number.""")
105
def lowerCAmelCase_ ( ): return [ a * b * (1_0_0_0 - a - b) for a in range(1 , 9_9_9 ) for b in range(__lowerCamelCase , 9_9_9 ) if (a * a + b * b == (1_0_0_0 - a - b) ** 2) ][0] if __name__ == "__main__": print(f'''{solution() = }''')
81
0
from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case :Tuple =logging.get_logger(__name__) __snake_case :str ={ 'microsoft/markuplm-base': 'https://huggingface.co/microsoft/markuplm-base/resolve/main/config.json', 'microsoft/markuplm-large': 'https://huggingface.co/microsoft/markuplm-large/resolve/main/config.json', } class lowerCAmelCase__ ( _lowerCamelCase ): A_ : List[Any] = 'markuplm' def __init__( self : Optional[Any] , __UpperCamelCase : List[str]=30_522 , __UpperCamelCase : Tuple=768 , __UpperCamelCase : Union[str, Any]=12 , __UpperCamelCase : Any=12 , __UpperCamelCase : Optional[int]=3_072 , __UpperCamelCase : Tuple="gelu" , __UpperCamelCase : Optional[int]=0.1 , __UpperCamelCase : Tuple=0.1 , __UpperCamelCase : Any=512 , __UpperCamelCase : int=2 , __UpperCamelCase : List[Any]=0.0_2 , __UpperCamelCase : Optional[int]=1e-12 , __UpperCamelCase : Tuple=0 , __UpperCamelCase : List[str]=0 , __UpperCamelCase : List[str]=2 , __UpperCamelCase : Optional[Any]=256 , __UpperCamelCase : int=1_024 , __UpperCamelCase : Union[str, Any]=216 , __UpperCamelCase : Optional[Any]=1_001 , __UpperCamelCase : Any=32 , __UpperCamelCase : Union[str, Any]=50 , __UpperCamelCase : Tuple="absolute" , __UpperCamelCase : int=True , __UpperCamelCase : List[str]=None , **__UpperCamelCase : Optional[Any] , ) -> Tuple: super().__init__( pad_token_id=__UpperCamelCase , bos_token_id=__UpperCamelCase , eos_token_id=__UpperCamelCase , **__UpperCamelCase , ) A = vocab_size A = hidden_size A = num_hidden_layers A = num_attention_heads A = hidden_act A = intermediate_size A = hidden_dropout_prob A = attention_probs_dropout_prob A = max_position_embeddings A = type_vocab_size A = initializer_range A = layer_norm_eps A = position_embedding_type A = use_cache A = classifier_dropout # additional properties A = max_depth A = max_xpath_tag_unit_embeddings A = max_xpath_subs_unit_embeddings A = tag_pad_id A = subs_pad_id A = xpath_unit_hidden_size
106
from math import sqrt def lowerCAmelCase_ ( __lowerCamelCase = 1_0_0_0_0_0_0 ): __snake_case : int = 0 __snake_case : int = 0 __snake_case : int while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(__lowerCamelCase , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(f'''{solution() = }''')
81
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : Dict = logging.get_logger(__name__) _UpperCAmelCase : List[str] = { '''facebook/nllb-moe-54B''': '''https://huggingface.co/facebook/nllb-moe-54b/resolve/main/config.json''', } class lowercase_ ( _UpperCamelCase ): """simple docstring""" __lowerCAmelCase = "nllb-moe" __lowerCAmelCase = ["past_key_values"] __lowerCAmelCase = {"num_attention_heads": "encoder_attention_heads", "hidden_size": "d_model"} def __init__( self : List[Any], UpperCamelCase__ : str=12_81_12, UpperCamelCase__ : str=10_24, UpperCamelCase__ : Optional[int]=12, UpperCamelCase__ : Any=40_96, UpperCamelCase__ : Union[str, Any]=16, UpperCamelCase__ : Dict=12, UpperCamelCase__ : Union[str, Any]=40_96, UpperCamelCase__ : Any=16, UpperCamelCase__ : Tuple=0.05, UpperCamelCase__ : Dict=0.05, UpperCamelCase__ : List[Any]=True, UpperCamelCase__ : Union[str, Any]=True, UpperCamelCase__ : List[Any]="relu", UpperCamelCase__ : int=10_24, UpperCamelCase__ : Any=0.1, UpperCamelCase__ : Optional[Any]=0.1, UpperCamelCase__ : Dict=0.0, UpperCamelCase__ : List[Any]=0.02, UpperCamelCase__ : str=2, UpperCamelCase__ : Optional[int]=True, UpperCamelCase__ : Optional[int]=False, UpperCamelCase__ : int="float32", UpperCamelCase__ : Optional[Any]=False, UpperCamelCase__ : Dict=1_28, UpperCamelCase__ : Optional[int]=64, UpperCamelCase__ : Tuple=4, UpperCamelCase__ : Any=4, UpperCamelCase__ : Dict=0.001, UpperCamelCase__ : List[str]=0.001, UpperCamelCase__ : Union[str, Any]="all", UpperCamelCase__ : Tuple=False, UpperCamelCase__ : Dict=False, UpperCamelCase__ : Any=1.0, UpperCamelCase__ : Tuple=0.2, UpperCamelCase__ : Optional[int]=1, UpperCamelCase__ : int=0, UpperCamelCase__ : List[str]=2, UpperCamelCase__ : str=False, **UpperCamelCase__ : Union[str, Any], ) -> Tuple: _A = vocab_size _A = max_position_embeddings _A = d_model _A = encoder_ffn_dim _A = encoder_layers _A = encoder_attention_heads _A = decoder_ffn_dim _A = decoder_layers _A = decoder_attention_heads _A = dropout _A = attention_dropout _A = activation_dropout _A = activation_function _A = init_std _A = encoder_layerdrop _A = decoder_layerdrop _A = use_cache _A = encoder_layers _A = scale_embedding # scale factor will be sqrt(d_model) if True _A = router_z_loss_coef _A = router_aux_loss_coef _A = decoder_sparse_step _A = encoder_sparse_step _A = num_experts _A = expert_capacity _A = router_bias if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(f'`router_dtype` must be one of \'float32\', \'float16\' or \'bfloat16\', got {router_dtype}' ) _A = router_dtype _A = router_ignore_padding_tokens _A = batch_prioritized_routing _A = second_expert_policy _A = normalize_router_prob_before_dropping _A = moe_eval_capacity_token_fraction _A = moe_token_dropout _A = output_router_logits super().__init__( pad_token_id=UpperCamelCase__, bos_token_id=UpperCamelCase__, eos_token_id=UpperCamelCase__, is_encoder_decoder=UpperCamelCase__, decoder_start_token_id=UpperCamelCase__, **UpperCamelCase__, )
107
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class a (_lowerCAmelCase ): """simple docstring""" def __snake_case ( self : str ) -> str: __snake_case : Tuple = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCamelCase , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "neck_hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "num_attention_heads" ) ) class a : """simple docstring""" def __init__( self : Optional[int] , lowerCamelCase : List[str] , lowerCamelCase : Tuple=13 , lowerCamelCase : str=32 , lowerCamelCase : Dict=2 , lowerCamelCase : List[str]=3 , lowerCamelCase : Any=640 , lowerCamelCase : Optional[Any]=4 , lowerCamelCase : Tuple="silu" , lowerCamelCase : int=3 , lowerCamelCase : Dict=32 , lowerCamelCase : str=0.1 , lowerCamelCase : Optional[int]=0.1 , lowerCamelCase : Optional[Any]=0.1 , lowerCamelCase : Dict=0.02 , lowerCamelCase : Union[str, Any]=True , lowerCamelCase : Optional[int]=True , lowerCamelCase : Union[str, Any]=10 , lowerCamelCase : int=None , ) -> str: __snake_case : Optional[Any] = parent __snake_case : Optional[Any] = batch_size __snake_case : Any = image_size __snake_case : List[Any] = patch_size __snake_case : Any = num_channels __snake_case : Union[str, Any] = last_hidden_size __snake_case : Any = num_attention_heads __snake_case : Any = hidden_act __snake_case : Tuple = conv_kernel_size __snake_case : Any = output_stride __snake_case : Any = hidden_dropout_prob __snake_case : List[Any] = attention_probs_dropout_prob __snake_case : Optional[Any] = classifier_dropout_prob __snake_case : Union[str, Any] = use_labels __snake_case : Optional[int] = is_training __snake_case : Dict = num_labels __snake_case : Any = initializer_range __snake_case : Optional[int] = scope def __snake_case ( self : str ) -> Union[str, Any]: __snake_case : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : List[Any] = None __snake_case : Optional[int] = None if self.use_labels: __snake_case : Optional[int] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : Any = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __snake_case : Optional[Any] = self.get_config() return config, pixel_values, labels, pixel_labels def __snake_case ( self : Any ) -> Union[str, Any]: return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __snake_case ( self : Any , lowerCamelCase : Tuple , lowerCamelCase : int , lowerCamelCase : Tuple , lowerCamelCase : Optional[int] ) -> Dict: __snake_case : List[Any] = MobileViTModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[str] = model(lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[Any] , lowerCamelCase : List[str] , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] , lowerCamelCase : Tuple ) -> List[str]: __snake_case : str = self.num_labels __snake_case : List[Any] = MobileViTForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : Dict , lowerCamelCase : Any , lowerCamelCase : Any , lowerCamelCase : Optional[Any] , lowerCamelCase : Dict ) -> Dict: __snake_case : Union[str, Any] = self.num_labels __snake_case : Optional[int] = MobileViTForSemanticSegmentation(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Tuple = model(lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[int] ) -> List[Any]: __snake_case : Optional[Any] = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case , __snake_case : Union[str, Any] = config_and_inputs __snake_case : Dict = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) __UpperCAmelCase : Optional[Any] = ( { "feature-extraction": MobileViTModel, "image-classification": MobileViTForImageClassification, "image-segmentation": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) __UpperCAmelCase : List[str] = False __UpperCAmelCase : int = False __UpperCAmelCase : Optional[int] = False __UpperCAmelCase : Optional[int] = False def __snake_case ( self : Optional[int] ) -> Dict: __snake_case : Tuple = MobileViTModelTester(self ) __snake_case : Any = MobileViTConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase ) def __snake_case ( self : Optional[int] ) -> Dict: self.config_tester.run_common_tests() @unittest.skip(reason="MobileViT does not use inputs_embeds" ) def __snake_case ( self : Dict ) -> Any: pass @unittest.skip(reason="MobileViT does not support input and output embeddings" ) def __snake_case ( self : Dict ) -> List[Any]: pass @unittest.skip(reason="MobileViT does not output attentions" ) def __snake_case ( self : int ) -> Dict: pass def __snake_case ( self : int ) -> Union[str, Any]: __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Optional[Any] = model_class(lowerCamelCase ) __snake_case : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : int = [*signature.parameters.keys()] __snake_case : List[str] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __snake_case ( self : int ) -> Tuple: pass def __snake_case ( self : Any ) -> Tuple: __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def __snake_case ( self : Any ) -> str: def check_hidden_states_output(lowerCamelCase : Union[str, Any] , lowerCamelCase : Dict , lowerCamelCase : Any ): __snake_case : int = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : int = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : Union[str, Any] = outputs.hidden_states __snake_case : int = 5 self.assertEqual(len(lowerCamelCase ) , lowerCamelCase ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. __snake_case : List[Any] = 2 for i in range(len(lowerCamelCase ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) __snake_case , __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : str = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : List[Any] = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Any ) -> Any: __snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) def __snake_case ( self : List[str] ) -> List[str]: __snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowerCamelCase ) @slow def __snake_case ( self : List[str] ) -> Any: for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : List[str] = MobileViTModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def lowerCAmelCase_ ( ): __snake_case : Optional[int] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a (unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self : str ) -> Dict: return MobileViTImageProcessor.from_pretrained("apple/mobilevit-xx-small" ) if is_vision_available() else None @slow def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : Tuple = MobileViTForImageClassification.from_pretrained("apple/mobilevit-xx-small" ).to(lowerCamelCase ) __snake_case : Optional[Any] = self.default_image_processor __snake_case : Union[str, Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Dict = model(**lowerCamelCase ) # verify the logits __snake_case : Union[str, Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([-1.93_64, -1.23_27, -0.46_53] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : str ) -> Optional[int]: __snake_case : Optional[int] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : str = model.to(lowerCamelCase ) __snake_case : int = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Optional[int] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : List[str] = model(**lowerCamelCase ) __snake_case : Union[str, Any] = outputs.logits # verify the logits __snake_case : Tuple = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , lowerCamelCase ) __snake_case : Union[str, Any] = torch.tensor( [ [[6.97_13, 6.97_86, 7.24_22], [7.28_93, 7.28_25, 7.44_46], [7.65_80, 7.87_97, 7.94_20]], [[-10.68_69, -10.32_50, -10.34_71], [-10.42_28, -9.98_68, -9.71_32], [-11.04_05, -11.02_21, -10.73_18]], [[-3.30_89, -2.85_39, -2.67_40], [-3.27_06, -2.56_21, -2.51_08], [-3.25_34, -2.66_15, -2.66_51]], ] , device=lowerCamelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : Union[str, Any] ) -> Optional[int]: __snake_case : Optional[Any] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Tuple = model.to(lowerCamelCase ) __snake_case : Dict = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : List[Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Any = model(**lowerCamelCase ) __snake_case : Dict = outputs.logits.detach().cpu() __snake_case : Any = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase , target_sizes=[(50, 60)] ) __snake_case : int = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , lowerCamelCase ) __snake_case : List[str] = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase ) __snake_case : Optional[int] = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , lowerCamelCase )
81
0
def _SCREAMING_SNAKE_CASE ( __snake_case , __snake_case ) -> str: _UpperCAmelCase = len(__snake_case ) _UpperCAmelCase = len(__snake_case ) _UpperCAmelCase = ( first_str_length if first_str_length > second_str_length else second_str_length ) _UpperCAmelCase = [] for char_count in range(__snake_case ): if char_count < first_str_length: output_list.append(first_str[char_count] ) if char_count < second_str_length: output_list.append(second_str[char_count] ) return "".join(__snake_case ) if __name__ == "__main__": print(alternative_string_arrange('''AB''', '''XYZ'''), end=''' ''')
108
import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor _snake_case : int = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[int] , *lowerCamelCase : List[Any] , **lowerCamelCase : int ) -> None: warnings.warn( "The class ImageGPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use ImageGPTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
'''simple docstring''' import inspect import unittest from transformers import RegNetConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import RegNetForImageClassification, RegNetModel from transformers.models.regnet.modeling_regnet import REGNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __a : def __init__( self : Any ,lowerCamelCase : Any ,lowerCamelCase : int=3 ,lowerCamelCase : Optional[int]=32 ,lowerCamelCase : Any=3 ,lowerCamelCase : Dict=10 ,lowerCamelCase : Union[str, Any]=[10, 20, 30, 40] ,lowerCamelCase : Any=[1, 1, 2, 1] ,lowerCamelCase : int=True ,lowerCamelCase : Optional[Any]=True ,lowerCamelCase : Dict="relu" ,lowerCamelCase : str=3 ,lowerCamelCase : Union[str, Any]=None ,): '''simple docstring''' __SCREAMING_SNAKE_CASE = parent __SCREAMING_SNAKE_CASE = batch_size __SCREAMING_SNAKE_CASE = image_size __SCREAMING_SNAKE_CASE = num_channels __SCREAMING_SNAKE_CASE = embeddings_size __SCREAMING_SNAKE_CASE = hidden_sizes __SCREAMING_SNAKE_CASE = depths __SCREAMING_SNAKE_CASE = is_training __SCREAMING_SNAKE_CASE = use_labels __SCREAMING_SNAKE_CASE = hidden_act __SCREAMING_SNAKE_CASE = num_labels __SCREAMING_SNAKE_CASE = scope __SCREAMING_SNAKE_CASE = len(lowerCamelCase ) def UpperCAmelCase__ ( self : Union[str, Any] ): '''simple docstring''' __SCREAMING_SNAKE_CASE = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __SCREAMING_SNAKE_CASE = None if self.use_labels: __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] ,self.num_labels ) __SCREAMING_SNAKE_CASE = self.get_config() return config, pixel_values, labels def UpperCAmelCase__ ( self : Any ): '''simple docstring''' return RegNetConfig( num_channels=self.num_channels ,embeddings_size=self.embeddings_size ,hidden_sizes=self.hidden_sizes ,depths=self.depths ,hidden_act=self.hidden_act ,num_labels=self.num_labels ,) def UpperCAmelCase__ ( self : Tuple ,lowerCamelCase : int ,lowerCamelCase : List[str] ,lowerCamelCase : Optional[int] ): '''simple docstring''' __SCREAMING_SNAKE_CASE = RegNetModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __SCREAMING_SNAKE_CASE = model(lowerCamelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape ,(self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) ,) def UpperCAmelCase__ ( self : List[Any] ,lowerCamelCase : int ,lowerCamelCase : Dict ,lowerCamelCase : Any ): '''simple docstring''' __SCREAMING_SNAKE_CASE = self.num_labels __SCREAMING_SNAKE_CASE = RegNetForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __SCREAMING_SNAKE_CASE = model(lowerCamelCase ,labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def UpperCAmelCase__ ( self : Dict ): '''simple docstring''' __SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = config_and_inputs __SCREAMING_SNAKE_CASE = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __a ( _snake_case, _snake_case, unittest.TestCase ): __UpperCamelCase : List[str] = (RegNetModel, RegNetForImageClassification) if is_torch_available() else () __UpperCamelCase : Optional[int] = ( {'feature-extraction': RegNetModel, 'image-classification': RegNetForImageClassification} if is_torch_available() else {} ) __UpperCamelCase : Dict = False __UpperCamelCase : str = False __UpperCamelCase : Any = False __UpperCamelCase : Tuple = False def UpperCAmelCase__ ( self : List[str] ): '''simple docstring''' __SCREAMING_SNAKE_CASE = RegNetModelTester(self ) __SCREAMING_SNAKE_CASE = ConfigTester(self ,config_class=lowerCamelCase ,has_text_modality=lowerCamelCase ) def UpperCAmelCase__ ( self : Optional[int] ): '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCAmelCase__ ( self : Any ): '''simple docstring''' return @unittest.skip(reason="""RegNet does not use inputs_embeds""" ) def UpperCAmelCase__ ( self : str ): '''simple docstring''' pass @unittest.skip(reason="""RegNet does not support input and output embeddings""" ) def UpperCAmelCase__ ( self : Tuple ): '''simple docstring''' pass def UpperCAmelCase__ ( self : Tuple ): '''simple docstring''' __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __SCREAMING_SNAKE_CASE = model_class(lowerCamelCase ) __SCREAMING_SNAKE_CASE = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __SCREAMING_SNAKE_CASE = [*signature.parameters.keys()] __SCREAMING_SNAKE_CASE = ["""pixel_values"""] self.assertListEqual(arg_names[:1] ,lowerCamelCase ) def UpperCAmelCase__ ( self : Any ): '''simple docstring''' __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def UpperCAmelCase__ ( self : int ): '''simple docstring''' __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __SCREAMING_SNAKE_CASE = model_class(config=lowerCamelCase ) for name, module in model.named_modules(): if isinstance(lowerCamelCase ,(nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) ,msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" ,) self.assertTrue( torch.all(module.bias == 0 ) ,msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" ,) def UpperCAmelCase__ ( self : Tuple ): '''simple docstring''' def check_hidden_states_output(lowerCamelCase : Union[str, Any] ,lowerCamelCase : Tuple ,lowerCamelCase : Optional[int] ): __SCREAMING_SNAKE_CASE = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(**self._prepare_for_class(lowerCamelCase ,lowerCamelCase ) ) __SCREAMING_SNAKE_CASE = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __SCREAMING_SNAKE_CASE = self.model_tester.num_stages self.assertEqual(len(lowerCamelCase ) ,expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) ,[self.model_tester.image_size // 2, self.model_tester.image_size // 2] ,) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() __SCREAMING_SNAKE_CASE = ["""basic""", """bottleneck"""] for model_class in self.all_model_classes: for layer_type in layers_type: __SCREAMING_SNAKE_CASE = layer_type __SCREAMING_SNAKE_CASE = True check_hidden_states_output(lowerCamelCase ,lowerCamelCase ,lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __SCREAMING_SNAKE_CASE = True check_hidden_states_output(lowerCamelCase ,lowerCamelCase ,lowerCamelCase ) def UpperCAmelCase__ ( self : Dict ): '''simple docstring''' __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) @slow def UpperCAmelCase__ ( self : Union[str, Any] ): '''simple docstring''' for model_name in REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __SCREAMING_SNAKE_CASE = RegNetModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def __magic_name__ ( ) -> List[Any]: '''simple docstring''' __SCREAMING_SNAKE_CASE = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class __a ( unittest.TestCase ): @cached_property def UpperCAmelCase__ ( self : List[str] ): '''simple docstring''' return ( AutoImageProcessor.from_pretrained(REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def UpperCAmelCase__ ( self : int ): '''simple docstring''' __SCREAMING_SNAKE_CASE = RegNetForImageClassification.from_pretrained(REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(lowerCamelCase ) __SCREAMING_SNAKE_CASE = self.default_image_processor __SCREAMING_SNAKE_CASE = prepare_img() __SCREAMING_SNAKE_CASE = image_processor(images=lowerCamelCase ,return_tensors="""pt""" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(**lowerCamelCase ) # verify the logits __SCREAMING_SNAKE_CASE = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape ,lowerCamelCase ) __SCREAMING_SNAKE_CASE = torch.tensor([-0.4_180, -1.5_051, -3.4_836] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowerCamelCase ,atol=1E-4 ) )
109
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _snake_case : Optional[Any] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[str] = ["pixel_values"] def __init__( self : List[Any] , lowerCamelCase : bool = True , lowerCamelCase : Union[int, float] = 1 / 255 , lowerCamelCase : bool = True , lowerCamelCase : int = 8 , **lowerCamelCase : Tuple , ) -> None: super().__init__(**lowerCamelCase ) __snake_case : Dict = do_rescale __snake_case : Dict = rescale_factor __snake_case : Optional[Any] = do_pad __snake_case : Tuple = pad_size def __snake_case ( self : Dict , lowerCamelCase : np.ndarray , lowerCamelCase : float , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase : Optional[int] ) -> np.ndarray: return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def __snake_case ( self : Optional[Any] , lowerCamelCase : np.ndarray , lowerCamelCase : int , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None ) -> Tuple: __snake_case , __snake_case : List[str] = get_image_size(lowerCamelCase ) __snake_case : Optional[Any] = (old_height // size + 1) * size - old_height __snake_case : List[Any] = (old_width // size + 1) * size - old_width return pad(lowerCamelCase , ((0, pad_height), (0, pad_width)) , mode="symmetric" , data_format=lowerCamelCase ) def __snake_case ( self : Tuple , lowerCamelCase : ImageInput , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[float] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[int] = None , lowerCamelCase : Optional[Union[str, TensorType]] = None , lowerCamelCase : Union[str, ChannelDimension] = ChannelDimension.FIRST , **lowerCamelCase : Union[str, Any] , ) -> List[str]: __snake_case : int = do_rescale if do_rescale is not None else self.do_rescale __snake_case : Dict = rescale_factor if rescale_factor is not None else self.rescale_factor __snake_case : str = do_pad if do_pad is not None else self.do_pad __snake_case : Any = pad_size if pad_size is not None else self.pad_size __snake_case : int = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) # All transformations expect numpy arrays. __snake_case : str = [to_numpy_array(lowerCamelCase ) for image in images] if do_rescale: __snake_case : Optional[int] = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_pad: __snake_case : Optional[Any] = [self.pad(lowerCamelCase , size=lowerCamelCase ) for image in images] __snake_case : int = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __snake_case : Union[str, Any] = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
81
0
"""simple docstring""" UpperCamelCase__ = { 'meter': 'm', 'kilometer': 'km', 'megametre': 'Mm', 'gigametre': 'Gm', 'terametre': 'Tm', 'petametre': 'Pm', 'exametre': 'Em', 'zettametre': 'Zm', 'yottametre': 'Ym', } # Exponent of the factor(meter) UpperCamelCase__ = { 'm': 0, 'km': 3, 'Mm': 6, 'Gm': 9, 'Tm': 12, 'Pm': 15, 'Em': 18, 'Zm': 21, 'Ym': 24, } def lowerCamelCase ( _snake_case ,_snake_case ,_snake_case ): UpperCAmelCase__ : int = from_type.lower().strip('s' ) UpperCAmelCase__ : List[Any] = to_type.lower().strip('s' ) UpperCAmelCase__ : Optional[int] = UNIT_SYMBOL.get(_snake_case ,_snake_case ) UpperCAmelCase__ : Optional[Any] = UNIT_SYMBOL.get(_snake_case ,_snake_case ) if from_sanitized not in METRIC_CONVERSION: UpperCAmelCase__ : Optional[int] = ( F'''Invalid \'from_type\' value: {from_type!r}.\n''' F'''Conversion abbreviations are: {', '.join(_snake_case )}''' ) raise ValueError(_snake_case ) if to_sanitized not in METRIC_CONVERSION: UpperCAmelCase__ : int = ( F'''Invalid \'to_type\' value: {to_type!r}.\n''' F'''Conversion abbreviations are: {', '.join(_snake_case )}''' ) raise ValueError(_snake_case ) UpperCAmelCase__ : str = METRIC_CONVERSION[from_sanitized] UpperCAmelCase__ : Optional[int] = METRIC_CONVERSION[to_sanitized] UpperCAmelCase__ : int = 1 if from_exponent > to_exponent: UpperCAmelCase__ : List[str] = from_exponent - to_exponent else: UpperCAmelCase__ : int = -(to_exponent - from_exponent) return value * pow(10 ,_snake_case ) if __name__ == "__main__": from doctest import testmod testmod()
110
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case : Any = { "configuration_informer": [ "INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "InformerConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "InformerForPrediction", "InformerModel", "InformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys _snake_case : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
0
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) _UpperCAmelCase : List[Any] = "\\n Text data.\n Second line of data." _UpperCAmelCase : Tuple = "file" @pytest.fixture(scope='''session''' ) def lowerCAmelCase_ (lowercase__ : List[Any] ) -> Tuple: '''simple docstring''' lowerCAmelCase__ = tmp_path_factory.mktemp('''data''' ) / (FILE_PATH + ".zstd") lowerCAmelCase__ = bytes(__lowerCamelCase , '''utf-8''' ) with zstd.open(__lowerCamelCase , '''wb''' ) as f: f.write(__lowerCamelCase ) return path @pytest.fixture def lowerCAmelCase_ (lowercase__ : Tuple ) -> Optional[int]: '''simple docstring''' with open(os.path.join(tmpfs.local_root_dir , __lowerCamelCase ) , '''w''' ) as f: f.write(__lowerCamelCase ) return FILE_PATH @pytest.mark.parametrize('''compression_format''' , ['''gzip''', '''xz''', '''zstd'''] ) def lowerCAmelCase_ (lowercase__ : Dict , lowercase__ : Optional[Any] , lowercase__ : int , lowercase__ : List[str] , lowercase__ : Dict , lowercase__ : str ) -> List[str]: '''simple docstring''' lowerCAmelCase__ = {"gzip": gz_file, "xz": xz_file, "zstd": zstd_path} lowerCAmelCase__ = input_paths[compression_format] lowerCAmelCase__ = tmp_path / "cache" lowerCAmelCase__ = DownloadConfig(cache_dir=__lowerCamelCase , extract_compressed_file=__lowerCamelCase ) lowerCAmelCase__ = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) with open(__lowerCamelCase ) as f: lowerCAmelCase__ = f.read() with open(__lowerCamelCase ) as f: lowerCAmelCase__ = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize('''default_extracted''' , [True, False] ) @pytest.mark.parametrize('''default_cache_dir''' , [True, False] ) def lowerCAmelCase_ (lowercase__ : int , lowercase__ : Any , lowercase__ : Union[str, Any] , lowercase__ : int , lowercase__ : str ) -> int: '''simple docstring''' lowerCAmelCase__ = "custom_cache" lowerCAmelCase__ = "custom_extracted_dir" lowerCAmelCase__ = tmp_path / "custom_extracted_path" if default_extracted: lowerCAmelCase__ = ("downloads" if default_cache_dir else custom_cache_dir, "extracted") else: monkeypatch.setattr('''datasets.config.EXTRACTED_DATASETS_DIR''' , __lowerCamelCase ) monkeypatch.setattr('''datasets.config.EXTRACTED_DATASETS_PATH''' , str(__lowerCamelCase ) ) lowerCAmelCase__ = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) lowerCAmelCase__ = xz_file lowerCAmelCase__ = ( DownloadConfig(extract_compressed_file=__lowerCamelCase ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=__lowerCamelCase ) ) lowerCAmelCase__ = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) assert Path(__lowerCamelCase ).parent.parts[-2:] == expected def lowerCAmelCase_ (lowercase__ : Optional[int] ) -> Dict: '''simple docstring''' lowerCAmelCase__ = str(Path(__lowerCamelCase ).resolve() ) assert cached_path(__lowerCamelCase ) == text_file # relative path lowerCAmelCase__ = str(Path(__lowerCamelCase ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(__lowerCamelCase ) == text_file def lowerCAmelCase_ (lowercase__ : Any ) -> List[Any]: '''simple docstring''' lowerCAmelCase__ = str(tmp_path.resolve() / '''__missing_file__.txt''' ) with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) # relative path lowerCAmelCase__ = "./__missing_file__.txt" with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) def lowerCAmelCase_ (lowercase__ : List[Any] ) -> int: '''simple docstring''' lowerCAmelCase__ = get_from_cache(f'tmp://{tmpfs_file}' ) with open(__lowerCamelCase ) as f: lowerCAmelCase__ = f.read() assert output_file_content == FILE_CONTENT @patch('''datasets.config.HF_DATASETS_OFFLINE''' , __lowerCamelCase ) def lowerCAmelCase_ () -> str: '''simple docstring''' with pytest.raises(__lowerCamelCase ): cached_path('''https://huggingface.co''' ) @patch('''datasets.config.HF_DATASETS_OFFLINE''' , __lowerCamelCase ) def lowerCAmelCase_ (lowercase__ : Dict ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase__ = tmp_path_factory.mktemp('''data''' ) / "file.html" with pytest.raises(__lowerCamelCase ): http_get('''https://huggingface.co''' , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): http_head('''https://huggingface.co''' ) @patch('''datasets.config.HF_DATASETS_OFFLINE''' , __lowerCamelCase ) def lowerCAmelCase_ (lowercase__ : Optional[Any] ) -> Dict: '''simple docstring''' lowerCAmelCase__ = tmp_path_factory.mktemp('''data''' ) / "file.html" with pytest.raises(__lowerCamelCase ): ftp_get('''ftp://huggingface.co''' , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): ftp_head('''ftp://huggingface.co''' ) @patch('''datasets.config.HF_DATASETS_OFFLINE''' , __lowerCamelCase ) def lowerCAmelCase_ (lowercase__ : Optional[int] ) -> Tuple: '''simple docstring''' lowerCAmelCase__ = tmp_path_factory.mktemp('''data''' ) / "file.html" with pytest.raises(__lowerCamelCase ): fsspec_get('''s3://huggingface.co''' , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): fsspec_head('''s3://huggingface.co''' )
668
from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging _snake_case : Any = logging.get_logger(__name__) # pylint: disable=invalid-name class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[Any] , lowerCamelCase : CLIPSegForImageSegmentation , lowerCamelCase : CLIPSegProcessor , lowerCamelCase : AutoencoderKL , lowerCamelCase : CLIPTextModel , lowerCamelCase : CLIPTokenizer , lowerCamelCase : UNetaDConditionModel , lowerCamelCase : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , lowerCamelCase : StableDiffusionSafetyChecker , lowerCamelCase : CLIPImageProcessor , ) -> Tuple: super().__init__() if hasattr(scheduler.config , "steps_offset" ) and scheduler.config.steps_offset != 1: __snake_case : Tuple = ( F'The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`' F' should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure ' "to update the config accordingly as leaving `steps_offset` might led to incorrect results" " in future versions. If you have downloaded this checkpoint from the Hugging Face Hub," " it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`" " file" ) deprecate("steps_offset!=1" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : Any = dict(scheduler.config ) __snake_case : List[Any] = 1 __snake_case : Tuple = FrozenDict(lowerCamelCase ) if hasattr(scheduler.config , "skip_prk_steps" ) and scheduler.config.skip_prk_steps is False: __snake_case : List[str] = ( F'The configuration file of this scheduler: {scheduler} has not set the configuration' " `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make" " sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to" " incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face" " Hub, it would be very nice if you could open a Pull request for the" " `scheduler/scheduler_config.json` file" ) deprecate("skip_prk_steps not set" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : List[str] = dict(scheduler.config ) __snake_case : List[str] = True __snake_case : Any = FrozenDict(lowerCamelCase ) if safety_checker is None: logger.warning( F'You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure' " that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered" " results in services or applications open to the public. Both the diffusers team and Hugging Face" " strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling" " it only for use-cases that involve analyzing network behavior or auditing its results. For more" " information, please have a look at https://github.com/huggingface/diffusers/pull/254 ." ) self.register_modules( segmentation_model=lowerCamelCase , segmentation_processor=lowerCamelCase , vae=lowerCamelCase , text_encoder=lowerCamelCase , tokenizer=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase , safety_checker=lowerCamelCase , feature_extractor=lowerCamelCase , ) def __snake_case ( self : Dict , lowerCamelCase : Optional[Union[str, int]] = "auto" ) -> Tuple: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __snake_case : Any = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowerCamelCase ) def __snake_case ( self : List[Any] ) -> Any: self.enable_attention_slicing(lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> str: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) __snake_case : Optional[int] = torch.device("cuda" ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(lowerCamelCase , lowerCamelCase ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __snake_case ( self : int ) -> Any: if self.device != torch.device("meta" ) or not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowerCamelCase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self : List[Any] , lowerCamelCase : Union[str, List[str]] , lowerCamelCase : Union[torch.FloatTensor, PIL.Image.Image] , lowerCamelCase : str , lowerCamelCase : int = 512 , lowerCamelCase : int = 512 , lowerCamelCase : int = 50 , lowerCamelCase : float = 7.5 , lowerCamelCase : Optional[Union[str, List[str]]] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : float = 0.0 , lowerCamelCase : Optional[torch.Generator] = None , lowerCamelCase : Optional[torch.FloatTensor] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , lowerCamelCase : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , lowerCamelCase : int = 1 , **lowerCamelCase : Dict , ) -> List[str]: __snake_case : Tuple = self.segmentation_processor( text=[text] , images=[image] , padding="max_length" , return_tensors="pt" ).to(self.device ) __snake_case : str = self.segmentation_model(**lowerCamelCase ) __snake_case : Dict = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() __snake_case : List[Any] = self.numpy_to_pil(lowerCamelCase )[0].resize(image.size ) # Run inpainting pipeline with the generated mask __snake_case : Tuple = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=lowerCamelCase , image=lowerCamelCase , mask_image=lowerCamelCase , height=lowerCamelCase , width=lowerCamelCase , num_inference_steps=lowerCamelCase , guidance_scale=lowerCamelCase , negative_prompt=lowerCamelCase , num_images_per_prompt=lowerCamelCase , eta=lowerCamelCase , generator=lowerCamelCase , latents=lowerCamelCase , output_type=lowerCamelCase , return_dict=lowerCamelCase , callback=lowerCamelCase , callback_steps=lowerCamelCase , )
81
0
'''simple docstring''' from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass class _A ( _lowerCAmelCase ): lowercase__: torch.FloatTensor class _A ( nn.Module ): def __init__( self : int , __magic_name__ : List[Any]=3 , __magic_name__ : str=3 , __magic_name__ : Optional[int]=("DownEncoderBlock2D",) , __magic_name__ : str=(64,) , __magic_name__ : Optional[Any]=2 , __magic_name__ : Any=32 , __magic_name__ : Any="silu" , __magic_name__ : List[str]=True , ) -> str: """simple docstring""" super().__init__() __snake_case : Union[str, Any] = layers_per_block __snake_case : str = torch.nn.Convad( __magic_name__ , block_out_channels[0] , kernel_size=3 , stride=1 , padding=1 , ) __snake_case : Optional[int] = None __snake_case : Any = nn.ModuleList([] ) # down __snake_case : Any = block_out_channels[0] for i, down_block_type in enumerate(__magic_name__ ): __snake_case : List[str] = output_channel __snake_case : Any = block_out_channels[i] __snake_case : Tuple = i == len(__magic_name__ ) - 1 __snake_case : Dict = get_down_block( __magic_name__ , num_layers=self.layers_per_block , in_channels=__magic_name__ , out_channels=__magic_name__ , add_downsample=not is_final_block , resnet_eps=1E-6 , downsample_padding=0 , resnet_act_fn=__magic_name__ , resnet_groups=__magic_name__ , attention_head_dim=__magic_name__ , temb_channels=__magic_name__ , ) self.down_blocks.append(__magic_name__ ) # mid __snake_case : str = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=__magic_name__ , output_scale_factor=1 , resnet_time_scale_shift="""default""" , attention_head_dim=block_out_channels[-1] , resnet_groups=__magic_name__ , temb_channels=__magic_name__ , ) # out __snake_case : Optional[int] = nn.GroupNorm(num_channels=block_out_channels[-1] , num_groups=__magic_name__ , eps=1E-6 ) __snake_case : int = nn.SiLU() __snake_case : Tuple = 2 * out_channels if double_z else out_channels __snake_case : str = nn.Convad(block_out_channels[-1] , __magic_name__ , 3 , padding=1 ) __snake_case : Any = False def lowercase__ ( self : List[Any] , __magic_name__ : int ) -> int: """simple docstring""" __snake_case : Any = x __snake_case : Union[str, Any] = self.conv_in(__magic_name__ ) if self.training and self.gradient_checkpointing: def create_custom_forward(__magic_name__ : Dict ): def custom_forward(*__magic_name__ : str ): return module(*__magic_name__ ) return custom_forward # down if is_torch_version(""">=""" , """1.11.0""" ): for down_block in self.down_blocks: __snake_case : str = torch.utils.checkpoint.checkpoint( create_custom_forward(__magic_name__ ) , __magic_name__ , use_reentrant=__magic_name__ ) # middle __snake_case : Optional[Any] = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , __magic_name__ , use_reentrant=__magic_name__ ) else: for down_block in self.down_blocks: __snake_case : Union[str, Any] = torch.utils.checkpoint.checkpoint(create_custom_forward(__magic_name__ ) , __magic_name__ ) # middle __snake_case : List[str] = torch.utils.checkpoint.checkpoint(create_custom_forward(self.mid_block ) , __magic_name__ ) else: # down for down_block in self.down_blocks: __snake_case : Union[str, Any] = down_block(__magic_name__ ) # middle __snake_case : Dict = self.mid_block(__magic_name__ ) # post-process __snake_case : Union[str, Any] = self.conv_norm_out(__magic_name__ ) __snake_case : int = self.conv_act(__magic_name__ ) __snake_case : List[Any] = self.conv_out(__magic_name__ ) return sample class _A ( nn.Module ): def __init__( self : Optional[int] , __magic_name__ : int=3 , __magic_name__ : int=3 , __magic_name__ : Any=("UpDecoderBlock2D",) , __magic_name__ : List[str]=(64,) , __magic_name__ : List[str]=2 , __magic_name__ : Any=32 , __magic_name__ : Union[str, Any]="silu" , __magic_name__ : Union[str, Any]="group" , ) -> str: """simple docstring""" super().__init__() __snake_case : Tuple = layers_per_block __snake_case : Optional[int] = nn.Convad( __magic_name__ , block_out_channels[-1] , kernel_size=3 , stride=1 , padding=1 , ) __snake_case : Dict = None __snake_case : Optional[int] = nn.ModuleList([] ) __snake_case : List[Any] = in_channels if norm_type == "spatial" else None # mid __snake_case : int = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=__magic_name__ , output_scale_factor=1 , resnet_time_scale_shift="""default""" if norm_type == """group""" else norm_type , attention_head_dim=block_out_channels[-1] , resnet_groups=__magic_name__ , temb_channels=__magic_name__ , ) # up __snake_case : Dict = list(reversed(__magic_name__ ) ) __snake_case : List[str] = reversed_block_out_channels[0] for i, up_block_type in enumerate(__magic_name__ ): __snake_case : Any = output_channel __snake_case : Tuple = reversed_block_out_channels[i] __snake_case : int = i == len(__magic_name__ ) - 1 __snake_case : List[str] = get_up_block( __magic_name__ , num_layers=self.layers_per_block + 1 , in_channels=__magic_name__ , out_channels=__magic_name__ , prev_output_channel=__magic_name__ , add_upsample=not is_final_block , resnet_eps=1E-6 , resnet_act_fn=__magic_name__ , resnet_groups=__magic_name__ , attention_head_dim=__magic_name__ , temb_channels=__magic_name__ , resnet_time_scale_shift=__magic_name__ , ) self.up_blocks.append(__magic_name__ ) __snake_case : int = output_channel # out if norm_type == "spatial": __snake_case : List[Any] = SpatialNorm(block_out_channels[0] , __magic_name__ ) else: __snake_case : List[str] = nn.GroupNorm(num_channels=block_out_channels[0] , num_groups=__magic_name__ , eps=1E-6 ) __snake_case : Any = nn.SiLU() __snake_case : Optional[Any] = nn.Convad(block_out_channels[0] , __magic_name__ , 3 , padding=1 ) __snake_case : Any = False def lowercase__ ( self : Dict , __magic_name__ : Any , __magic_name__ : Dict=None ) -> int: """simple docstring""" __snake_case : Union[str, Any] = z __snake_case : int = self.conv_in(__magic_name__ ) __snake_case : Optional[int] = next(iter(self.up_blocks.parameters() ) ).dtype if self.training and self.gradient_checkpointing: def create_custom_forward(__magic_name__ : str ): def custom_forward(*__magic_name__ : Optional[Any] ): return module(*__magic_name__ ) return custom_forward if is_torch_version(""">=""" , """1.11.0""" ): # middle __snake_case : str = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , __magic_name__ , __magic_name__ , use_reentrant=__magic_name__ ) __snake_case : Dict = sample.to(__magic_name__ ) # up for up_block in self.up_blocks: __snake_case : str = torch.utils.checkpoint.checkpoint( create_custom_forward(__magic_name__ ) , __magic_name__ , __magic_name__ , use_reentrant=__magic_name__ ) else: # middle __snake_case : Union[str, Any] = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , __magic_name__ , __magic_name__ ) __snake_case : List[Any] = sample.to(__magic_name__ ) # up for up_block in self.up_blocks: __snake_case : Any = torch.utils.checkpoint.checkpoint(create_custom_forward(__magic_name__ ) , __magic_name__ , __magic_name__ ) else: # middle __snake_case : Tuple = self.mid_block(__magic_name__ , __magic_name__ ) __snake_case : Union[str, Any] = sample.to(__magic_name__ ) # up for up_block in self.up_blocks: __snake_case : int = up_block(__magic_name__ , __magic_name__ ) # post-process if latent_embeds is None: __snake_case : List[str] = self.conv_norm_out(__magic_name__ ) else: __snake_case : List[Any] = self.conv_norm_out(__magic_name__ , __magic_name__ ) __snake_case : Optional[Any] = self.conv_act(__magic_name__ ) __snake_case : List[Any] = self.conv_out(__magic_name__ ) return sample class _A ( nn.Module ): def __init__( self : Optional[Any] , __magic_name__ : int , __magic_name__ : Union[str, Any] , __magic_name__ : List[Any] , __magic_name__ : Tuple=None , __magic_name__ : Dict="random" , __magic_name__ : List[str]=False , __magic_name__ : int=True ) -> List[str]: """simple docstring""" super().__init__() __snake_case : Any = n_e __snake_case : Optional[int] = vq_embed_dim __snake_case : List[str] = beta __snake_case : List[Any] = legacy __snake_case : Tuple = nn.Embedding(self.n_e , self.vq_embed_dim ) self.embedding.weight.data.uniform_(-1.0 / self.n_e , 1.0 / self.n_e ) __snake_case : Union[str, Any] = remap if self.remap is not None: self.register_buffer("""used""" , torch.tensor(np.load(self.remap ) ) ) __snake_case : List[str] = self.used.shape[0] __snake_case : Union[str, Any] = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": __snake_case : List[str] = self.re_embed __snake_case : str = self.re_embed + 1 print( f'''Remapping {self.n_e} indices to {self.re_embed} indices. ''' f'''Using {self.unknown_index} for unknown indices.''' ) else: __snake_case : List[Any] = n_e __snake_case : Union[str, Any] = sane_index_shape def lowercase__ ( self : str , __magic_name__ : Optional[Any] ) -> int: """simple docstring""" __snake_case : Union[str, Any] = inds.shape assert len(__magic_name__ ) > 1 __snake_case : Tuple = inds.reshape(ishape[0] , -1 ) __snake_case : Union[str, Any] = self.used.to(__magic_name__ ) __snake_case : Union[str, Any] = (inds[:, :, None] == used[None, None, ...]).long() __snake_case : Optional[Any] = match.argmax(-1 ) __snake_case : Tuple = match.sum(2 ) < 1 if self.unknown_index == "random": __snake_case : List[str] = torch.randint(0 , self.re_embed , size=new[unknown].shape ).to(device=new.device ) else: __snake_case : Any = self.unknown_index return new.reshape(__magic_name__ ) def lowercase__ ( self : Dict , __magic_name__ : int ) -> List[Any]: """simple docstring""" __snake_case : str = inds.shape assert len(__magic_name__ ) > 1 __snake_case : Union[str, Any] = inds.reshape(ishape[0] , -1 ) __snake_case : Any = self.used.to(__magic_name__ ) if self.re_embed > self.used.shape[0]: # extra token __snake_case : List[str] = 0 # simply set to zero __snake_case : List[Any] = torch.gather(used[None, :][inds.shape[0] * [0], :] , 1 , __magic_name__ ) return back.reshape(__magic_name__ ) def lowercase__ ( self : Union[str, Any] , __magic_name__ : Tuple ) -> Any: """simple docstring""" __snake_case : Any = z.permute(0 , 2 , 3 , 1 ).contiguous() __snake_case : Optional[Any] = z.view(-1 , self.vq_embed_dim ) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z __snake_case : Optional[int] = torch.argmin(torch.cdist(__magic_name__ , self.embedding.weight ) , dim=1 ) __snake_case : Union[str, Any] = self.embedding(__magic_name__ ).view(z.shape ) __snake_case : Any = None __snake_case : Optional[Any] = None # compute loss for embedding if not self.legacy: __snake_case : Union[str, Any] = self.beta * torch.mean((z_q.detach() - z) ** 2 ) + torch.mean((z_q - z.detach()) ** 2 ) else: __snake_case : int = torch.mean((z_q.detach() - z) ** 2 ) + self.beta * torch.mean((z_q - z.detach()) ** 2 ) # preserve gradients __snake_case : Dict = z + (z_q - z).detach() # reshape back to match original input shape __snake_case : Dict = z_q.permute(0 , 3 , 1 , 2 ).contiguous() if self.remap is not None: __snake_case : int = min_encoding_indices.reshape(z.shape[0] , -1 ) # add batch axis __snake_case : Union[str, Any] = self.remap_to_used(__magic_name__ ) __snake_case : Union[str, Any] = min_encoding_indices.reshape(-1 , 1 ) # flatten if self.sane_index_shape: __snake_case : List[Any] = min_encoding_indices.reshape(z_q.shape[0] , z_q.shape[2] , z_q.shape[3] ) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def lowercase__ ( self : Optional[Any] , __magic_name__ : Dict , __magic_name__ : Dict ) -> List[str]: """simple docstring""" if self.remap is not None: __snake_case : List[Any] = indices.reshape(shape[0] , -1 ) # add batch axis __snake_case : Dict = self.unmap_to_all(__magic_name__ ) __snake_case : List[Any] = indices.reshape(-1 ) # flatten again # get quantized latent vectors __snake_case : Optional[Any] = self.embedding(__magic_name__ ) if shape is not None: __snake_case : List[Any] = z_q.view(__magic_name__ ) # reshape back to match original input shape __snake_case : Optional[int] = z_q.permute(0 , 3 , 1 , 2 ).contiguous() return z_q class _A ( _lowerCAmelCase ): def __init__( self : List[Any] , __magic_name__ : List[Any] , __magic_name__ : List[Any]=False ) -> Optional[Any]: """simple docstring""" __snake_case : List[str] = parameters __snake_case : List[Any] = torch.chunk(__magic_name__ , 2 , dim=1 ) __snake_case : int = torch.clamp(self.logvar , -30.0 , 20.0 ) __snake_case : List[Any] = deterministic __snake_case : Tuple = torch.exp(0.5 * self.logvar ) __snake_case : Tuple = torch.exp(self.logvar ) if self.deterministic: __snake_case : str = torch.zeros_like( self.mean , device=self.parameters.device , dtype=self.parameters.dtype ) def lowercase__ ( self : int , __magic_name__ : Optional[torch.Generator] = None ) -> torch.FloatTensor: """simple docstring""" __snake_case : Tuple = randn_tensor( self.mean.shape , generator=__magic_name__ , device=self.parameters.device , dtype=self.parameters.dtype ) __snake_case : Optional[int] = self.mean + self.std * sample return x def lowercase__ ( self : List[Any] , __magic_name__ : Union[str, Any]=None ) -> Dict: """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean , 2 ) + self.var - 1.0 - self.logvar , dim=[1, 2, 3] ) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean , 2 ) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar , dim=[1, 2, 3] , ) def lowercase__ ( self : str , __magic_name__ : Union[str, Any] , __magic_name__ : Optional[Any]=[1, 2, 3] ) -> List[str]: """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) __snake_case : List[str] = np.log(2.0 * np.pi ) return 0.5 * torch.sum(logtwopi + self.logvar + torch.pow(sample - self.mean , 2 ) / self.var , dim=__magic_name__ ) def lowercase__ ( self : int ) -> Union[str, Any]: """simple docstring""" return self.mean
26
class a : """simple docstring""" def __init__( self : Tuple , lowerCamelCase : list ) -> None: __snake_case : str = set_counts __snake_case : Union[str, Any] = max(lowerCamelCase ) __snake_case : List[Any] = len(lowerCamelCase ) __snake_case : Tuple = [1] * num_sets __snake_case : Dict = list(range(lowerCamelCase ) ) def __snake_case ( self : str , lowerCamelCase : int , lowerCamelCase : int ) -> bool: __snake_case : List[Any] = self.get_parent(lowerCamelCase ) __snake_case : Tuple = self.get_parent(lowerCamelCase ) if src_parent == dst_parent: return False if self.ranks[dst_parent] >= self.ranks[src_parent]: self.set_counts[dst_parent] += self.set_counts[src_parent] __snake_case : List[str] = 0 __snake_case : List[Any] = dst_parent if self.ranks[dst_parent] == self.ranks[src_parent]: self.ranks[dst_parent] += 1 __snake_case : Dict = self.set_counts[dst_parent] else: self.set_counts[src_parent] += self.set_counts[dst_parent] __snake_case : Union[str, Any] = 0 __snake_case : Optional[int] = src_parent __snake_case : Tuple = self.set_counts[src_parent] __snake_case : str = max(self.max_set , lowerCamelCase ) return True def __snake_case ( self : int , lowerCamelCase : int ) -> int: if self.parents[disj_set] == disj_set: return disj_set __snake_case : Optional[int] = self.get_parent(self.parents[disj_set] ) return self.parents[disj_set]
81
0
'''simple docstring''' UpperCAmelCase = "\n# Transformers installation\n! pip install transformers datasets\n# To install from source instead of the last release, comment the command above and uncomment the following one.\n# ! pip install git+https://github.com/huggingface/transformers.git\n" UpperCAmelCase = [{"type": "code", "content": INSTALL_CONTENT}] UpperCAmelCase = { "{processor_class}": "FakeProcessorClass", "{model_class}": "FakeModelClass", "{object_class}": "FakeObjectClass", }
119
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class a (unittest.TestCase ): """simple docstring""" def __init__( self : int , lowerCamelCase : Any , lowerCamelCase : int=7 , lowerCamelCase : str=3 , lowerCamelCase : Optional[int]=30 , lowerCamelCase : Dict=400 , lowerCamelCase : str=True , lowerCamelCase : str=None , lowerCamelCase : Any=True , lowerCamelCase : Union[str, Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[str]=True , lowerCamelCase : Optional[int]=1 / 255 , lowerCamelCase : Any=True , ) -> str: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __snake_case : Optional[Any] = size if size is not None else {"shortest_edge": 18, "longest_edge": 1333} __snake_case : Optional[Any] = parent __snake_case : List[Any] = batch_size __snake_case : Optional[int] = num_channels __snake_case : str = min_resolution __snake_case : int = max_resolution __snake_case : int = do_resize __snake_case : Tuple = size __snake_case : Any = do_normalize __snake_case : int = image_mean __snake_case : Tuple = image_std __snake_case : Dict = do_rescale __snake_case : Optional[Any] = rescale_factor __snake_case : str = do_pad def __snake_case ( self : Any ) -> int: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def __snake_case ( self : Dict , lowerCamelCase : List[Any] , lowerCamelCase : Optional[Any]=False ) -> List[str]: if not batched: __snake_case : Dict = image_inputs[0] if isinstance(lowerCamelCase , Image.Image ): __snake_case , __snake_case : Dict = image.size else: __snake_case , __snake_case : List[str] = image.shape[1], image.shape[2] if w < h: __snake_case : Optional[int] = int(self.size["shortest_edge"] * h / w ) __snake_case : int = self.size["shortest_edge"] elif w > h: __snake_case : List[str] = self.size["shortest_edge"] __snake_case : Optional[Any] = int(self.size["shortest_edge"] * w / h ) else: __snake_case : List[Any] = self.size["shortest_edge"] __snake_case : Any = self.size["shortest_edge"] else: __snake_case : int = [] for image in image_inputs: __snake_case , __snake_case : List[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[0] )[0] __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class a (_lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ConditionalDetrImageProcessor if is_vision_available() else None def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : Optional[Any] = ConditionalDetrImageProcessingTester(self ) @property def __snake_case ( self : Any ) -> str: return self.image_processor_tester.prepare_image_processor_dict() def __snake_case ( self : Optional[Any] ) -> Optional[int]: __snake_case : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase , "image_mean" ) ) self.assertTrue(hasattr(lowerCamelCase , "image_std" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_normalize" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase , "size" ) ) def __snake_case ( self : Any ) -> Dict: __snake_case : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 1333} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) __snake_case : str = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=lowerCamelCase ) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> Dict: pass def __snake_case ( self : Tuple ) -> str: # Initialize image_processing __snake_case : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , Image.Image ) # Test not batched input __snake_case : int = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) __snake_case : Dict = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> str: # Initialize image_processing __snake_case : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __snake_case : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , numpify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , np.ndarray ) # Test not batched input __snake_case : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : List[Any] = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> List[str]: # Initialize image_processing __snake_case : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , torchify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , torch.Tensor ) # Test not batched input __snake_case : Optional[Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : List[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : int = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def __snake_case ( self : Any ) -> Optional[int]: # prepare image and target __snake_case : List[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : List[Any] = {"image_id": 39769, "annotations": target} # encode them __snake_case : List[str] = ConditionalDetrImageProcessor.from_pretrained("microsoft/conditional-detr-resnet-50" ) __snake_case : List[str] = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : Tuple = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Optional[int] = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : List[Any] = torch.tensor([58_87.96_00, 1_12_50.20_61, 48_93_53.84_38, 83_71_22.75_00, 14_79_67.51_56, 16_57_32.34_38] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : Optional[int] = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([0.55_03, 0.27_65, 0.06_04, 0.22_15] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Optional[Any] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : List[Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : Dict = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify orig_size __snake_case : int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Tuple = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) ) @slow def __snake_case ( self : str ) -> Tuple: # prepare image, target and masks_path __snake_case : Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : str = {"file_name": "000000039769.png", "image_id": 39769, "segments_info": target} __snake_case : Optional[int] = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them __snake_case : int = ConditionalDetrImageProcessor(format="coco_panoptic" ) __snake_case : str = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , masks_path=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : List[str] = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Dict = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : Any = torch.tensor([14_79_79.68_75, 16_55_27.04_69, 48_46_38.59_38, 1_12_92.93_75, 58_79.65_62, 76_34.11_47] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : str = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : Optional[Any] = torch.tensor([0.26_25, 0.54_37, 0.46_88, 0.86_25] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Tuple = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : Dict = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : int = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify masks __snake_case : List[Any] = 822873 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , lowerCamelCase ) # verify orig_size __snake_case : List[str] = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Any = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) )
81
0
import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import DetrConfig, MaskFormerConfig, SwinConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskFormerForInstanceSegmentation, MaskFormerModel if is_vision_available(): from transformers import MaskFormerImageProcessor if is_vision_available(): from PIL import Image class lowerCAmelCase__ : def __init__( self : Optional[int] , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Tuple=2 , SCREAMING_SNAKE_CASE__ : Any=True , SCREAMING_SNAKE_CASE__ : Tuple=False , SCREAMING_SNAKE_CASE__ : str=10 , SCREAMING_SNAKE_CASE__ : Optional[Any]=3 , SCREAMING_SNAKE_CASE__ : Tuple=32 * 4 , SCREAMING_SNAKE_CASE__ : List[Any]=32 * 6 , SCREAMING_SNAKE_CASE__ : Tuple=4 , SCREAMING_SNAKE_CASE__ : Tuple=32 , ) -> int: __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = is_training __lowerCamelCase = use_auxiliary_loss __lowerCamelCase = num_queries __lowerCamelCase = num_channels __lowerCamelCase = min_size __lowerCamelCase = max_size __lowerCamelCase = num_labels __lowerCamelCase = mask_feature_size def __A ( self : List[Any] ) -> int: __lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = torch.ones([self.batch_size, self.min_size, self.max_size] , device=SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=SCREAMING_SNAKE_CASE__ ) > 0.5 ).float() __lowerCamelCase = (torch.rand((self.batch_size, self.num_labels) , device=SCREAMING_SNAKE_CASE__ ) > 0.5).long() __lowerCamelCase = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def __A ( self : Any ) -> Tuple: return MaskFormerConfig.from_backbone_and_decoder_configs( backbone_config=SwinConfig( depths=[1, 1, 1, 1] , ) , decoder_config=DetrConfig( decoder_ffn_dim=1_28 , num_queries=self.num_queries , decoder_attention_heads=2 , d_model=self.mask_feature_size , ) , mask_feature_size=self.mask_feature_size , fpn_feature_size=self.mask_feature_size , num_channels=self.num_channels , num_labels=self.num_labels , ) def __A ( self : int ) -> Dict: __lowerCamelCase = self.prepare_config_and_inputs() __lowerCamelCase = {"pixel_values": pixel_values, "pixel_mask": pixel_mask} return config, inputs_dict def __A ( self : Tuple , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : List[str] ) -> List[str]: __lowerCamelCase = output.encoder_hidden_states __lowerCamelCase = output.pixel_decoder_hidden_states __lowerCamelCase = output.transformer_decoder_hidden_states self.parent.assertTrue(len(SCREAMING_SNAKE_CASE__ ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(SCREAMING_SNAKE_CASE__ ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(SCREAMING_SNAKE_CASE__ ) , config.decoder_config.decoder_layers ) def __A ( self : Union[str, Any] , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Dict=False ) -> Dict: with torch.no_grad(): __lowerCamelCase = MaskFormerModel(config=SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() __lowerCamelCase = model(pixel_values=SCREAMING_SNAKE_CASE__ , pixel_mask=SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = model(SCREAMING_SNAKE_CASE__ , output_hidden_states=SCREAMING_SNAKE_CASE__ ) # the correct shape of output.transformer_decoder_hidden_states ensure the correcteness of the # encoder and pixel decoder self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.mask_feature_size) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def __A ( self : Optional[int] , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Union[str, Any] ) -> Optional[Any]: __lowerCamelCase = MaskFormerForInstanceSegmentation(config=SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() def comm_check_on_output(SCREAMING_SNAKE_CASE__ : Union[str, Any] ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): __lowerCamelCase = model(pixel_values=SCREAMING_SNAKE_CASE__ , pixel_mask=SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = model(SCREAMING_SNAKE_CASE__ ) comm_check_on_output(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = model( pixel_values=SCREAMING_SNAKE_CASE__ , pixel_mask=SCREAMING_SNAKE_CASE__ , mask_labels=SCREAMING_SNAKE_CASE__ , class_labels=SCREAMING_SNAKE_CASE__ ) comm_check_on_output(SCREAMING_SNAKE_CASE__ ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class lowerCAmelCase__ ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): a__ : Tuple = (MaskFormerModel, MaskFormerForInstanceSegmentation) if is_torch_available() else () a__ : Dict = ( {"feature-extraction": MaskFormerModel, "image-segmentation": MaskFormerForInstanceSegmentation} if is_torch_available() else {} ) a__ : Dict = False a__ : int = False a__ : List[str] = False a__ : str = False def __A ( self : Optional[int] ) -> List[str]: __lowerCamelCase = MaskFormerModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE__ , has_text_modality=SCREAMING_SNAKE_CASE__ ) def __A ( self : int ) -> Dict: self.config_tester.run_common_tests() def __A ( self : Any ) -> Union[str, Any]: __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ , output_hidden_states=SCREAMING_SNAKE_CASE__ ) def __A ( self : Any ) -> Optional[Any]: __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskformer_instance_segmentation_head_model(*SCREAMING_SNAKE_CASE__ ) @unittest.skip(reason='''MaskFormer does not use inputs_embeds''' ) def __A ( self : List[Any] ) -> Tuple: pass @unittest.skip(reason='''MaskFormer does not have a get_input_embeddings method''' ) def __A ( self : Optional[int] ) -> List[str]: pass @unittest.skip(reason='''MaskFormer is not a generative model''' ) def __A ( self : Union[str, Any] ) -> Tuple: pass @unittest.skip(reason='''MaskFormer does not use token embeddings''' ) def __A ( self : List[Any] ) -> Union[str, Any]: pass @require_torch_multi_gpu @unittest.skip( reason='''MaskFormer has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`''' ) def __A ( self : str ) -> List[Any]: pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def __A ( self : int ) -> Dict: pass def __A ( self : int ) -> List[Any]: __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCamelCase = model_class(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCamelCase = [*signature.parameters.keys()] __lowerCamelCase = ["pixel_values"] self.assertListEqual(arg_names[:1] , SCREAMING_SNAKE_CASE__ ) @slow def __A ( self : List[Any] ) -> int: for model_name in ["facebook/maskformer-swin-small-coco"]: __lowerCamelCase = MaskFormerModel.from_pretrained(SCREAMING_SNAKE_CASE__ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE__ ) def __A ( self : Union[str, Any] ) -> str: __lowerCamelCase = (self.model_tester.min_size,) * 2 __lowerCamelCase = { "pixel_values": torch.randn((2, 3, *size) , device=SCREAMING_SNAKE_CASE__ ), "mask_labels": torch.randn((2, 10, *size) , device=SCREAMING_SNAKE_CASE__ ), "class_labels": torch.zeros(2 , 10 , device=SCREAMING_SNAKE_CASE__ ).long(), } __lowerCamelCase = MaskFormerForInstanceSegmentation(MaskFormerConfig() ).to(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = model(**SCREAMING_SNAKE_CASE__ ) self.assertTrue(outputs.loss is not None ) def __A ( self : Any ) -> Optional[Any]: __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ , output_hidden_states=SCREAMING_SNAKE_CASE__ ) def __A ( self : Any ) -> List[str]: __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCamelCase = model_class(SCREAMING_SNAKE_CASE__ ).to(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = model(**SCREAMING_SNAKE_CASE__ , output_attentions=SCREAMING_SNAKE_CASE__ ) self.assertTrue(outputs.attentions is not None ) def __A ( self : Optional[int] ) -> str: if not self.model_tester.is_training: return # only MaskFormerForInstanceSegmentation has the loss __lowerCamelCase = self.all_model_classes[1] __lowerCamelCase = self.model_tester.prepare_config_and_inputs() __lowerCamelCase = model_class(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.train() __lowerCamelCase = model(SCREAMING_SNAKE_CASE__ , mask_labels=SCREAMING_SNAKE_CASE__ , class_labels=SCREAMING_SNAKE_CASE__ ).loss loss.backward() def __A ( self : Any ) -> str: # only MaskFormerForInstanceSegmentation has the loss __lowerCamelCase = self.all_model_classes[1] __lowerCamelCase = self.model_tester.prepare_config_and_inputs() __lowerCamelCase = True __lowerCamelCase = True __lowerCamelCase = model_class(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.train() __lowerCamelCase = model(SCREAMING_SNAKE_CASE__ , mask_labels=SCREAMING_SNAKE_CASE__ , class_labels=SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() __lowerCamelCase = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() # we requires_grad=True in inputs_embeds (line 2152), the original implementation don't __lowerCamelCase = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() __lowerCamelCase = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=SCREAMING_SNAKE_CASE__ ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) SCREAMING_SNAKE_CASE__ : int = 1E-4 def __magic_name__ ( ) -> int: __lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_vision @slow class lowerCAmelCase__ ( unittest.TestCase ): @cached_property def __A ( self : Optional[Any] ) -> List[str]: return ( MaskFormerImageProcessor.from_pretrained('''facebook/maskformer-swin-small-coco''' ) if is_vision_available() else None ) def __A ( self : Dict ) -> Optional[int]: __lowerCamelCase = MaskFormerModel.from_pretrained('''facebook/maskformer-swin-small-coco''' ).to(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = self.default_image_processor __lowerCamelCase = prepare_img() __lowerCamelCase = image_processor(SCREAMING_SNAKE_CASE__ , return_tensors='''pt''' ).to(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = inputs["pixel_values"].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(SCREAMING_SNAKE_CASE__ , (1, 3, 8_00, 10_88) ) with torch.no_grad(): __lowerCamelCase = model(**SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = torch.tensor( [[-0.0482, 0.9228, 0.4951], [-0.2547, 0.8017, 0.8527], [-0.0069, 0.3385, -0.0089]] ).to(SCREAMING_SNAKE_CASE__ ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , SCREAMING_SNAKE_CASE__ , atol=SCREAMING_SNAKE_CASE__ ) ) __lowerCamelCase = torch.tensor( [[-0.8422, -0.8434, -0.9718], [-1.0144, -0.5565, -0.4195], [-1.0038, -0.4484, -0.1961]] ).to(SCREAMING_SNAKE_CASE__ ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , SCREAMING_SNAKE_CASE__ , atol=SCREAMING_SNAKE_CASE__ ) ) __lowerCamelCase = torch.tensor( [[0.2852, -0.0159, 0.9735], [0.6254, 0.1858, 0.8529], [-0.0680, -0.4116, 1.8413]] ).to(SCREAMING_SNAKE_CASE__ ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , SCREAMING_SNAKE_CASE__ , atol=SCREAMING_SNAKE_CASE__ ) ) def __A ( self : Dict ) -> List[Any]: __lowerCamelCase = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-swin-small-coco''' ) .to(SCREAMING_SNAKE_CASE__ ) .eval() ) __lowerCamelCase = self.default_image_processor __lowerCamelCase = prepare_img() __lowerCamelCase = image_processor(SCREAMING_SNAKE_CASE__ , return_tensors='''pt''' ).to(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = inputs["pixel_values"].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(SCREAMING_SNAKE_CASE__ , (1, 3, 8_00, 10_88) ) with torch.no_grad(): __lowerCamelCase = model(**SCREAMING_SNAKE_CASE__ ) # masks_queries_logits __lowerCamelCase = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) __lowerCamelCase = [ [-1.3737124, -1.7724937, -1.9364233], [-1.5977281, -1.9867939, -2.1523695], [-1.5795398, -1.9269832, -2.093942], ] __lowerCamelCase = torch.tensor(SCREAMING_SNAKE_CASE__ ).to(SCREAMING_SNAKE_CASE__ ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , SCREAMING_SNAKE_CASE__ , atol=SCREAMING_SNAKE_CASE__ ) ) # class_queries_logits __lowerCamelCase = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) __lowerCamelCase = torch.tensor( [ [1.6_512e00, -5.2_572e00, -3.3_519e00], [3.6_169e-02, -5.9_025e00, -2.9_313e00], [1.0_766e-04, -7.7_630e00, -5.1_263e00], ] ).to(SCREAMING_SNAKE_CASE__ ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , SCREAMING_SNAKE_CASE__ , atol=SCREAMING_SNAKE_CASE__ ) ) def __A ( self : Any ) -> int: __lowerCamelCase = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-resnet101-coco-stuff''' ) .to(SCREAMING_SNAKE_CASE__ ) .eval() ) __lowerCamelCase = self.default_image_processor __lowerCamelCase = prepare_img() __lowerCamelCase = image_processor(SCREAMING_SNAKE_CASE__ , return_tensors='''pt''' ).to(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = inputs["pixel_values"].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(SCREAMING_SNAKE_CASE__ , (1, 3, 8_00, 10_88) ) with torch.no_grad(): __lowerCamelCase = model(**SCREAMING_SNAKE_CASE__ ) # masks_queries_logits __lowerCamelCase = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) __lowerCamelCase = [[-0.9046, -2.6366, -4.6062], [-3.4179, -5.7890, -8.8057], [-4.9179, -7.6560, -10.7711]] __lowerCamelCase = torch.tensor(SCREAMING_SNAKE_CASE__ ).to(SCREAMING_SNAKE_CASE__ ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , SCREAMING_SNAKE_CASE__ , atol=SCREAMING_SNAKE_CASE__ ) ) # class_queries_logits __lowerCamelCase = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) __lowerCamelCase = torch.tensor( [[4.7188, -3.2585, -2.8857], [6.6871, -2.9181, -1.2487], [7.2449, -2.2764, -2.1874]] ).to(SCREAMING_SNAKE_CASE__ ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , SCREAMING_SNAKE_CASE__ , atol=SCREAMING_SNAKE_CASE__ ) ) def __A ( self : Optional[Any] ) -> List[str]: __lowerCamelCase = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-swin-small-coco''' ) .to(SCREAMING_SNAKE_CASE__ ) .eval() ) __lowerCamelCase = self.default_image_processor __lowerCamelCase = image_processor( [np.zeros((3, 8_00, 13_33) ), np.zeros((3, 8_00, 13_33) )] , segmentation_maps=[np.zeros((3_84, 3_84) ).astype(np.floataa ), np.zeros((3_84, 3_84) ).astype(np.floataa )] , return_tensors='''pt''' , ) __lowerCamelCase = inputs["pixel_values"].to(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = [el.to(SCREAMING_SNAKE_CASE__ ) for el in inputs["mask_labels"]] __lowerCamelCase = [el.to(SCREAMING_SNAKE_CASE__ ) for el in inputs["class_labels"]] with torch.no_grad(): __lowerCamelCase = model(**SCREAMING_SNAKE_CASE__ ) self.assertTrue(outputs.loss is not None )
298
from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from ... import AutoBackbone from ...modeling_outputs import SemanticSegmenterOutput from ...modeling_utils import PreTrainedModel from ...utils import add_start_docstrings, add_start_docstrings_to_model_forward, replace_return_docstrings from ...utils.backbone_utils import BackboneMixin from .configuration_upernet import UperNetConfig _snake_case : Optional[Any] = [ "openmmlab/upernet-convnext-tiny", # See all UperNet models at https://huggingface.co/models?filter=upernet ] # General docstring _snake_case : Dict = "UperNetConfig" class a (nn.Module ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : Union[int, Tuple[int, int]] , lowerCamelCase : Union[int, Tuple[int, int], str] = 0 , lowerCamelCase : bool = False , lowerCamelCase : Union[int, Tuple[int, int]] = 1 , ) -> None: super().__init__() __snake_case : Union[str, Any] = nn.Convad( in_channels=lowerCamelCase , out_channels=lowerCamelCase , kernel_size=lowerCamelCase , padding=lowerCamelCase , bias=lowerCamelCase , dilation=lowerCamelCase , ) __snake_case : Dict = nn.BatchNormad(lowerCamelCase ) __snake_case : List[Any] = nn.ReLU() def __snake_case ( self : List[Any] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : Dict = self.conv(lowerCamelCase ) __snake_case : int = self.batch_norm(lowerCamelCase ) __snake_case : Optional[Any] = self.activation(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : str , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : int ) -> None: super().__init__() __snake_case : Tuple = [ nn.AdaptiveAvgPoolad(lowerCamelCase ), UperNetConvModule(lowerCamelCase , lowerCamelCase , kernel_size=1 ), ] for i, layer in enumerate(self.layers ): self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : Dict , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : List[str] = input for layer in self.layers: __snake_case : Tuple = layer(lowerCamelCase ) return hidden_state class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Tuple[int, ...] , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : bool ) -> None: super().__init__() __snake_case : Dict = pool_scales __snake_case : List[str] = align_corners __snake_case : List[Any] = in_channels __snake_case : str = channels __snake_case : Optional[Any] = [] for i, pool_scale in enumerate(lowerCamelCase ): __snake_case : Union[str, Any] = UperNetPyramidPoolingBlock(pool_scale=lowerCamelCase , in_channels=lowerCamelCase , channels=lowerCamelCase ) self.blocks.append(lowerCamelCase ) self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> List[torch.Tensor]: __snake_case : Tuple = [] for ppm in self.blocks: __snake_case : Any = ppm(lowerCamelCase ) __snake_case : List[Any] = nn.functional.interpolate( lowerCamelCase , size=x.size()[2:] , mode="bilinear" , align_corners=self.align_corners ) ppm_outs.append(lowerCamelCase ) return ppm_outs class a (nn.Module ): """simple docstring""" def __init__( self : int , lowerCamelCase : str , lowerCamelCase : str ) -> List[str]: super().__init__() __snake_case : Dict = config __snake_case : List[str] = config.pool_scales # e.g. (1, 2, 3, 6) __snake_case : Tuple = in_channels __snake_case : str = config.hidden_size __snake_case : List[str] = False __snake_case : Any = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) # PSP Module __snake_case : Tuple = UperNetPyramidPoolingModule( self.pool_scales , self.in_channels[-1] , self.channels , align_corners=self.align_corners , ) __snake_case : List[str] = UperNetConvModule( self.in_channels[-1] + len(self.pool_scales ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) # FPN Module __snake_case : List[Any] = nn.ModuleList() __snake_case : Dict = nn.ModuleList() for in_channels in self.in_channels[:-1]: # skip the top layer __snake_case : Union[str, Any] = UperNetConvModule(lowerCamelCase , self.channels , kernel_size=1 ) __snake_case : Optional[int] = UperNetConvModule(self.channels , self.channels , kernel_size=3 , padding=1 ) self.lateral_convs.append(lowerCamelCase ) self.fpn_convs.append(lowerCamelCase ) __snake_case : int = UperNetConvModule( len(self.in_channels ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) def __snake_case ( self : List[str] ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> str: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : List[Any] , lowerCamelCase : Tuple ) -> Optional[int]: __snake_case : str = inputs[-1] __snake_case : int = [x] psp_outs.extend(self.psp_modules(lowerCamelCase ) ) __snake_case : Tuple = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Union[str, Any] = self.bottleneck(lowerCamelCase ) return output def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # build laterals __snake_case : Any = [lateral_conv(encoder_hidden_states[i] ) for i, lateral_conv in enumerate(self.lateral_convs )] laterals.append(self.psp_forward(lowerCamelCase ) ) # build top-down path __snake_case : Dict = len(lowerCamelCase ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Union[str, Any] = laterals[i - 1].shape[2:] __snake_case : Optional[int] = laterals[i - 1] + nn.functional.interpolate( laterals[i] , size=lowerCamelCase , mode="bilinear" , align_corners=self.align_corners ) # build outputs __snake_case : str = [self.fpn_convs[i](laterals[i] ) for i in range(used_backbone_levels - 1 )] # append psp feature fpn_outs.append(laterals[-1] ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Tuple = nn.functional.interpolate( fpn_outs[i] , size=fpn_outs[0].shape[2:] , mode="bilinear" , align_corners=self.align_corners ) __snake_case : str = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Optional[Any] = self.fpn_bottleneck(lowerCamelCase ) __snake_case : Tuple = self.classifier(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Optional[int] , lowerCamelCase : int = 2 , lowerCamelCase : int = 3 , lowerCamelCase : Union[int, Tuple[int, int]] = 1 ) -> None: super().__init__() __snake_case : List[Any] = config __snake_case : List[str] = config.auxiliary_in_channels __snake_case : List[Any] = config.auxiliary_channels __snake_case : Tuple = config.auxiliary_num_convs __snake_case : int = config.auxiliary_concat_input __snake_case : Optional[int] = in_index __snake_case : Tuple = (kernel_size // 2) * dilation __snake_case : Optional[int] = [] convs.append( UperNetConvModule( self.in_channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) for i in range(self.num_convs - 1 ): convs.append( UperNetConvModule( self.channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) if self.num_convs == 0: __snake_case : Union[str, Any] = nn.Identity() else: __snake_case : Any = nn.Sequential(*lowerCamelCase ) if self.concat_input: __snake_case : int = UperNetConvModule( self.in_channels + self.channels , self.channels , kernel_size=lowerCamelCase , padding=kernel_size // 2 ) __snake_case : Optional[Any] = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) def __snake_case ( self : Dict ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : Tuple ) -> Optional[int]: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : Optional[int] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # just take the relevant feature maps __snake_case : List[str] = encoder_hidden_states[self.in_index] __snake_case : Optional[Any] = self.convs(lowerCamelCase ) if self.concat_input: __snake_case : Optional[Any] = self.conv_cat(torch.cat([hidden_states, output] , dim=1 ) ) __snake_case : Union[str, Any] = self.classifier(lowerCamelCase ) return output class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = UperNetConfig __UpperCAmelCase : int = "pixel_values" __UpperCAmelCase : str = True def __snake_case ( self : Union[str, Any] , lowerCamelCase : List[Any] ) -> List[Any]: if isinstance(lowerCamelCase , lowerCamelCase ): module.backbone.init_weights() module.decode_head.init_weights() module.auxiliary_head.init_weights() def __snake_case ( self : Optional[Any] ) -> List[str]: self.backbone.init_weights() self.decode_head.init_weights() self.auxiliary_head.init_weights() def __snake_case ( self : List[Any] , lowerCamelCase : Any , lowerCamelCase : Optional[Any]=False ) -> Dict: if isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : Union[str, Any] = value _snake_case : Dict = R"\n Parameters:\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n config ([`UperNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" _snake_case : Tuple = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using\n [`AutoImageProcessor`]. See [`SegformerImageProcessor.__call__`] for details.\n output_attentions (`bool`, *optional*):\n Whether or not to return the attentions tensors of all attention layers in case the backbone has them. See\n `attentions` under returned tensors for more detail.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers of the backbone. See `hidden_states` under\n returned tensors for more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n" @add_start_docstrings( "UperNet framework leveraging any vision backbone e.g. for ADE20k, CityScapes." , _lowerCAmelCase , ) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Dict , lowerCamelCase : int ) -> Optional[int]: super().__init__(lowerCamelCase ) __snake_case : Any = AutoBackbone.from_config(config.backbone_config ) # Semantic segmentation head(s) __snake_case : Union[str, Any] = UperNetHead(lowerCamelCase , in_channels=self.backbone.channels ) __snake_case : Any = UperNetFCNHead(lowerCamelCase ) if config.use_auxiliary_head else None # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UPERNET_INPUTS_DOCSTRING.format("batch_size, sequence_length" ) ) @replace_return_docstrings(output_type=lowerCamelCase , config_class=_CONFIG_FOR_DOC ) def __snake_case ( self : Union[str, Any] , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , ) -> Union[tuple, SemanticSegmenterOutput]: __snake_case : Any = return_dict if return_dict is not None else self.config.use_return_dict __snake_case : int = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __snake_case : str = output_attentions if output_attentions is not None else self.config.output_attentions __snake_case : Tuple = self.backbone.forward_with_filtered_kwargs( lowerCamelCase , output_hidden_states=lowerCamelCase , output_attentions=lowerCamelCase ) __snake_case : List[Any] = outputs.feature_maps __snake_case : List[Any] = self.decode_head(lowerCamelCase ) __snake_case : List[str] = nn.functional.interpolate(lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : Optional[int] = None if self.auxiliary_head is not None: __snake_case : Dict = self.auxiliary_head(lowerCamelCase ) __snake_case : Dict = nn.functional.interpolate( lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : int = None if labels is not None: if self.config.num_labels == 1: raise ValueError("The number of labels should be greater than one" ) else: # compute weighted loss __snake_case : Any = CrossEntropyLoss(ignore_index=self.config.loss_ignore_index ) __snake_case : Union[str, Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = main_loss + self.config.auxiliary_loss_weight * auxiliary_loss if not return_dict: if output_hidden_states: __snake_case : Any = (logits,) + outputs[1:] else: __snake_case : str = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return SemanticSegmenterOutput( loss=lowerCamelCase , logits=lowerCamelCase , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
81
0
"""simple docstring""" import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer __SCREAMING_SNAKE_CASE =logging.get_logger(__name__) class UpperCamelCase ( _lowerCAmelCase ): lowercase = "AutoTokenizer" lowercase = ["tokenizer"] lowercase = { "semantic_prompt": 1, "coarse_prompt": 2, "fine_prompt": 2, } def __init__( self ,__UpperCamelCase ,__UpperCamelCase=None ) -> Optional[Any]: '''simple docstring''' super().__init__(__UpperCamelCase ) lowercase_ : Tuple = speaker_embeddings @classmethod def _UpperCAmelCase ( cls ,__UpperCamelCase ,__UpperCamelCase="speaker_embeddings_path.json" ,**__UpperCamelCase ) -> List[str]: '''simple docstring''' if speaker_embeddings_dict_path is not None: lowercase_ : Any = get_file_from_repo( __UpperCamelCase ,__UpperCamelCase ,subfolder=kwargs.pop('subfolder' ,__UpperCamelCase ) ,cache_dir=kwargs.pop('cache_dir' ,__UpperCamelCase ) ,force_download=kwargs.pop('force_download' ,__UpperCamelCase ) ,proxies=kwargs.pop('proxies' ,__UpperCamelCase ) ,resume_download=kwargs.pop('resume_download' ,__UpperCamelCase ) ,local_files_only=kwargs.pop('local_files_only' ,__UpperCamelCase ) ,use_auth_token=kwargs.pop('use_auth_token' ,__UpperCamelCase ) ,revision=kwargs.pop('revision' ,__UpperCamelCase ) ,) if speaker_embeddings_path is None: logger.warning( f'''`{os.path.join(__UpperCamelCase ,__UpperCamelCase )}` does not exists\n , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json\n dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.''' ) lowercase_ : List[str] = None else: with open(__UpperCamelCase ) as speaker_embeddings_json: lowercase_ : List[Any] = json.load(__UpperCamelCase ) else: lowercase_ : Dict = None lowercase_ : Union[str, Any] = AutoTokenizer.from_pretrained(__UpperCamelCase ,**__UpperCamelCase ) return cls(tokenizer=__UpperCamelCase ,speaker_embeddings=__UpperCamelCase ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase="speaker_embeddings_path.json" ,__UpperCamelCase="speaker_embeddings" ,__UpperCamelCase = False ,**__UpperCamelCase ,) -> str: '''simple docstring''' if self.speaker_embeddings is not None: os.makedirs(os.path.join(__UpperCamelCase ,__UpperCamelCase ,'v2' ) ,exist_ok=__UpperCamelCase ) lowercase_ : List[Any] = {} lowercase_ : Optional[int] = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": lowercase_ : List[str] = self._load_voice_preset(__UpperCamelCase ) lowercase_ : int = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict['repo_or_path'] ,__UpperCamelCase ,f'''{prompt_key}_{key}''' ) ,voice_preset[key] ,allow_pickle=__UpperCamelCase ,) lowercase_ : Optional[Any] = os.path.join(__UpperCamelCase ,f'''{prompt_key}_{key}.npy''' ) lowercase_ : List[str] = tmp_dict with open(os.path.join(__UpperCamelCase ,__UpperCamelCase ) ,'w' ) as fp: json.dump(__UpperCamelCase ,__UpperCamelCase ) super().save_pretrained(__UpperCamelCase ,__UpperCamelCase ,**__UpperCamelCase ) def _UpperCAmelCase ( self ,__UpperCamelCase = None ,**__UpperCamelCase ) -> Union[str, Any]: '''simple docstring''' lowercase_ : Optional[int] = self.speaker_embeddings[voice_preset] lowercase_ : List[str] = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( f'''Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].''' ) lowercase_ : int = get_file_from_repo( self.speaker_embeddings.get('repo_or_path' ,'/' ) ,voice_preset_paths[key] ,subfolder=kwargs.pop('subfolder' ,__UpperCamelCase ) ,cache_dir=kwargs.pop('cache_dir' ,__UpperCamelCase ) ,force_download=kwargs.pop('force_download' ,__UpperCamelCase ) ,proxies=kwargs.pop('proxies' ,__UpperCamelCase ) ,resume_download=kwargs.pop('resume_download' ,__UpperCamelCase ) ,local_files_only=kwargs.pop('local_files_only' ,__UpperCamelCase ) ,use_auth_token=kwargs.pop('use_auth_token' ,__UpperCamelCase ) ,revision=kwargs.pop('revision' ,__UpperCamelCase ) ,) if path is None: raise ValueError( f'''`{os.path.join(self.speaker_embeddings.get("repo_or_path" ,"/" ) ,voice_preset_paths[key] )}` does not exists\n , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset}\n embeddings.''' ) lowercase_ : Tuple = np.load(__UpperCamelCase ) return voice_preset_dict def _UpperCAmelCase ( self ,__UpperCamelCase = None ) -> Tuple: '''simple docstring''' for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(f'''Voice preset unrecognized, missing {key} as a key.''' ) if not isinstance(voice_preset[key] ,np.ndarray ): raise ValueError(f'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(f'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) def __call__( self ,__UpperCamelCase=None ,__UpperCamelCase=None ,__UpperCamelCase="pt" ,__UpperCamelCase=256 ,__UpperCamelCase=False ,__UpperCamelCase=True ,__UpperCamelCase=False ,**__UpperCamelCase ,) -> List[Any]: '''simple docstring''' if voice_preset is not None and not isinstance(__UpperCamelCase ,__UpperCamelCase ): if ( isinstance(__UpperCamelCase ,__UpperCamelCase ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): lowercase_ : int = self._load_voice_preset(__UpperCamelCase ) else: if isinstance(__UpperCamelCase ,__UpperCamelCase ) and not voice_preset.endswith('.npz' ): lowercase_ : List[str] = voice_preset + ".npz" lowercase_ : Union[str, Any] = np.load(__UpperCamelCase ) if voice_preset is not None: self._validate_voice_preset_dict(__UpperCamelCase ,**__UpperCamelCase ) lowercase_ : str = BatchFeature(data=__UpperCamelCase ,tensor_type=__UpperCamelCase ) lowercase_ : Dict = self.tokenizer( __UpperCamelCase ,return_tensors=__UpperCamelCase ,padding='max_length' ,max_length=__UpperCamelCase ,return_attention_mask=__UpperCamelCase ,return_token_type_ids=__UpperCamelCase ,add_special_tokens=__UpperCamelCase ,**__UpperCamelCase ,) if voice_preset is not None: lowercase_ : Tuple = voice_preset return encoded_text
425
from __future__ import annotations def lowerCAmelCase_ ( __lowerCamelCase ): if not nums: return 0 __snake_case : Any = nums[0] __snake_case : str = 0 for num in nums[1:]: __snake_case , __snake_case : List[str] = ( max_excluding + num, max(__lowerCamelCase , __lowerCamelCase ), ) return max(__lowerCamelCase , __lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
81
0
'''simple docstring''' import inspect import unittest from transformers import ConvNextVaConfig from transformers.models.auto import get_values from transformers.models.auto.modeling_auto import MODEL_FOR_BACKBONE_MAPPING_NAMES, MODEL_MAPPING_NAMES from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextVaBackbone, ConvNextVaForImageClassification, ConvNextVaModel from transformers.models.convnextva.modeling_convnextva import CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowercase__ : '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=3_2 , lowerCamelCase__=3 , lowerCamelCase__=4 , lowerCamelCase__=[1_0, 2_0, 3_0, 4_0] , lowerCamelCase__=[2, 2, 3, 2] , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=3_7 , lowerCamelCase__="gelu" , lowerCamelCase__=1_0 , lowerCamelCase__=0.02 , lowerCamelCase__=["stage2", "stage3", "stage4"] , lowerCamelCase__=[2, 3, 4] , lowerCamelCase__=None , ): '''simple docstring''' UpperCamelCase = parent UpperCamelCase = batch_size UpperCamelCase = image_size UpperCamelCase = num_channels UpperCamelCase = num_stages UpperCamelCase = hidden_sizes UpperCamelCase = depths UpperCamelCase = is_training UpperCamelCase = use_labels UpperCamelCase = intermediate_size UpperCamelCase = hidden_act UpperCamelCase = num_labels UpperCamelCase = initializer_range UpperCamelCase = out_features UpperCamelCase = out_indices UpperCamelCase = scope def UpperCAmelCase ( self ): '''simple docstring''' UpperCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase = None if self.use_labels: UpperCamelCase = ids_tensor([self.batch_size] , self.num_labels ) UpperCamelCase = self.get_config() return config, pixel_values, labels def UpperCAmelCase ( self ): '''simple docstring''' return ConvNextVaConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def UpperCAmelCase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' UpperCamelCase = ConvNextVaModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() UpperCamelCase = model(lowerCamelCase__ ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 3_2, self.image_size // 3_2) , ) def UpperCAmelCase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' UpperCamelCase = ConvNextVaForImageClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() UpperCamelCase = model(lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' UpperCamelCase = ConvNextVaBackbone(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() UpperCamelCase = model(lowerCamelCase__ ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None UpperCamelCase = None UpperCamelCase = ConvNextVaBackbone(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() UpperCamelCase = model(lowerCamelCase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def UpperCAmelCase ( self ): '''simple docstring''' UpperCamelCase = self.prepare_config_and_inputs() UpperCamelCase = config_and_inputs UpperCamelCase = {"pixel_values": pixel_values} return config, inputs_dict def UpperCAmelCase ( self ): '''simple docstring''' UpperCamelCase = self.prepare_config_and_inputs() UpperCamelCase = config_and_inputs UpperCamelCase = {"pixel_values": pixel_values, "labels": labels} return config, inputs_dict @require_torch class lowercase__ ( _lowerCAmelCase, _lowerCAmelCase, unittest.TestCase ): '''simple docstring''' _snake_case = ( ( ConvNextVaModel, ConvNextVaForImageClassification, ConvNextVaBackbone, ) if is_torch_available() else () ) _snake_case = ( {"feature-extraction": ConvNextVaModel, "image-classification": ConvNextVaForImageClassification} if is_torch_available() else {} ) _snake_case = False _snake_case = False _snake_case = False _snake_case = False _snake_case = False def UpperCAmelCase ( self ): '''simple docstring''' UpperCamelCase = ConvNextVaModelTester(self ) UpperCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ , hidden_size=3_7 ) def UpperCAmelCase ( self ): '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCAmelCase ( self ): '''simple docstring''' return @unittest.skip(reason='''ConvNextV2 does not use inputs_embeds''' ) def UpperCAmelCase ( self ): '''simple docstring''' pass @unittest.skip(reason='''ConvNextV2 does not support input and output embeddings''' ) def UpperCAmelCase ( self ): '''simple docstring''' pass @unittest.skip(reason='''ConvNextV2 does not use feedforward chunking''' ) def UpperCAmelCase ( self ): '''simple docstring''' pass def UpperCAmelCase ( self ): '''simple docstring''' if not self.model_tester.is_training: return for model_class in self.all_model_classes: UpperCamelCase = self.model_tester.prepare_config_and_inputs_with_labels() UpperCamelCase = True if model_class.__name__ in [ *get_values(lowerCamelCase__ ), *get_values(lowerCamelCase__ ), ]: continue UpperCamelCase = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.train() UpperCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ , return_labels=lowerCamelCase__ ) UpperCamelCase = model(**lowerCamelCase__ ).loss loss.backward() def UpperCAmelCase ( self ): '''simple docstring''' if not self.model_tester.is_training: return for model_class in self.all_model_classes: UpperCamelCase = self.model_tester.prepare_config_and_inputs_with_labels() UpperCamelCase = False UpperCamelCase = True if ( model_class.__name__ in [*get_values(lowerCamelCase__ ), *get_values(lowerCamelCase__ )] or not model_class.supports_gradient_checkpointing ): continue UpperCamelCase = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.gradient_checkpointing_enable() model.train() UpperCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ , return_labels=lowerCamelCase__ ) UpperCamelCase = model(**lowerCamelCase__ ).loss loss.backward() def UpperCAmelCase ( self ): '''simple docstring''' UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase = model_class(lowerCamelCase__ ) UpperCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase = [*signature.parameters.keys()] UpperCamelCase = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def UpperCAmelCase ( self ): '''simple docstring''' UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def UpperCAmelCase ( self ): '''simple docstring''' def check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): UpperCamelCase = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): UpperCamelCase = model(**self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) UpperCamelCase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states UpperCamelCase = self.model_tester.num_stages self.assertEqual(len(lowerCamelCase__ ) , expected_num_stages + 1 ) # ConvNextV2's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase = True check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCamelCase = True check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def UpperCAmelCase ( self ): '''simple docstring''' UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase__ ) @slow def UpperCAmelCase ( self ): '''simple docstring''' for model_name in CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase = ConvNextVaModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def __snake_case ( ): UpperCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''') return image @require_torch @require_vision class lowercase__ ( unittest.TestCase ): '''simple docstring''' @cached_property def UpperCAmelCase ( self ): '''simple docstring''' return AutoImageProcessor.from_pretrained('''facebook/convnextv2-tiny-1k-224''' ) if is_vision_available() else None @slow def UpperCAmelCase ( self ): '''simple docstring''' UpperCamelCase = ConvNextVaForImageClassification.from_pretrained('''facebook/convnextv2-tiny-1k-224''' ).to(lowerCamelCase__ ) UpperCamelCase = self.default_image_processor UpperCamelCase = prepare_img() UpperCamelCase = preprocessor(images=lowerCamelCase__ , return_tensors='''pt''' ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): UpperCamelCase = model(**lowerCamelCase__ ) # verify the logits UpperCamelCase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) UpperCamelCase = torch.tensor([0.9996, 0.1966, -0.4386] ).to(lowerCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase__ , atol=1e-4 ) )
212
from __future__ import annotations from typing import Any def lowerCAmelCase_ ( __lowerCamelCase ): create_state_space_tree(__lowerCamelCase , [] , 0 ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): if index == len(__lowerCamelCase ): print(__lowerCamelCase ) return create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.pop() if __name__ == "__main__": _snake_case : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["A", "B", "C"]) generate_all_subsequences(seq)
81
0
def __snake_case ( _lowerCAmelCase : Dict , _lowerCAmelCase : Any ) -> int: return int((input_a, input_a).count(0 ) == 0 ) def __snake_case ( ) -> List[str]: assert and_gate(0 , 0 ) == 0 assert and_gate(0 , 1 ) == 0 assert and_gate(1 , 0 ) == 0 assert and_gate(1 , 1 ) == 1 if __name__ == "__main__": test_and_gate() print(and_gate(1, 0)) print(and_gate(0, 0)) print(and_gate(0, 1)) print(and_gate(1, 1))
454
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case , __snake_case : List[Any] = image.size __snake_case , __snake_case : Tuple = (x - x % 3_2 for x in (w, h)) # resize to integer multiple of 32 __snake_case : str = image.resize((w, h) , resample=PIL_INTERPOLATION["lanczos"] ) __snake_case : int = np.array(__lowerCamelCase ).astype(np.floataa ) / 2_5_5.0 __snake_case : Union[str, Any] = image[None].transpose(0 , 3 , 1 , 2 ) __snake_case : Union[str, Any] = torch.from_numpy(__lowerCamelCase ) return 2.0 * image - 1.0 class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : VQModel , lowerCamelCase : UNetaDModel , lowerCamelCase : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ) -> Union[str, Any]: super().__init__() self.register_modules(vqvae=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase ) @torch.no_grad() def __call__( self : List[str] , lowerCamelCase : Union[torch.Tensor, PIL.Image.Image] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : Optional[int] = 100 , lowerCamelCase : Optional[float] = 0.0 , lowerCamelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , ) -> Union[Tuple, ImagePipelineOutput]: if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : Any = 1 elif isinstance(lowerCamelCase , torch.Tensor ): __snake_case : Any = image.shape[0] else: raise ValueError(F'`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(lowerCamelCase )}' ) if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : List[Any] = preprocess(lowerCamelCase ) __snake_case , __snake_case : int = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image __snake_case : str = (batch_size, self.unet.config.in_channels // 2, height, width) __snake_case : str = next(self.unet.parameters() ).dtype __snake_case : Tuple = randn_tensor(lowerCamelCase , generator=lowerCamelCase , device=self.device , dtype=lowerCamelCase ) __snake_case : List[Any] = image.to(device=self.device , dtype=lowerCamelCase ) # set timesteps and move to the correct device self.scheduler.set_timesteps(lowerCamelCase , device=self.device ) __snake_case : str = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler __snake_case : Dict = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] __snake_case : Union[str, Any] = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __snake_case : int = {} if accepts_eta: __snake_case : List[str] = eta for t in self.progress_bar(lowerCamelCase ): # concat latents and low resolution image in the channel dimension. __snake_case : Union[str, Any] = torch.cat([latents, image] , dim=1 ) __snake_case : Optional[Any] = self.scheduler.scale_model_input(lowerCamelCase , lowerCamelCase ) # predict the noise residual __snake_case : int = self.unet(lowerCamelCase , lowerCamelCase ).sample # compute the previous noisy sample x_t -> x_t-1 __snake_case : Union[str, Any] = self.scheduler.step(lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase ).prev_sample # decode the image latents with the VQVAE __snake_case : List[Any] = self.vqvae.decode(lowerCamelCase ).sample __snake_case : Dict = torch.clamp(lowerCamelCase , -1.0 , 1.0 ) __snake_case : Any = image / 2 + 0.5 __snake_case : Tuple = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __snake_case : Tuple = self.numpy_to_pil(lowerCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCamelCase )
81
0
"""simple docstring""" from ....configuration_utils import PretrainedConfig from ....utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { "Visual-Attention-Network/van-base": ( "https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json" ), } class _UpperCamelCase ( _lowerCAmelCase): __lowerCamelCase = "van" def __init__(self , lowerCamelCase__=2_2_4 , lowerCamelCase__=3 , lowerCamelCase__=[7, 3, 3, 3] , lowerCamelCase__=[4, 2, 2, 2] , lowerCamelCase__=[6_4, 1_2_8, 3_2_0, 5_1_2] , lowerCamelCase__=[3, 3, 1_2, 3] , lowerCamelCase__=[8, 8, 4, 4] , lowerCamelCase__="gelu" , lowerCamelCase__=0.0_2 , lowerCamelCase__=1E-6 , lowerCamelCase__=1E-2 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , **lowerCamelCase__ , ): """simple docstring""" super().__init__(**lowerCamelCase__ ) A__ = image_size A__ = num_channels A__ = patch_sizes A__ = strides A__ = hidden_sizes A__ = depths A__ = mlp_ratios A__ = hidden_act A__ = initializer_range A__ = layer_norm_eps A__ = layer_scale_init_value A__ = drop_path_rate A__ = dropout_rate
574
import gc import unittest import torch from parameterized import parameterized from diffusers import AutoencoderKL from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = AutoencoderKL __UpperCAmelCase : Optional[Any] = "sample" __UpperCAmelCase : Optional[int] = 1e-2 @property def __snake_case ( self : Dict ) -> Optional[Any]: __snake_case : Optional[Any] = 4 __snake_case : Tuple = 3 __snake_case : List[str] = (32, 32) __snake_case : str = floats_tensor((batch_size, num_channels) + sizes ).to(lowerCamelCase ) return {"sample": image} @property def __snake_case ( self : Union[str, Any] ) -> Tuple: return (3, 32, 32) @property def __snake_case ( self : int ) -> int: return (3, 32, 32) def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Optional[Any] = { "block_out_channels": [32, 64], "in_channels": 3, "out_channels": 3, "down_block_types": ["DownEncoderBlock2D", "DownEncoderBlock2D"], "up_block_types": ["UpDecoderBlock2D", "UpDecoderBlock2D"], "latent_channels": 4, } __snake_case : Any = self.dummy_input return init_dict, inputs_dict def __snake_case ( self : str ) -> Dict: pass def __snake_case ( self : Tuple ) -> List[str]: pass @unittest.skipIf(torch_device == "mps" , "Gradient checkpointing skipped on MPS" ) def __snake_case ( self : Any ) -> Optional[Any]: # enable deterministic behavior for gradient checkpointing __snake_case , __snake_case : int = self.prepare_init_args_and_inputs_for_common() __snake_case : str = self.model_class(**lowerCamelCase ) model.to(lowerCamelCase ) assert not model.is_gradient_checkpointing and model.training __snake_case : str = model(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model.zero_grad() __snake_case : Any = torch.randn_like(lowerCamelCase ) __snake_case : str = (out - labels).mean() loss.backward() # re-instantiate the model now enabling gradient checkpointing __snake_case : Optional[int] = self.model_class(**lowerCamelCase ) # clone model model_a.load_state_dict(model.state_dict() ) model_a.to(lowerCamelCase ) model_a.enable_gradient_checkpointing() assert model_a.is_gradient_checkpointing and model_a.training __snake_case : int = model_a(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model_a.zero_grad() __snake_case : Union[str, Any] = (out_a - labels).mean() loss_a.backward() # compare the output and parameters gradients self.assertTrue((loss - loss_a).abs() < 1E-5 ) __snake_case : Optional[int] = dict(model.named_parameters() ) __snake_case : List[Any] = dict(model_a.named_parameters() ) for name, param in named_params.items(): self.assertTrue(torch_all_close(param.grad.data , named_params_a[name].grad.data , atol=5E-5 ) ) def __snake_case ( self : List[Any] ) -> Optional[int]: __snake_case , __snake_case : Optional[Any] = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" , output_loading_info=lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(lowerCamelCase ) __snake_case : Optional[Any] = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def __snake_case ( self : Optional[Any] ) -> Union[str, Any]: __snake_case : Tuple = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" ) __snake_case : Dict = model.to(lowerCamelCase ) model.eval() if torch_device == "mps": __snake_case : int = torch.manual_seed(0 ) else: __snake_case : str = torch.Generator(device=lowerCamelCase ).manual_seed(0 ) __snake_case : List[str] = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) __snake_case : Union[str, Any] = image.to(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , sample_posterior=lowerCamelCase , generator=lowerCamelCase ).sample __snake_case : List[Any] = output[0, -1, -3:, -3:].flatten().cpu() # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. if torch_device == "mps": __snake_case : Union[str, Any] = torch.tensor( [ -4.0078E-01, -3.8323E-04, -1.2681E-01, -1.1462E-01, 2.0095E-01, 1.0893E-01, -8.8247E-02, -3.0361E-01, -9.8644E-03, ] ) elif torch_device == "cpu": __snake_case : Tuple = torch.tensor( [-0.13_52, 0.08_78, 0.04_19, -0.08_18, -0.10_69, 0.06_88, -0.14_58, -0.44_46, -0.00_26] ) else: __snake_case : List[str] = torch.tensor( [-0.24_21, 0.46_42, 0.25_07, -0.04_38, 0.06_82, 0.31_60, -0.20_18, -0.07_27, 0.24_85] ) self.assertTrue(torch_all_close(lowerCamelCase , lowerCamelCase , rtol=1E-2 ) ) @slow class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : int , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] ) -> List[str]: return F'gaussian_noise_s={seed}_shape={"_".join([str(lowerCamelCase ) for s in shape] )}.npy' def __snake_case ( self : List[Any] ) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __snake_case ( self : Tuple , lowerCamelCase : List[Any]=0 , lowerCamelCase : Tuple=(4, 3, 512, 512) , lowerCamelCase : Optional[int]=False ) -> str: __snake_case : List[Any] = torch.floataa if fpaa else torch.floataa __snake_case : Tuple = torch.from_numpy(load_hf_numpy(self.get_file_format(lowerCamelCase , lowerCamelCase ) ) ).to(lowerCamelCase ).to(lowerCamelCase ) return image def __snake_case ( self : Optional[Any] , lowerCamelCase : int="CompVis/stable-diffusion-v1-4" , lowerCamelCase : int=False ) -> int: __snake_case : str = "fp16" if fpaa else None __snake_case : int = torch.floataa if fpaa else torch.floataa __snake_case : int = AutoencoderKL.from_pretrained( lowerCamelCase , subfolder="vae" , torch_dtype=lowerCamelCase , revision=lowerCamelCase , ) model.to(lowerCamelCase ).eval() return model def __snake_case ( self : str , lowerCamelCase : int=0 ) -> Optional[Any]: if torch_device == "mps": return torch.manual_seed(lowerCamelCase ) return torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) @parameterized.expand( [ # fmt: off [33, [-0.16_03, 0.98_78, -0.04_95, -0.07_90, -0.27_09, 0.83_75, -0.20_60, -0.08_24], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_76, 0.11_68, 0.13_32, -0.48_40, -0.25_08, -0.07_91, -0.04_93, -0.40_89], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[str] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] ) -> List[Any]: __snake_case : Optional[Any] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) __snake_case : Tuple = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : List[Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : int = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [33, [-0.05_13, 0.02_89, 1.37_99, 0.21_66, -0.25_73, -0.08_71, 0.51_03, -0.09_99]], [47, [-0.41_28, -0.13_20, -0.37_04, 0.19_65, -0.41_16, -0.23_32, -0.33_40, 0.22_47]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : Any , lowerCamelCase : List[str] , lowerCamelCase : List[str] ) -> Tuple: __snake_case : Any = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Any = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.16_09, 0.98_66, -0.04_87, -0.07_77, -0.27_16, 0.83_68, -0.20_55, -0.08_14], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_77, 0.11_47, 0.13_33, -0.48_41, -0.25_06, -0.08_05, -0.04_91, -0.40_85], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Any , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) with torch.no_grad(): __snake_case : int = model(lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Union[str, Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : List[str] = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [13, [-0.20_51, -0.18_03, -0.23_11, -0.21_14, -0.32_92, -0.35_74, -0.29_53, -0.33_23]], [37, [-0.26_32, -0.26_25, -0.21_99, -0.27_41, -0.45_39, -0.49_90, -0.37_20, -0.49_25]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : List[str] , lowerCamelCase : Tuple , lowerCamelCase : Any ) -> Optional[Any]: __snake_case : List[str] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : str = sample[-1, -2:, :2, -2:].flatten().cpu() __snake_case : Optional[int] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-3 ) @parameterized.expand( [ # fmt: off [27, [-0.03_69, 0.02_07, -0.07_76, -0.06_82, -0.17_47, -0.19_30, -0.14_65, -0.20_39]], [16, [-0.16_28, -0.21_34, -0.27_47, -0.26_42, -0.37_74, -0.44_04, -0.36_87, -0.42_77]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : str , lowerCamelCase : Optional[int] , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[str] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : Union[str, Any] = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Optional[Any] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=5E-3 ) @parameterized.expand([(13,), (16,), (27,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> Tuple: __snake_case : Dict = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : Any = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Any = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-1 ) @parameterized.expand([(13,), (16,), (37,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : List[Any] , lowerCamelCase : Any ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : Union[str, Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : List[Any] = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Dict = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.30_01, 0.09_18, -2.69_84, -3.97_20, -3.20_99, -5.03_53, 1.73_38, -0.20_65, 3.42_67]], [47, [-1.50_30, -4.38_71, -6.03_55, -9.11_57, -1.66_61, -2.78_53, 2.16_07, -5.08_23, 2.56_33]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Dict ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : int = self.get_sd_image(lowerCamelCase ) __snake_case : int = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model.encode(lowerCamelCase ).latent_dist __snake_case : Dict = dist.sample(generator=lowerCamelCase ) assert list(sample.shape ) == [image.shape[0], 4] + [i // 8 for i in image.shape[2:]] __snake_case : List[str] = sample[0, -1, -3:, -3:].flatten().cpu() __snake_case : Dict = torch.tensor(lowerCamelCase ) __snake_case : Dict = 3E-3 if torch_device != "mps" else 1E-2 assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=lowerCamelCase )
81
0
'''simple docstring''' def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase ): print('''\nThe shortest path matrix using Floyd Warshall algorithm\n''' ) for i in range(__lowerCamelCase ): for j in range(__lowerCamelCase ): if dist[i][j] != float('''inf''' ): print(int(dist[i][j] ) , end='''\t''' ) else: print('''INF''' , end='''\t''' ) print() def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase ): lowercase__ : int = [[float('''inf''' ) for _ in range(__lowerCamelCase )] for _ in range(__lowerCamelCase )] for i in range(__lowerCamelCase ): for j in range(__lowerCamelCase ): lowercase__ : Optional[Any] = graph[i][j] # check vertex k against all other vertices (i, j) for k in range(__lowerCamelCase ): # looping through rows of graph array for i in range(__lowerCamelCase ): # looping through columns of graph array for j in range(__lowerCamelCase ): if ( dist[i][k] != float('''inf''' ) and dist[k][j] != float('''inf''' ) and dist[i][k] + dist[k][j] < dist[i][j] ): lowercase__ : List[Any] = dist[i][k] + dist[k][j] _print_dist(__lowerCamelCase , __lowerCamelCase ) return dist, v if __name__ == "__main__": __a: int = int(input("""Enter number of vertices: """)) __a: List[str] = int(input("""Enter number of edges: """)) __a: str = [[float("""inf""") for i in range(v)] for j in range(v)] for i in range(v): __a: Dict = 0.0 # src and dst are indices that must be within the array size graph[e][v] # failure to follow this will result in an error for i in range(e): print("""\nEdge """, i + 1) __a: str = int(input("""Enter source:""")) __a: Any = int(input("""Enter destination:""")) __a: Tuple = float(input("""Enter weight:""")) __a: Union[str, Any] = weight floyd_warshall(graph, v) # Example Input # Enter number of vertices: 3 # Enter number of edges: 2 # # generated graph from vertex and edge inputs # [[inf, inf, inf], [inf, inf, inf], [inf, inf, inf]] # [[0.0, inf, inf], [inf, 0.0, inf], [inf, inf, 0.0]] # specify source, destination and weight for edge #1 # Edge 1 # Enter source:1 # Enter destination:2 # Enter weight:2 # specify source, destination and weight for edge #2 # Edge 2 # Enter source:2 # Enter destination:1 # Enter weight:1 # # Expected Output from the vertice, edge and src, dst, weight inputs!! # 0 INF INF # INF 0 2 # INF 1 0
152
import warnings from ...utils import logging from .image_processing_flava import FlavaImageProcessor _snake_case : Optional[int] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[Any] , *lowerCamelCase : Any , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class FlavaFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use FlavaImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
import argparse from torch import nn # transformers_old should correspond to branch `save_old_prophetnet_model_structure` here # original prophetnet_checkpoints are saved under `patrickvonplaten/..._old` respectively from transformers_old.modeling_prophetnet import ( ProphetNetForConditionalGeneration as ProphetNetForConditionalGenerationOld, ) from transformers_old.modeling_xlm_prophetnet import ( XLMProphetNetForConditionalGeneration as XLMProphetNetForConditionalGenerationOld, ) from transformers import ProphetNetForConditionalGeneration, XLMProphetNetForConditionalGeneration, logging lowerCAmelCase_ = logging.get_logger(__name__) logging.set_verbosity_info() def A_ ( lowercase_ , lowercase_ ) -> List[Any]: if "xprophetnet" in prophetnet_checkpoint_path: _snake_case : Any = XLMProphetNetForConditionalGenerationOld.from_pretrained(__lowerCamelCase ) _snake_case : Optional[Any] = XLMProphetNetForConditionalGeneration.from_pretrained( __lowerCamelCase , output_loading_info=__lowerCamelCase ) else: _snake_case : Optional[int] = ProphetNetForConditionalGenerationOld.from_pretrained(__lowerCamelCase ) _snake_case : Dict = ProphetNetForConditionalGeneration.from_pretrained( __lowerCamelCase , output_loading_info=__lowerCamelCase ) _snake_case : Union[str, Any] = ["key_proj", "value_proj", "query_proj"] _snake_case : Tuple = { "self_attn": "ngram_self_attn", "cross_attn": "encoder_attn", "cross_attn_layer_norm": "encoder_attn_layer_norm", "feed_forward_layer_norm": "final_layer_norm", "feed_forward": "", "intermediate": "fc1", "output": "fc2", "key_proj": "k_proj", "query_proj": "q_proj", "value_proj": "v_proj", "word_embeddings": "embed_tokens", "embeddings_layer_norm": "emb_layer_norm", "relative_pos_embeddings": "relative_linear", "ngram_embeddings": "ngram_input_embed", "position_embeddings": "embed_positions", } for key in loading_info["missing_keys"]: _snake_case : Optional[Any] = key.split('''.''' ) if attributes[0] == "lm_head": _snake_case : Dict = prophet _snake_case : Tuple = prophet_old else: _snake_case : str = prophet.prophetnet _snake_case : Any = prophet_old.model _snake_case : Any = False for attribute in attributes: if attribute in mapping: _snake_case : Optional[int] = mapping[attribute] if not hasattr(__lowerCamelCase , __lowerCamelCase ) and len(__lowerCamelCase ) > 0: _snake_case : List[str] = attribute elif hasattr(__lowerCamelCase , __lowerCamelCase ): _snake_case : List[str] = attribute if attribute == "weight": assert old_model.weight.shape == model.weight.shape, "Shapes have to match!" _snake_case : Dict = old_model.weight logger.info(f'''{attribute} is initialized.''' ) _snake_case : int = True break elif attribute == "bias": assert old_model.bias.shape == model.bias.shape, "Shapes have to match!" _snake_case : int = old_model.bias logger.info(f'''{attribute} is initialized''' ) _snake_case : Dict = True break elif attribute in special_keys and hasattr(__lowerCamelCase , '''in_proj_weight''' ): _snake_case : Any = old_model.in_proj_weight.shape[0] // 3 _snake_case : Dict = getattr(__lowerCamelCase , __lowerCamelCase ) param.weight.shape == old_model.in_proj_weight[:embed_dim, :].shape, "Shapes have to match" param.bias.shape == old_model.in_proj_bias[:embed_dim].shape, "Shapes have to match" if attribute == "query_proj": _snake_case : Any = nn.Parameter(old_model.in_proj_weight[:embed_dim, :] ) _snake_case : Tuple = nn.Parameter(old_model.in_proj_bias[:embed_dim] ) elif attribute == "key_proj": _snake_case : str = nn.Parameter(old_model.in_proj_weight[embed_dim : 2 * embed_dim, :] ) _snake_case : List[Any] = nn.Parameter(old_model.in_proj_bias[embed_dim : 2 * embed_dim] ) elif attribute == "value_proj": _snake_case : List[Any] = nn.Parameter(old_model.in_proj_weight[2 * embed_dim :, :] ) _snake_case : Optional[Any] = nn.Parameter(old_model.in_proj_bias[2 * embed_dim :] ) _snake_case : int = True break elif attribute == "position_embeddings": assert ( model.position_embeddings.weight.shape[-1] == old_model.embed_positions.weight.shape[-1] ), "Hidden size has to match" assert model.position_embeddings.weight.shape[0] == 512, "We want 512 position_embeddings." _snake_case : Tuple = nn.Parameter(old_model.embed_positions.weight[:512, :] ) _snake_case : List[str] = True break if attribute.isdigit(): _snake_case : List[str] = model[int(__lowerCamelCase )] _snake_case : int = old_model[int(__lowerCamelCase )] else: _snake_case : Optional[int] = getattr(__lowerCamelCase , __lowerCamelCase ) if old_attribute == "": _snake_case : Optional[Any] = old_model else: if not hasattr(__lowerCamelCase , __lowerCamelCase ): raise ValueError(f'''{old_model} does not have {old_attribute}''' ) _snake_case : str = getattr(__lowerCamelCase , __lowerCamelCase ) if not is_key_init: raise ValueError(f'''{key} was not correctly initialized!''' ) print(f'''Saving model to {pytorch_dump_folder_path}''' ) prophet.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": lowerCAmelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( "--prophetnet_checkpoint_path", default=None, type=str, required=True, help="Path the official PyTorch dump." ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) lowerCAmelCase_ = parser.parse_args() convert_prophetnet_checkpoint_to_pytorch(args.prophetnet_checkpoint_path, args.pytorch_dump_folder_path)
326
import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : Tuple ) -> Optional[Any]: __snake_case : Dict = tempfile.mkdtemp() __snake_case : Any = SamImageProcessor() __snake_case : Optional[int] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : Optional[Any] , **lowerCamelCase : Optional[int] ) -> Optional[Any]: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[Any] ) -> Dict: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : int ) -> List[Any]: __snake_case : List[str] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : int = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : List[Any] ) -> Dict: __snake_case : int = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Dict = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Optional[Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : List[str] ) -> Tuple: __snake_case : int = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Optional[int] = self.prepare_image_inputs() __snake_case : List[str] = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : Dict = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_torch def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Tuple = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[str] = [torch.ones((1, 3, 5, 5) )] __snake_case : Tuple = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : int = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , torch.tensor(lowerCamelCase ) , torch.tensor(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : List[str] = [np.ones((1, 3, 5, 5) )] __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : str = [[1, 0], [0, 1]] with self.assertRaises(lowerCamelCase ): __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) @require_vision @require_tf class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[Any] ) -> Union[str, Any]: __snake_case : int = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : str , **lowerCamelCase : Any ) -> Tuple: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> Any: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : str ) -> List[Any]: __snake_case : int = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : Dict = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : int ) -> List[str]: __snake_case : List[Any] = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Optional[Any] = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Tuple = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : Union[str, Any] ) -> List[Any]: __snake_case : str = self.get_image_processor() __snake_case : Union[str, Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : int = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : List[str] = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_tf def __snake_case ( self : Any ) -> Optional[int]: __snake_case : List[str] = self.get_image_processor() __snake_case : Dict = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Union[str, Any] = [tf.ones((1, 3, 5, 5) )] __snake_case : List[Any] = [[1764, 2646]] __snake_case : Dict = [[683, 1024]] __snake_case : List[str] = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , tf.convert_to_tensor(lowerCamelCase ) , tf.convert_to_tensor(lowerCamelCase ) , return_tensors="tf" , ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : Union[str, Any] = [np.ones((1, 3, 5, 5) )] __snake_case : List[str] = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Tuple = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): __snake_case : Dict = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[str] ) -> str: __snake_case : Optional[int] = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : List[str] , **lowerCamelCase : Any ) -> Dict: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> List[Any]: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : str = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : List[Any] = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : str = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[Any] = np.random.randint(0 , 2 , size=(1, 3, 5, 5) ).astype(np.floataa ) __snake_case : Dict = [tf.convert_to_tensor(lowerCamelCase )] __snake_case : List[Any] = [torch.tensor(lowerCamelCase )] __snake_case : Optional[Any] = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : Union[str, Any] = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) __snake_case : Dict = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="pt" ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def __snake_case ( self : List[Any] ) -> List[str]: __snake_case : Any = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : Any = image_processor(lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Optional[Any] = processor(images=lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Tuple = image_processor(lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() __snake_case : List[Any] = processor(images=lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) )
81
0
"""simple docstring""" import numpy as np from cva import destroyAllWindows, imread, imshow, waitKey class SCREAMING_SNAKE_CASE_ : """simple docstring""" def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): if dst_width < 0 or dst_height < 0: raise ValueError("""Destination width/height should be > 0""") __SCREAMING_SNAKE_CASE = img __SCREAMING_SNAKE_CASE = img.shape[1] __SCREAMING_SNAKE_CASE = img.shape[0] __SCREAMING_SNAKE_CASE = dst_width __SCREAMING_SNAKE_CASE = dst_height __SCREAMING_SNAKE_CASE = self.src_w / self.dst_w __SCREAMING_SNAKE_CASE = self.src_h / self.dst_h __SCREAMING_SNAKE_CASE = ( np.ones((self.dst_h, self.dst_w, 3) , np.uinta) * 2_5_5 ) def snake_case_ ( self): for i in range(self.dst_h): for j in range(self.dst_w): __SCREAMING_SNAKE_CASE = self.img[self.get_y(lowerCAmelCase__)][self.get_x(lowerCAmelCase__)] def snake_case_ ( self , lowerCAmelCase__): return int(self.ratio_x * x) def snake_case_ ( self , lowerCAmelCase__): return int(self.ratio_y * y) if __name__ == "__main__": __magic_name__ = 800, 600 __magic_name__ = imread("image_data/lena.jpg", 1) __magic_name__ = NearestNeighbour(im, dst_w, dst_h) n.process() imshow( F"""Image resized from: {im.shape[1]}x{im.shape[0]} to {dst_w}x{dst_h}""", n.output ) waitKey(0) destroyAllWindows()
155
import warnings from ...utils import logging from .image_processing_deit import DeiTImageProcessor _snake_case : List[str] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[str] , *lowerCamelCase : str , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class DeiTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use DeiTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class lowerCAmelCase_ ( unittest.TestCase ): @slow def __snake_case ( self : Optional[Any] ): lowerCAmelCase__ = FlaxMTaForConditionalGeneration.from_pretrained('''google/mt5-small''' ) lowerCAmelCase__ = AutoTokenizer.from_pretrained('''google/mt5-small''' ) lowerCAmelCase__ = tokenizer('''Hello there''' , return_tensors='''np''' ).input_ids lowerCAmelCase__ = tokenizer('''Hi I am''' , return_tensors='''np''' ).input_ids lowerCAmelCase__ = shift_tokens_right(SCREAMING_SNAKE_CASE_ , model.config.pad_token_id , model.config.decoder_start_token_id ) lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE_ , decoder_input_ids=SCREAMING_SNAKE_CASE_ ).logits lowerCAmelCase__ = optax.softmax_cross_entropy(SCREAMING_SNAKE_CASE_ , onehot(SCREAMING_SNAKE_CASE_ , logits.shape[-1] ) ).mean() lowerCAmelCase__ = -(labels.shape[-1] * loss.item()) lowerCAmelCase__ = -84.9_127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1e-4 )
668
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case : Union[str, Any] = { "configuration_owlvit": [ "OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "OwlViTConfig", "OwlViTOnnxConfig", "OwlViTTextConfig", "OwlViTVisionConfig", ], "processing_owlvit": ["OwlViTProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[Any] = ["OwlViTFeatureExtractor"] _snake_case : Optional[int] = ["OwlViTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "OwlViTModel", "OwlViTPreTrainedModel", "OwlViTTextModel", "OwlViTVisionModel", "OwlViTForObjectDetection", ] if TYPE_CHECKING: from .configuration_owlvit import ( OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, OwlViTConfig, OwlViTOnnxConfig, OwlViTTextConfig, OwlViTVisionConfig, ) from .processing_owlvit import OwlViTProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_owlvit import OwlViTFeatureExtractor from .image_processing_owlvit import OwlViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_owlvit import ( OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST, OwlViTForObjectDetection, OwlViTModel, OwlViTPreTrainedModel, OwlViTTextModel, OwlViTVisionModel, ) else: import sys _snake_case : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
0
'''simple docstring''' from io import BytesIO from typing import List, Union import requests from ..utils import add_end_docstrings, is_decord_available, is_torch_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_decord_available(): import numpy as np from decord import VideoReader if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING __UpperCamelCase = logging.get_logger(__name__) @add_end_docstrings(_lowerCAmelCase ) class _A ( _lowerCAmelCase ): def __init__( self : int , *__magic_name__ : List[Any] , **__magic_name__ : List[Any] ) -> Optional[int]: """simple docstring""" super().__init__(*__magic_name__ , **__magic_name__ ) requires_backends(self , """decord""" ) self.check_model_type(__magic_name__ ) def lowercase__ ( self : str , __magic_name__ : str=None , __magic_name__ : Any=None , __magic_name__ : Any=None ) -> Optional[int]: """simple docstring""" __snake_case : Any = {} if frame_sampling_rate is not None: __snake_case : Union[str, Any] = frame_sampling_rate if num_frames is not None: __snake_case : Optional[Any] = num_frames __snake_case : int = {} if top_k is not None: __snake_case : Union[str, Any] = top_k return preprocess_params, {}, postprocess_params def __call__( self : Tuple , __magic_name__ : Union[str, List[str]] , **__magic_name__ : Optional[int] ) -> Tuple: """simple docstring""" return super().__call__(__magic_name__ , **__magic_name__ ) def lowercase__ ( self : int , __magic_name__ : Any , __magic_name__ : Optional[Any]=None , __magic_name__ : Dict=1 ) -> str: """simple docstring""" if num_frames is None: __snake_case : List[str] = self.model.config.num_frames if video.startswith("""http://""" ) or video.startswith("""https://""" ): __snake_case : Optional[Any] = BytesIO(requests.get(__magic_name__ ).content ) __snake_case : Optional[int] = VideoReader(__magic_name__ ) videoreader.seek(0 ) __snake_case : Optional[Any] = 0 __snake_case : Dict = num_frames * frame_sampling_rate - 1 __snake_case : Union[str, Any] = np.linspace(__magic_name__ , __magic_name__ , num=__magic_name__ , dtype=np.intaa ) __snake_case : Optional[Any] = videoreader.get_batch(__magic_name__ ).asnumpy() __snake_case : Any = list(__magic_name__ ) __snake_case : List[Any] = self.image_processor(__magic_name__ , return_tensors=self.framework ) return model_inputs def lowercase__ ( self : Dict , __magic_name__ : Tuple ) -> Tuple: """simple docstring""" __snake_case : List[str] = self.model(**__magic_name__ ) return model_outputs def lowercase__ ( self : int , __magic_name__ : Any , __magic_name__ : int=5 ) -> List[Any]: """simple docstring""" if top_k > self.model.config.num_labels: __snake_case : str = self.model.config.num_labels if self.framework == "pt": __snake_case : List[str] = model_outputs.logits.softmax(-1 )[0] __snake_case : Union[str, Any] = probs.topk(__magic_name__ ) else: raise ValueError(f'''Unsupported framework: {self.framework}''' ) __snake_case : Any = scores.tolist() __snake_case : Any = ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(__magic_name__ , __magic_name__ )]
26
from ....configuration_utils import PretrainedConfig from ....utils import logging _snake_case : int = logging.get_logger(__name__) _snake_case : str = { "Visual-Attention-Network/van-base": ( "https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json" ), } class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[Any] = "van" def __init__( self : Optional[int] , lowerCamelCase : Any=224 , lowerCamelCase : str=3 , lowerCamelCase : Any=[7, 3, 3, 3] , lowerCamelCase : Dict=[4, 2, 2, 2] , lowerCamelCase : List[Any]=[64, 128, 320, 512] , lowerCamelCase : str=[3, 3, 12, 3] , lowerCamelCase : Dict=[8, 8, 4, 4] , lowerCamelCase : Any="gelu" , lowerCamelCase : Optional[int]=0.02 , lowerCamelCase : Tuple=1E-6 , lowerCamelCase : Optional[int]=1E-2 , lowerCamelCase : int=0.0 , lowerCamelCase : Optional[Any]=0.0 , **lowerCamelCase : Optional[int] , ) -> int: super().__init__(**lowerCamelCase ) __snake_case : Union[str, Any] = image_size __snake_case : Any = num_channels __snake_case : Any = patch_sizes __snake_case : List[Any] = strides __snake_case : str = hidden_sizes __snake_case : Dict = depths __snake_case : Optional[int] = mlp_ratios __snake_case : Dict = hidden_act __snake_case : Union[str, Any] = initializer_range __snake_case : List[str] = layer_norm_eps __snake_case : Optional[int] = layer_scale_init_value __snake_case : List[Any] = drop_path_rate __snake_case : int = dropout_rate
81
0
'''simple docstring''' def __UpperCamelCase ( lowercase__ : List[Any], lowercase__ : Optional[Any] ): '''simple docstring''' __lowercase =len(__lowerCamelCase ) __lowercase =len(__lowerCamelCase ) __lowercase =[[False for _ in range(m + 1 )] for _ in range(n + 1 )] __lowercase =True for i in range(__lowerCamelCase ): for j in range(m + 1 ): if dp[i][j]: if j < m and a[i].upper() == b[j]: __lowercase =True if a[i].islower(): __lowercase =True return dp[n][m] if __name__ == "__main__": import doctest doctest.testmod()
119
import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer _snake_case : Union[str, Any] = logging.getLogger(__name__) def lowerCAmelCase_ ( ): __snake_case : int = argparse.ArgumentParser( description="Prepare TFRecord shards from pre-tokenized samples of the wikitext dataset." ) parser.add_argument( "--dataset_name" , type=__lowerCamelCase , default="wikitext" , help="Name of the training. Explore datasets at: hf.co/datasets." , ) parser.add_argument( "--dataset_config" , type=__lowerCamelCase , default="wikitext-103-raw-v1" , help="Configuration name of the dataset." ) parser.add_argument( "--tokenizer_name_or_path" , type=__lowerCamelCase , default="sayakpaul/unigram-tokenizer-wikitext" , help="Tokenizer identifier. Can be a local filepath or a Hub identifier." , ) parser.add_argument( "--shard_size" , type=__lowerCamelCase , default=1_0_0_0 , help="Number of entries to go in a single shard." , ) parser.add_argument("--split" , type=__lowerCamelCase , default="train" , choices=["train", "test", "validation"] ) parser.add_argument( "--limit" , default=__lowerCamelCase , type=__lowerCamelCase , help="Limit the number of shards (used for debugging)." , ) parser.add_argument( "--max_length" , type=__lowerCamelCase , default=5_1_2 , help="Maximum sequence length. For training on TPUs, it helps to have a maximum" " sequence length that is a multiple of 8." , ) parser.add_argument( "--output_dir" , default="tf-tpu" , type=__lowerCamelCase , help="Output directory where the TFRecord shards will be saved. If the" " path is appended with `gs://` ('gs://tf-tpu', for example) then the TFRecord" " shards will be directly saved to a Google Cloud Storage bucket." , ) __snake_case : List[str] = parser.parse_args() return args def lowerCAmelCase_ ( __lowerCamelCase ): def fn(__lowerCamelCase ): return tokenizer(examples["text"] ) return fn def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = [] for i in range(len(tokenized_data["input_ids"] ) ): __snake_case : Tuple = { "input_ids": tf.train.Feature(intaa_list=tf.train.IntaaList(value=tokenized_data["input_ids"][i] ) ), "attention_mask": tf.train.Feature( intaa_list=tf.train.IntaaList(value=tokenized_data["attention_mask"][i] ) ), } __snake_case : List[Any] = tf.train.Features(feature=__lowerCamelCase ) __snake_case : str = tf.train.Example(features=__lowerCamelCase ) __snake_case : List[str] = example.SerializeToString() records.append(__lowerCamelCase ) return records def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Optional[int] = datasets.load_dataset(args.dataset_name , args.dataset_config , split=args.split ) if args.limit is not None: __snake_case : Optional[Any] = min(len(__lowerCamelCase ) , args.limit ) __snake_case : Dict = dataset.select(range(__lowerCamelCase ) ) print(F'Limiting the dataset to {args.limit} entries.' ) __snake_case : Dict = AutoTokenizer.from_pretrained(args.tokenizer_name_or_path ) # Handle output directory creation. # For serializing into a Google Cloud Storage Bucket, one needs to first # create a bucket. if "gs" not in args.output_dir: if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) __snake_case : Dict = os.path.join(args.output_dir , args.split ) if not os.path.exists(__lowerCamelCase ): os.makedirs(__lowerCamelCase ) else: __snake_case : str = os.path.join(args.output_dir , args.split ) # Tokenize the whole dataset at once. __snake_case : Any = tokenize_function(__lowerCamelCase ) __snake_case : Optional[Any] = dataset.map(__lowerCamelCase , batched=__lowerCamelCase , num_proc=4 , remove_columns=["text"] ) # We need to concatenate all our texts together, and then split the result # into chunks of a fixed size, which we will call block_size. To do this, we # will use the map method again, with the option batched=True. When we use batched=True, # the function we pass to map() will be passed multiple inputs at once, allowing us # to group them into more or fewer examples than we had in the input. # This allows us to create our new fixed-length samples. The advantage of this # method is that we don't lose a whole lot of content from the dataset compared to the # case where we simply tokenize with a pre-defined max_length. def group_texts(__lowerCamelCase ): # Concatenate all texts. __snake_case : List[str] = {k: sum(examples[k] , [] ) for k in examples.keys()} __snake_case : List[Any] = len(concatenated_examples[list(examples.keys() )[0]] ) # We drop the small remainder, though you could add padding instead if the model supports it # In this, as in all things, we advise you to follow your heart 🫀 __snake_case : Any = (total_length // args.max_length) * args.max_length # Split by chunks of max_len. __snake_case : int = { k: [t[i : i + args.max_length] for i in range(0 , __lowerCamelCase , args.max_length )] for k, t in concatenated_examples.items() } return result __snake_case : Any = dataset_tokenized.map(__lowerCamelCase , batched=__lowerCamelCase , batch_size=1_0_0_0 , num_proc=4 ) __snake_case : Optional[Any] = 0 __snake_case : Optional[Any] = 0 for shard in range(0 , len(__lowerCamelCase ) , args.shard_size ): __snake_case : List[str] = grouped_dataset[shard : shard + args.shard_size] __snake_case : Any = len(dataset_snapshot["input_ids"] ) __snake_case : List[Any] = os.path.join(__lowerCamelCase , F'dataset-{shard_count}-{records_containing}.tfrecord' ) __snake_case : Optional[Any] = get_serialized_examples(__lowerCamelCase ) with tf.io.TFRecordWriter(__lowerCamelCase ) as out_file: for i in range(len(__lowerCamelCase ) ): __snake_case : Union[str, Any] = serialized_examples[i] out_file.write(__lowerCamelCase ) print("Wrote file {} containing {} records".format(__lowerCamelCase , __lowerCamelCase ) ) shard_count += 1 total_records += records_containing with open(F'split-{args.split}-records-count.txt' , "w" ) as f: print(F'Total {args.split} records: {total_records}' , file=__lowerCamelCase ) if __name__ == "__main__": _snake_case : List[Any] = parse_args() main(args)
81
0
import torch from diffusers import CMStochasticIterativeScheduler from .test_schedulers import SchedulerCommonTest class lowerCAmelCase__ ( _lowerCAmelCase ): a__ : Optional[int] = (CMStochasticIterativeScheduler,) a__ : List[Any] = 10 def __A ( self : Any , **SCREAMING_SNAKE_CASE__ : Union[str, Any] ) -> Optional[Any]: __lowerCamelCase = { "num_train_timesteps": 2_01, "sigma_min": 0.002, "sigma_max": 80.0, } config.update(**SCREAMING_SNAKE_CASE__ ) return config def __A ( self : Optional[int] ) -> List[str]: __lowerCamelCase = 10 __lowerCamelCase = self.get_scheduler_config() __lowerCamelCase = self.scheduler_classes[0](**SCREAMING_SNAKE_CASE__ ) scheduler.set_timesteps(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = scheduler.timesteps[0] __lowerCamelCase = scheduler.timesteps[1] __lowerCamelCase = self.dummy_sample __lowerCamelCase = 0.1 * sample __lowerCamelCase = scheduler.step(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ).prev_sample __lowerCamelCase = scheduler.step(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def __A ( self : Dict ) -> str: for timesteps in [10, 50, 1_00, 10_00]: self.check_over_configs(num_train_timesteps=SCREAMING_SNAKE_CASE__ ) def __A ( self : Optional[int] ) -> Dict: for clip_denoised in [True, False]: self.check_over_configs(clip_denoised=SCREAMING_SNAKE_CASE__ ) def __A ( self : Dict ) -> Tuple: __lowerCamelCase = self.scheduler_classes[0] __lowerCamelCase = self.get_scheduler_config() __lowerCamelCase = scheduler_class(**SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = 1 scheduler.set_timesteps(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = scheduler.timesteps __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = self.dummy_model() __lowerCamelCase = self.dummy_sample_deter * scheduler.init_noise_sigma for i, t in enumerate(SCREAMING_SNAKE_CASE__ ): # 1. scale model input __lowerCamelCase = scheduler.scale_model_input(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # 2. predict noise residual __lowerCamelCase = model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # 3. predict previous sample x_t-1 __lowerCamelCase = scheduler.step(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , generator=SCREAMING_SNAKE_CASE__ ).prev_sample __lowerCamelCase = pred_prev_sample __lowerCamelCase = torch.sum(torch.abs(SCREAMING_SNAKE_CASE__ ) ) __lowerCamelCase = torch.mean(torch.abs(SCREAMING_SNAKE_CASE__ ) ) assert abs(result_sum.item() - 1_92.76_14 ) < 1e-2 assert abs(result_mean.item() - 0.2510 ) < 1e-3 def __A ( self : str ) -> Optional[Any]: __lowerCamelCase = self.scheduler_classes[0] __lowerCamelCase = self.get_scheduler_config() __lowerCamelCase = scheduler_class(**SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = [1_06, 0] scheduler.set_timesteps(timesteps=SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = scheduler.timesteps __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = self.dummy_model() __lowerCamelCase = self.dummy_sample_deter * scheduler.init_noise_sigma for t in timesteps: # 1. scale model input __lowerCamelCase = scheduler.scale_model_input(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # 2. predict noise residual __lowerCamelCase = model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # 3. predict previous sample x_t-1 __lowerCamelCase = scheduler.step(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , generator=SCREAMING_SNAKE_CASE__ ).prev_sample __lowerCamelCase = pred_prev_sample __lowerCamelCase = torch.sum(torch.abs(SCREAMING_SNAKE_CASE__ ) ) __lowerCamelCase = torch.mean(torch.abs(SCREAMING_SNAKE_CASE__ ) ) assert abs(result_sum.item() - 3_47.63_57 ) < 1e-2 assert abs(result_mean.item() - 0.4527 ) < 1e-3 def __A ( self : List[str] ) -> List[Any]: __lowerCamelCase = self.scheduler_classes[0] __lowerCamelCase = self.get_scheduler_config() __lowerCamelCase = scheduler_class(**SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = [39, 30, 12, 15, 0] with self.assertRaises(SCREAMING_SNAKE_CASE__ , msg='''`timesteps` must be in descending order.''' ): scheduler.set_timesteps(timesteps=SCREAMING_SNAKE_CASE__ ) def __A ( self : Tuple ) -> Union[str, Any]: __lowerCamelCase = self.scheduler_classes[0] __lowerCamelCase = self.get_scheduler_config() __lowerCamelCase = scheduler_class(**SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = [39, 30, 12, 1, 0] __lowerCamelCase = len(SCREAMING_SNAKE_CASE__ ) with self.assertRaises(SCREAMING_SNAKE_CASE__ , msg='''Can only pass one of `num_inference_steps` or `timesteps`.''' ): scheduler.set_timesteps(num_inference_steps=SCREAMING_SNAKE_CASE__ , timesteps=SCREAMING_SNAKE_CASE__ ) def __A ( self : List[Any] ) -> Optional[int]: __lowerCamelCase = self.scheduler_classes[0] __lowerCamelCase = self.get_scheduler_config() __lowerCamelCase = scheduler_class(**SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = [scheduler.config.num_train_timesteps] with self.assertRaises( SCREAMING_SNAKE_CASE__ , msg='''`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}''' , ): scheduler.set_timesteps(timesteps=SCREAMING_SNAKE_CASE__ )
298
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) _snake_case : List[Any] = "\\n Text data.\n Second line of data." _snake_case : Tuple = "file" @pytest.fixture(scope="session" ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / (FILE_PATH + ".zstd") __snake_case : Optional[Any] = bytes(__lowerCamelCase , "utf-8" ) with zstd.open(__lowerCamelCase , "wb" ) as f: f.write(__lowerCamelCase ) return path @pytest.fixture def lowerCAmelCase_ ( __lowerCamelCase ): with open(os.path.join(tmpfs.local_root_dir , __lowerCamelCase ) , "w" ) as f: f.write(__lowerCamelCase ) return FILE_PATH @pytest.mark.parametrize("compression_format" , ["gzip", "xz", "zstd"] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Optional[int] = {"gzip": gz_file, "xz": xz_file, "zstd": zstd_path} __snake_case : str = input_paths[compression_format] __snake_case : Optional[Any] = tmp_path / "cache" __snake_case : Optional[int] = DownloadConfig(cache_dir=__lowerCamelCase , extract_compressed_file=__lowerCamelCase ) __snake_case : Union[str, Any] = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) with open(__lowerCamelCase ) as f: __snake_case : Dict = f.read() with open(__lowerCamelCase ) as f: __snake_case : Tuple = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("default_extracted" , [True, False] ) @pytest.mark.parametrize("default_cache_dir" , [True, False] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Tuple = "custom_cache" __snake_case : List[str] = "custom_extracted_dir" __snake_case : Any = tmp_path / "custom_extracted_path" if default_extracted: __snake_case : List[Any] = ("downloads" if default_cache_dir else custom_cache_dir, "extracted") else: monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_DIR" , __lowerCamelCase ) monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_PATH" , str(__lowerCamelCase ) ) __snake_case : Optional[Any] = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) __snake_case : Optional[int] = xz_file __snake_case : Optional[int] = ( DownloadConfig(extract_compressed_file=__lowerCamelCase ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=__lowerCamelCase ) ) __snake_case : str = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) assert Path(__lowerCamelCase ).parent.parts[-2:] == expected def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : Optional[Any] = str(Path(__lowerCamelCase ).resolve() ) assert cached_path(__lowerCamelCase ) == text_file # relative path __snake_case : Any = str(Path(__lowerCamelCase ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(__lowerCamelCase ) == text_file def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : List[Any] = str(tmp_path.resolve() / "__missing_file__.txt" ) with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) # relative path __snake_case : Optional[int] = "./__missing_file__.txt" with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : str = get_from_cache(F'tmp://{tmpfs_file}' ) with open(__lowerCamelCase ) as f: __snake_case : Union[str, Any] = f.read() assert output_file_content == FILE_CONTENT @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( ): with pytest.raises(__lowerCamelCase ): cached_path("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[Any] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): http_get("https://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): http_head("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[str] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): ftp_get("ftp://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): ftp_head("ftp://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): fsspec_get("s3://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): fsspec_head("s3://huggingface.co" )
81
0
"""simple docstring""" from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging __SCREAMING_SNAKE_CASE =logging.get_logger(__name__) # pylint: disable=invalid-name class UpperCamelCase ( _lowerCAmelCase ): def __init__( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,) -> Tuple: '''simple docstring''' super().__init__() if hasattr(scheduler.config ,'steps_offset' ) and scheduler.config.steps_offset != 1: lowercase_ : Tuple = ( f'''The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`''' f''' should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure ''' "to update the config accordingly as leaving `steps_offset` might led to incorrect results" " in future versions. If you have downloaded this checkpoint from the Hugging Face Hub," " it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`" " file" ) deprecate('steps_offset!=1' ,'1.0.0' ,__UpperCamelCase ,standard_warn=__UpperCamelCase ) lowercase_ : Any = dict(scheduler.config ) lowercase_ : List[Any] = 1 lowercase_ : Tuple = FrozenDict(__UpperCamelCase ) if hasattr(scheduler.config ,'skip_prk_steps' ) and scheduler.config.skip_prk_steps is False: lowercase_ : List[str] = ( f'''The configuration file of this scheduler: {scheduler} has not set the configuration''' " `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make" " sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to" " incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face" " Hub, it would be very nice if you could open a Pull request for the" " `scheduler/scheduler_config.json` file" ) deprecate('skip_prk_steps not set' ,'1.0.0' ,__UpperCamelCase ,standard_warn=__UpperCamelCase ) lowercase_ : List[str] = dict(scheduler.config ) lowercase_ : List[str] = True lowercase_ : Any = FrozenDict(__UpperCamelCase ) if safety_checker is None: logger.warning( f'''You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure''' ' that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered' ' results in services or applications open to the public. Both the diffusers team and Hugging Face' ' strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling' ' it only for use-cases that involve analyzing network behavior or auditing its results. For more' ' information, please have a look at https://github.com/huggingface/diffusers/pull/254 .' ) self.register_modules( segmentation_model=__UpperCamelCase ,segmentation_processor=__UpperCamelCase ,vae=__UpperCamelCase ,text_encoder=__UpperCamelCase ,tokenizer=__UpperCamelCase ,unet=__UpperCamelCase ,scheduler=__UpperCamelCase ,safety_checker=__UpperCamelCase ,feature_extractor=__UpperCamelCase ,) def _UpperCAmelCase ( self ,__UpperCamelCase = "auto" ) -> Tuple: '''simple docstring''' if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory lowercase_ : Any = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(__UpperCamelCase ) def _UpperCAmelCase ( self ) -> Any: '''simple docstring''' self.enable_attention_slicing(__UpperCamelCase ) def _UpperCAmelCase ( self ) -> str: '''simple docstring''' if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) lowercase_ : Optional[int] = torch.device('cuda' ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(__UpperCamelCase ,__UpperCamelCase ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def _UpperCAmelCase ( self ) -> Any: '''simple docstring''' if self.device != torch.device('meta' ) or not hasattr(self.unet ,'_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(__UpperCamelCase ,'_hf_hook' ) and hasattr(module._hf_hook ,'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase = 512 ,__UpperCamelCase = 512 ,__UpperCamelCase = 50 ,__UpperCamelCase = 7.5 ,__UpperCamelCase = None ,__UpperCamelCase = 1 ,__UpperCamelCase = 0.0 ,__UpperCamelCase = None ,__UpperCamelCase = None ,__UpperCamelCase = "pil" ,__UpperCamelCase = True ,__UpperCamelCase = None ,__UpperCamelCase = 1 ,**__UpperCamelCase ,) -> List[str]: '''simple docstring''' lowercase_ : Tuple = self.segmentation_processor( text=[text] ,images=[image] ,padding='max_length' ,return_tensors='pt' ).to(self.device ) lowercase_ : str = self.segmentation_model(**__UpperCamelCase ) lowercase_ : Dict = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() lowercase_ : List[Any] = self.numpy_to_pil(__UpperCamelCase )[0].resize(image.size ) # Run inpainting pipeline with the generated mask lowercase_ : Tuple = StableDiffusionInpaintPipeline( vae=self.vae ,text_encoder=self.text_encoder ,tokenizer=self.tokenizer ,unet=self.unet ,scheduler=self.scheduler ,safety_checker=self.safety_checker ,feature_extractor=self.feature_extractor ,) return inpainting_pipeline( prompt=__UpperCamelCase ,image=__UpperCamelCase ,mask_image=__UpperCamelCase ,height=__UpperCamelCase ,width=__UpperCamelCase ,num_inference_steps=__UpperCamelCase ,guidance_scale=__UpperCamelCase ,negative_prompt=__UpperCamelCase ,num_images_per_prompt=__UpperCamelCase ,eta=__UpperCamelCase ,generator=__UpperCamelCase ,latents=__UpperCamelCase ,output_type=__UpperCamelCase ,return_dict=__UpperCamelCase ,callback=__UpperCamelCase ,callback_steps=__UpperCamelCase ,)
425
_snake_case : Optional[int] = {"a": ["c", "b"], "b": ["d", "e"], "c": [], "d": [], "e": []} _snake_case : Dict = ["a", "b", "c", "d", "e"] def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : List[str] = start # add current to visited visited.append(__lowerCamelCase ) __snake_case : List[Any] = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: __snake_case : Tuple = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # if all neighbors visited add current to sort sort.append(__lowerCamelCase ) # if all vertices haven't been visited select a new one to visit if len(__lowerCamelCase ) != len(__lowerCamelCase ): for vertice in vertices: if vertice not in visited: __snake_case : int = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # return sort return sort if __name__ == "__main__": _snake_case : List[Any] = topological_sort("a", [], []) print(sort)
81
0
'''simple docstring''' from __future__ import annotations import math def __snake_case ( _UpperCAmelCase : Optional[int]): if num <= 0: UpperCamelCase = f'{num}: Invalid input, please enter a positive integer.' raise ValueError(__lowerCamelCase) UpperCamelCase = [True] * (num + 1) UpperCamelCase = [] UpperCamelCase = 2 UpperCamelCase = int(math.sqrt(__lowerCamelCase)) while start <= end: # If start is a prime if sieve[start] is True: prime.append(__lowerCamelCase) # Set multiples of start be False for i in range(start * start, num + 1, __lowerCamelCase): if sieve[i] is True: UpperCamelCase = False start += 1 for j in range(end + 1, num + 1): if sieve[j] is True: prime.append(__lowerCamelCase) return prime if __name__ == "__main__": print(prime_sieve(int(input('Enter a positive integer: ').strip())))
212
def lowerCAmelCase_ ( ): return [ a * b * (1_0_0_0 - a - b) for a in range(1 , 9_9_9 ) for b in range(__lowerCamelCase , 9_9_9 ) if (a * a + b * b == (1_0_0_0 - a - b) ** 2) ][0] if __name__ == "__main__": print(f'''{solution() = }''')
81
0
import unittest from transformers.testing_utils import CaptureStdout from transformers.tools.python_interpreter import evaluate def __snake_case ( _lowerCAmelCase : int ) -> Any: return x + 2 class __magic_name__ ( unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self :Tuple ): '''simple docstring''' A_ : Optional[Any] = "x = 3" A_ : Any = {} A_ : Optional[int] = evaluate(snake_case , {} , state=snake_case ) assert result == 3 self.assertDictEqual(snake_case , {"x": 3} ) A_ : str = "x = y" A_ : List[Any] = {"y": 5} A_ : Any = evaluate(snake_case , {} , state=snake_case ) # evaluate returns the value of the last assignment. assert result == 5 self.assertDictEqual(snake_case , {"x": 5, "y": 5} ) def SCREAMING_SNAKE_CASE ( self :Any ): '''simple docstring''' A_ : int = "y = add_two(x)" A_ : Any = {"x": 3} A_ : str = evaluate(snake_case , {"add_two": add_two} , state=snake_case ) assert result == 5 self.assertDictEqual(snake_case , {"x": 3, "y": 5} ) # Won't work without the tool with CaptureStdout() as out: A_ : str = evaluate(snake_case , {} , state=snake_case ) assert result is None assert "tried to execute add_two" in out.out def SCREAMING_SNAKE_CASE ( self :Dict ): '''simple docstring''' A_ : str = "x = 3" A_ : List[Any] = {} A_ : List[Any] = evaluate(snake_case , {} , state=snake_case ) assert result == 3 self.assertDictEqual(snake_case , {"x": 3} ) def SCREAMING_SNAKE_CASE ( self :Optional[int] ): '''simple docstring''' A_ : Union[str, Any] = "test_dict = {'x': x, 'y': add_two(x)}" A_ : Tuple = {"x": 3} A_ : List[Any] = evaluate(snake_case , {"add_two": add_two} , state=snake_case ) self.assertDictEqual(snake_case , {"x": 3, "y": 5} ) self.assertDictEqual(snake_case , {"x": 3, "test_dict": {"x": 3, "y": 5}} ) def SCREAMING_SNAKE_CASE ( self :Optional[int] ): '''simple docstring''' A_ : int = "x = 3\ny = 5" A_ : Optional[int] = {} A_ : List[str] = evaluate(snake_case , {} , state=snake_case ) # evaluate returns the value of the last assignment. assert result == 5 self.assertDictEqual(snake_case , {"x": 3, "y": 5} ) def SCREAMING_SNAKE_CASE ( self :Dict ): '''simple docstring''' A_ : List[Any] = "text = f'This is x: {x}.'" A_ : List[Any] = {"x": 3} A_ : List[str] = evaluate(snake_case , {} , state=snake_case ) # evaluate returns the value of the last assignment. assert result == "This is x: 3." self.assertDictEqual(snake_case , {"x": 3, "text": "This is x: 3."} ) def SCREAMING_SNAKE_CASE ( self :Any ): '''simple docstring''' A_ : List[str] = "if x <= 3:\n y = 2\nelse:\n y = 5" A_ : Tuple = {"x": 3} A_ : int = evaluate(snake_case , {} , state=snake_case ) # evaluate returns the value of the last assignment. assert result == 2 self.assertDictEqual(snake_case , {"x": 3, "y": 2} ) A_ : str = {"x": 8} A_ : List[str] = evaluate(snake_case , {} , state=snake_case ) # evaluate returns the value of the last assignment. assert result == 5 self.assertDictEqual(snake_case , {"x": 8, "y": 5} ) def SCREAMING_SNAKE_CASE ( self :int ): '''simple docstring''' A_ : Tuple = "test_list = [x, add_two(x)]" A_ : List[str] = {"x": 3} A_ : Any = evaluate(snake_case , {"add_two": add_two} , state=snake_case ) self.assertListEqual(snake_case , [3, 5] ) self.assertDictEqual(snake_case , {"x": 3, "test_list": [3, 5]} ) def SCREAMING_SNAKE_CASE ( self :Union[str, Any] ): '''simple docstring''' A_ : Optional[int] = "y = x" A_ : Any = {"x": 3} A_ : Union[str, Any] = evaluate(snake_case , {} , state=snake_case ) assert result == 3 self.assertDictEqual(snake_case , {"x": 3, "y": 3} ) def SCREAMING_SNAKE_CASE ( self :Any ): '''simple docstring''' A_ : Optional[Any] = "test_list = [x, add_two(x)]\ntest_list[1]" A_ : str = {"x": 3} A_ : Optional[int] = evaluate(snake_case , {"add_two": add_two} , state=snake_case ) assert result == 5 self.assertDictEqual(snake_case , {"x": 3, "test_list": [3, 5]} ) A_ : str = "test_dict = {'x': x, 'y': add_two(x)}\ntest_dict['y']" A_ : Optional[Any] = {"x": 3} A_ : Union[str, Any] = evaluate(snake_case , {"add_two": add_two} , state=snake_case ) assert result == 5 self.assertDictEqual(snake_case , {"x": 3, "test_dict": {"x": 3, "y": 5}} ) def SCREAMING_SNAKE_CASE ( self :Dict ): '''simple docstring''' A_ : Any = "x = 0\nfor i in range(3):\n x = i" A_ : Union[str, Any] = {} A_ : Any = evaluate(snake_case , {"range": range} , state=snake_case ) assert result == 2 self.assertDictEqual(snake_case , {"x": 2, "i": 2} )
454
from math import sqrt def lowerCAmelCase_ ( __lowerCamelCase = 1_0_0_0_0_0_0 ): __snake_case : int = 0 __snake_case : int = 0 __snake_case : int while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(__lowerCamelCase , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(f'''{solution() = }''')
81
0
"""simple docstring""" from transformers import BertTokenizer, EncoderDecoderModel, SeqaSeqTrainer, SeqaSeqTrainingArguments from transformers.testing_utils import TestCasePlus, require_torch, slow from transformers.utils import is_datasets_available if is_datasets_available(): import datasets class _UpperCamelCase ( _lowerCAmelCase): @slow @require_torch def A (self ): """simple docstring""" A__ = EncoderDecoderModel.from_encoder_decoder_pretrained("""prajjwal1/bert-tiny""" , """prajjwal1/bert-tiny""" ) A__ = BertTokenizer.from_pretrained("""bert-base-uncased""" ) A__ = bertabert.config.encoder.vocab_size A__ = tokenizer.sep_token_id A__ = tokenizer.cls_token_id A__ = 1_2_8 A__ = datasets.load_dataset("""cnn_dailymail""" , """3.0.0""" , split="""train[:1%]""" ) A__ = datasets.load_dataset("""cnn_dailymail""" , """3.0.0""" , split="""validation[:1%]""" ) A__ = train_dataset.select(range(3_2 ) ) A__ = val_dataset.select(range(1_6 ) ) A__ = 4 def _map_to_encoder_decoder_inputs(lowerCamelCase__ ): # Tokenizer will automatically set [BOS] <text> [EOS] A__ = tokenizer(batch["""article"""] , padding="""max_length""" , truncation=lowerCamelCase__ , max_length=5_1_2 ) A__ = tokenizer(batch["""highlights"""] , padding="""max_length""" , truncation=lowerCamelCase__ , max_length=1_2_8 ) A__ = inputs.input_ids A__ = inputs.attention_mask A__ = outputs.input_ids A__ = outputs.input_ids.copy() A__ = [ [-1_0_0 if token == tokenizer.pad_token_id else token for token in labels] for labels in batch["labels"] ] A__ = outputs.attention_mask assert all(len(lowerCamelCase__ ) == 5_1_2 for x in inputs.input_ids ) assert all(len(lowerCamelCase__ ) == 1_2_8 for x in outputs.input_ids ) return batch def _compute_metrics(lowerCamelCase__ ): A__ = pred.label_ids A__ = pred.predictions # all unnecessary tokens are removed A__ = tokenizer.batch_decode(lowerCamelCase__ , skip_special_tokens=lowerCamelCase__ ) A__ = tokenizer.batch_decode(lowerCamelCase__ , skip_special_tokens=lowerCamelCase__ ) A__ = sum([int(pred_str[i] == label_str[i] ) for i in range(len(lowerCamelCase__ ) )] ) / len(lowerCamelCase__ ) return {"accuracy": accuracy} # map train dataset A__ = train_dataset.map( _map_to_encoder_decoder_inputs , batched=lowerCamelCase__ , batch_size=lowerCamelCase__ , remove_columns=["""article""", """highlights"""] , ) train_dataset.set_format( type="""torch""" , columns=["""input_ids""", """attention_mask""", """decoder_input_ids""", """decoder_attention_mask""", """labels"""] , ) # same for validation dataset A__ = val_dataset.map( _map_to_encoder_decoder_inputs , batched=lowerCamelCase__ , batch_size=lowerCamelCase__ , remove_columns=["""article""", """highlights"""] , ) val_dataset.set_format( type="""torch""" , columns=["""input_ids""", """attention_mask""", """decoder_input_ids""", """decoder_attention_mask""", """labels"""] , ) A__ = self.get_auto_remove_tmp_dir() A__ = SeqaSeqTrainingArguments( output_dir=lowerCamelCase__ , per_device_train_batch_size=lowerCamelCase__ , per_device_eval_batch_size=lowerCamelCase__ , predict_with_generate=lowerCamelCase__ , evaluation_strategy="""steps""" , do_train=lowerCamelCase__ , do_eval=lowerCamelCase__ , warmup_steps=0 , eval_steps=2 , logging_steps=2 , ) # instantiate trainer A__ = SeqaSeqTrainer( model=lowerCamelCase__ , args=lowerCamelCase__ , compute_metrics=_compute_metrics , train_dataset=lowerCamelCase__ , eval_dataset=lowerCamelCase__ , tokenizer=lowerCamelCase__ , ) # start training trainer.train()
574
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class a (_lowerCAmelCase ): """simple docstring""" def __snake_case ( self : str ) -> str: __snake_case : Tuple = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCamelCase , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "neck_hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "num_attention_heads" ) ) class a : """simple docstring""" def __init__( self : Optional[int] , lowerCamelCase : List[str] , lowerCamelCase : Tuple=13 , lowerCamelCase : str=32 , lowerCamelCase : Dict=2 , lowerCamelCase : List[str]=3 , lowerCamelCase : Any=640 , lowerCamelCase : Optional[Any]=4 , lowerCamelCase : Tuple="silu" , lowerCamelCase : int=3 , lowerCamelCase : Dict=32 , lowerCamelCase : str=0.1 , lowerCamelCase : Optional[int]=0.1 , lowerCamelCase : Optional[Any]=0.1 , lowerCamelCase : Dict=0.02 , lowerCamelCase : Union[str, Any]=True , lowerCamelCase : Optional[int]=True , lowerCamelCase : Union[str, Any]=10 , lowerCamelCase : int=None , ) -> str: __snake_case : Optional[Any] = parent __snake_case : Optional[Any] = batch_size __snake_case : Any = image_size __snake_case : List[Any] = patch_size __snake_case : Any = num_channels __snake_case : Union[str, Any] = last_hidden_size __snake_case : Any = num_attention_heads __snake_case : Any = hidden_act __snake_case : Tuple = conv_kernel_size __snake_case : Any = output_stride __snake_case : Any = hidden_dropout_prob __snake_case : List[Any] = attention_probs_dropout_prob __snake_case : Optional[Any] = classifier_dropout_prob __snake_case : Union[str, Any] = use_labels __snake_case : Optional[int] = is_training __snake_case : Dict = num_labels __snake_case : Any = initializer_range __snake_case : Optional[int] = scope def __snake_case ( self : str ) -> Union[str, Any]: __snake_case : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : List[Any] = None __snake_case : Optional[int] = None if self.use_labels: __snake_case : Optional[int] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : Any = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __snake_case : Optional[Any] = self.get_config() return config, pixel_values, labels, pixel_labels def __snake_case ( self : Any ) -> Union[str, Any]: return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __snake_case ( self : Any , lowerCamelCase : Tuple , lowerCamelCase : int , lowerCamelCase : Tuple , lowerCamelCase : Optional[int] ) -> Dict: __snake_case : List[Any] = MobileViTModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[str] = model(lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[Any] , lowerCamelCase : List[str] , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] , lowerCamelCase : Tuple ) -> List[str]: __snake_case : str = self.num_labels __snake_case : List[Any] = MobileViTForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : Dict , lowerCamelCase : Any , lowerCamelCase : Any , lowerCamelCase : Optional[Any] , lowerCamelCase : Dict ) -> Dict: __snake_case : Union[str, Any] = self.num_labels __snake_case : Optional[int] = MobileViTForSemanticSegmentation(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Tuple = model(lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[int] ) -> List[Any]: __snake_case : Optional[Any] = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case , __snake_case : Union[str, Any] = config_and_inputs __snake_case : Dict = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) __UpperCAmelCase : Optional[Any] = ( { "feature-extraction": MobileViTModel, "image-classification": MobileViTForImageClassification, "image-segmentation": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) __UpperCAmelCase : List[str] = False __UpperCAmelCase : int = False __UpperCAmelCase : Optional[int] = False __UpperCAmelCase : Optional[int] = False def __snake_case ( self : Optional[int] ) -> Dict: __snake_case : Tuple = MobileViTModelTester(self ) __snake_case : Any = MobileViTConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase ) def __snake_case ( self : Optional[int] ) -> Dict: self.config_tester.run_common_tests() @unittest.skip(reason="MobileViT does not use inputs_embeds" ) def __snake_case ( self : Dict ) -> Any: pass @unittest.skip(reason="MobileViT does not support input and output embeddings" ) def __snake_case ( self : Dict ) -> List[Any]: pass @unittest.skip(reason="MobileViT does not output attentions" ) def __snake_case ( self : int ) -> Dict: pass def __snake_case ( self : int ) -> Union[str, Any]: __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Optional[Any] = model_class(lowerCamelCase ) __snake_case : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : int = [*signature.parameters.keys()] __snake_case : List[str] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __snake_case ( self : int ) -> Tuple: pass def __snake_case ( self : Any ) -> Tuple: __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def __snake_case ( self : Any ) -> str: def check_hidden_states_output(lowerCamelCase : Union[str, Any] , lowerCamelCase : Dict , lowerCamelCase : Any ): __snake_case : int = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : int = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : Union[str, Any] = outputs.hidden_states __snake_case : int = 5 self.assertEqual(len(lowerCamelCase ) , lowerCamelCase ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. __snake_case : List[Any] = 2 for i in range(len(lowerCamelCase ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) __snake_case , __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : str = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : List[Any] = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Any ) -> Any: __snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) def __snake_case ( self : List[str] ) -> List[str]: __snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowerCamelCase ) @slow def __snake_case ( self : List[str] ) -> Any: for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : List[str] = MobileViTModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def lowerCAmelCase_ ( ): __snake_case : Optional[int] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a (unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self : str ) -> Dict: return MobileViTImageProcessor.from_pretrained("apple/mobilevit-xx-small" ) if is_vision_available() else None @slow def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : Tuple = MobileViTForImageClassification.from_pretrained("apple/mobilevit-xx-small" ).to(lowerCamelCase ) __snake_case : Optional[Any] = self.default_image_processor __snake_case : Union[str, Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Dict = model(**lowerCamelCase ) # verify the logits __snake_case : Union[str, Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([-1.93_64, -1.23_27, -0.46_53] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : str ) -> Optional[int]: __snake_case : Optional[int] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : str = model.to(lowerCamelCase ) __snake_case : int = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Optional[int] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : List[str] = model(**lowerCamelCase ) __snake_case : Union[str, Any] = outputs.logits # verify the logits __snake_case : Tuple = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , lowerCamelCase ) __snake_case : Union[str, Any] = torch.tensor( [ [[6.97_13, 6.97_86, 7.24_22], [7.28_93, 7.28_25, 7.44_46], [7.65_80, 7.87_97, 7.94_20]], [[-10.68_69, -10.32_50, -10.34_71], [-10.42_28, -9.98_68, -9.71_32], [-11.04_05, -11.02_21, -10.73_18]], [[-3.30_89, -2.85_39, -2.67_40], [-3.27_06, -2.56_21, -2.51_08], [-3.25_34, -2.66_15, -2.66_51]], ] , device=lowerCamelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : Union[str, Any] ) -> Optional[int]: __snake_case : Optional[Any] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Tuple = model.to(lowerCamelCase ) __snake_case : Dict = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : List[Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Any = model(**lowerCamelCase ) __snake_case : Dict = outputs.logits.detach().cpu() __snake_case : Any = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase , target_sizes=[(50, 60)] ) __snake_case : int = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , lowerCamelCase ) __snake_case : List[str] = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase ) __snake_case : Optional[int] = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , lowerCamelCase )
81
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __a: Optional[Any] = { "configuration_timesformer": ["TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "TimesformerConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a: Dict = [ "TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "TimesformerModel", "TimesformerForVideoClassification", "TimesformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_timesformer import TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimesformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timesformer import ( TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimesformerForVideoClassification, TimesformerModel, TimesformerPreTrainedModel, ) else: import sys __a: str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
152
import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor _snake_case : int = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[int] , *lowerCamelCase : List[Any] , **lowerCamelCase : int ) -> None: warnings.warn( "The class ImageGPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use ImageGPTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
from math import sqrt def A_ ( lowercase_ = 1000000 ) -> int: _snake_case : int = 0 _snake_case : int = 0 _snake_case : int while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(__lowerCamelCase , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(F"""{solution() = }""")
326
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _snake_case : Optional[Any] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[str] = ["pixel_values"] def __init__( self : List[Any] , lowerCamelCase : bool = True , lowerCamelCase : Union[int, float] = 1 / 255 , lowerCamelCase : bool = True , lowerCamelCase : int = 8 , **lowerCamelCase : Tuple , ) -> None: super().__init__(**lowerCamelCase ) __snake_case : Dict = do_rescale __snake_case : Dict = rescale_factor __snake_case : Optional[Any] = do_pad __snake_case : Tuple = pad_size def __snake_case ( self : Dict , lowerCamelCase : np.ndarray , lowerCamelCase : float , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase : Optional[int] ) -> np.ndarray: return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def __snake_case ( self : Optional[Any] , lowerCamelCase : np.ndarray , lowerCamelCase : int , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None ) -> Tuple: __snake_case , __snake_case : List[str] = get_image_size(lowerCamelCase ) __snake_case : Optional[Any] = (old_height // size + 1) * size - old_height __snake_case : List[Any] = (old_width // size + 1) * size - old_width return pad(lowerCamelCase , ((0, pad_height), (0, pad_width)) , mode="symmetric" , data_format=lowerCamelCase ) def __snake_case ( self : Tuple , lowerCamelCase : ImageInput , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[float] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[int] = None , lowerCamelCase : Optional[Union[str, TensorType]] = None , lowerCamelCase : Union[str, ChannelDimension] = ChannelDimension.FIRST , **lowerCamelCase : Union[str, Any] , ) -> List[str]: __snake_case : int = do_rescale if do_rescale is not None else self.do_rescale __snake_case : Dict = rescale_factor if rescale_factor is not None else self.rescale_factor __snake_case : str = do_pad if do_pad is not None else self.do_pad __snake_case : Any = pad_size if pad_size is not None else self.pad_size __snake_case : int = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) # All transformations expect numpy arrays. __snake_case : str = [to_numpy_array(lowerCamelCase ) for image in images] if do_rescale: __snake_case : Optional[int] = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_pad: __snake_case : Optional[Any] = [self.pad(lowerCamelCase , size=lowerCamelCase ) for image in images] __snake_case : int = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __snake_case : Union[str, Any] = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
81
0
"""simple docstring""" import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __magic_name__ = logging.get_logger(__name__) __magic_name__ = { "vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_config_file": "tokenizer_config.json", } __magic_name__ = { "vocab_file": { "facebook/blenderbot_small-90M": "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json" }, "merges_file": { "facebook/blenderbot_small-90M": "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt" }, "tokenizer_config_file": { "facebook/blenderbot_small-90M": ( "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json" ) }, } __magic_name__ = {"facebook/blenderbot_small-90M": 512} def _lowerCAmelCase ( UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = set() __SCREAMING_SNAKE_CASE = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __SCREAMING_SNAKE_CASE = char __SCREAMING_SNAKE_CASE = set(__lowerCamelCase ) return pairs class SCREAMING_SNAKE_CASE_ ( _lowerCAmelCase ): """simple docstring""" __lowercase : Tuple = VOCAB_FILES_NAMES __lowercase : int = PRETRAINED_VOCAB_FILES_MAP __lowercase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowercase : Union[str, Any] = ["input_ids", "attention_mask"] def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__="__start__" , lowerCAmelCase__="__end__" , lowerCAmelCase__="__unk__" , lowerCAmelCase__="__null__" , **lowerCAmelCase__ , ): super().__init__(unk_token=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , **lowerCAmelCase__) with open(lowerCAmelCase__ , encoding="""utf-8""") as vocab_handle: __SCREAMING_SNAKE_CASE = json.load(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = {v: k for k, v in self.encoder.items()} with open(lowerCAmelCase__ , encoding="""utf-8""") as merges_handle: __SCREAMING_SNAKE_CASE = merges_handle.read().split("""\n""")[1:-1] __SCREAMING_SNAKE_CASE = [tuple(merge.split()) for merge in merges] __SCREAMING_SNAKE_CASE = dict(zip(lowerCAmelCase__ , range(len(lowerCAmelCase__)))) __SCREAMING_SNAKE_CASE = {} @property def snake_case_ ( self): return len(self.encoder) def snake_case_ ( self): return dict(self.encoder , **self.added_tokens_encoder) def snake_case_ ( self , lowerCAmelCase__): if token in self.cache: return self.cache[token] __SCREAMING_SNAKE_CASE = re.sub("""([.,!?()])""" , R""" \1""" , lowerCAmelCase__) __SCREAMING_SNAKE_CASE = re.sub("""(')""" , R""" \1 """ , lowerCAmelCase__) __SCREAMING_SNAKE_CASE = re.sub(R"""\s{2,}""" , """ """ , lowerCAmelCase__) if "\n" in token: __SCREAMING_SNAKE_CASE = token.replace("""\n""" , """ __newln__""") __SCREAMING_SNAKE_CASE = token.split(""" """) __SCREAMING_SNAKE_CASE = [] for token in tokens: if not len(lowerCAmelCase__): continue __SCREAMING_SNAKE_CASE = token.lower() __SCREAMING_SNAKE_CASE = tuple(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = tuple(list(word[:-1]) + [word[-1] + """</w>"""]) __SCREAMING_SNAKE_CASE = get_pairs(lowerCAmelCase__) if not pairs: words.append(lowerCAmelCase__) continue while True: __SCREAMING_SNAKE_CASE = min(lowerCAmelCase__ , key=lambda lowerCAmelCase__: self.bpe_ranks.get(lowerCAmelCase__ , float("""inf"""))) if bigram not in self.bpe_ranks: break __SCREAMING_SNAKE_CASE = bigram __SCREAMING_SNAKE_CASE = [] __SCREAMING_SNAKE_CASE = 0 while i < len(lowerCAmelCase__): try: __SCREAMING_SNAKE_CASE = word.index(lowerCAmelCase__ , lowerCAmelCase__) new_word.extend(word[i:j]) __SCREAMING_SNAKE_CASE = j except ValueError: new_word.extend(word[i:]) break if word[i] == first and i < len(lowerCAmelCase__) - 1 and word[i + 1] == second: new_word.append(first + second) i += 2 else: new_word.append(word[i]) i += 1 __SCREAMING_SNAKE_CASE = tuple(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = new_word if len(lowerCAmelCase__) == 1: break else: __SCREAMING_SNAKE_CASE = get_pairs(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = "@@ ".join(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = word[:-4] __SCREAMING_SNAKE_CASE = word words.append(lowerCAmelCase__) return " ".join(lowerCAmelCase__) def snake_case_ ( self , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = [] __SCREAMING_SNAKE_CASE = re.findall(R"""\S+\n?""" , lowerCAmelCase__) for token in words: split_tokens.extend(list(self.bpe(lowerCAmelCase__).split(""" """))) return split_tokens def snake_case_ ( self , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = token.lower() return self.encoder.get(lowerCAmelCase__ , self.encoder.get(self.unk_token)) def snake_case_ ( self , lowerCAmelCase__): return self.decoder.get(lowerCAmelCase__ , self.unk_token) def snake_case_ ( self , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = " ".join(lowerCAmelCase__).replace("""@@ """ , """""").strip() return out_string def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ = None): if not os.path.isdir(lowerCAmelCase__): logger.error(f"Vocabulary path ({save_directory}) should be a directory") return __SCREAMING_SNAKE_CASE = os.path.join( lowerCAmelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""]) __SCREAMING_SNAKE_CASE = os.path.join( lowerCAmelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""]) with open(lowerCAmelCase__ , """w""" , encoding="""utf-8""") as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=lowerCAmelCase__ , ensure_ascii=lowerCAmelCase__) + """\n""") __SCREAMING_SNAKE_CASE = 0 with open(lowerCAmelCase__ , """w""" , encoding="""utf-8""") as writer: writer.write("""#version: 0.2\n""") for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda lowerCAmelCase__: kv[1]): if index != token_index: logger.warning( f"Saving vocabulary to {merge_file}: BPE merge indices are not consecutive." """ Please check that the tokenizer is not corrupted!""") __SCREAMING_SNAKE_CASE = token_index writer.write(""" """.join(lowerCAmelCase__) + """\n""") index += 1 return vocab_file, merge_file
155
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case : Any = { "configuration_informer": [ "INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "InformerConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "InformerForPrediction", "InformerModel", "InformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys _snake_case : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
0
import math def lowerCAmelCase_ (lowercase__ : Optional[int] , lowercase__ : Optional[int] ) -> Dict: '''simple docstring''' lowerCAmelCase__ = len(__lowerCamelCase ) lowerCAmelCase__ = int(math.floor(math.sqrt(__lowerCamelCase ) ) ) lowerCAmelCase__ = 0 while arr[min(__lowerCamelCase , __lowerCamelCase ) - 1] < x: lowerCAmelCase__ = step step += int(math.floor(math.sqrt(__lowerCamelCase ) ) ) if prev >= n: return -1 while arr[prev] < x: lowerCAmelCase__ = prev + 1 if prev == min(__lowerCamelCase , __lowerCamelCase ): return -1 if arr[prev] == x: return prev return -1 if __name__ == "__main__": _UpperCAmelCase : int = input("Enter numbers separated by a comma:\n").strip() _UpperCAmelCase : Tuple = [int(item) for item in user_input.split(",")] _UpperCAmelCase : str = int(input("Enter the number to be searched:\n")) _UpperCAmelCase : List[str] = jump_search(arr, x) if res == -1: print("Number not found!") else: print(F'''Number {x} is at index {res}''')
668
from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging _snake_case : Any = logging.get_logger(__name__) # pylint: disable=invalid-name class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[Any] , lowerCamelCase : CLIPSegForImageSegmentation , lowerCamelCase : CLIPSegProcessor , lowerCamelCase : AutoencoderKL , lowerCamelCase : CLIPTextModel , lowerCamelCase : CLIPTokenizer , lowerCamelCase : UNetaDConditionModel , lowerCamelCase : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , lowerCamelCase : StableDiffusionSafetyChecker , lowerCamelCase : CLIPImageProcessor , ) -> Tuple: super().__init__() if hasattr(scheduler.config , "steps_offset" ) and scheduler.config.steps_offset != 1: __snake_case : Tuple = ( F'The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`' F' should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure ' "to update the config accordingly as leaving `steps_offset` might led to incorrect results" " in future versions. If you have downloaded this checkpoint from the Hugging Face Hub," " it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`" " file" ) deprecate("steps_offset!=1" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : Any = dict(scheduler.config ) __snake_case : List[Any] = 1 __snake_case : Tuple = FrozenDict(lowerCamelCase ) if hasattr(scheduler.config , "skip_prk_steps" ) and scheduler.config.skip_prk_steps is False: __snake_case : List[str] = ( F'The configuration file of this scheduler: {scheduler} has not set the configuration' " `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make" " sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to" " incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face" " Hub, it would be very nice if you could open a Pull request for the" " `scheduler/scheduler_config.json` file" ) deprecate("skip_prk_steps not set" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : List[str] = dict(scheduler.config ) __snake_case : List[str] = True __snake_case : Any = FrozenDict(lowerCamelCase ) if safety_checker is None: logger.warning( F'You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure' " that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered" " results in services or applications open to the public. Both the diffusers team and Hugging Face" " strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling" " it only for use-cases that involve analyzing network behavior or auditing its results. For more" " information, please have a look at https://github.com/huggingface/diffusers/pull/254 ." ) self.register_modules( segmentation_model=lowerCamelCase , segmentation_processor=lowerCamelCase , vae=lowerCamelCase , text_encoder=lowerCamelCase , tokenizer=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase , safety_checker=lowerCamelCase , feature_extractor=lowerCamelCase , ) def __snake_case ( self : Dict , lowerCamelCase : Optional[Union[str, int]] = "auto" ) -> Tuple: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __snake_case : Any = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowerCamelCase ) def __snake_case ( self : List[Any] ) -> Any: self.enable_attention_slicing(lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> str: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) __snake_case : Optional[int] = torch.device("cuda" ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(lowerCamelCase , lowerCamelCase ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __snake_case ( self : int ) -> Any: if self.device != torch.device("meta" ) or not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowerCamelCase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self : List[Any] , lowerCamelCase : Union[str, List[str]] , lowerCamelCase : Union[torch.FloatTensor, PIL.Image.Image] , lowerCamelCase : str , lowerCamelCase : int = 512 , lowerCamelCase : int = 512 , lowerCamelCase : int = 50 , lowerCamelCase : float = 7.5 , lowerCamelCase : Optional[Union[str, List[str]]] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : float = 0.0 , lowerCamelCase : Optional[torch.Generator] = None , lowerCamelCase : Optional[torch.FloatTensor] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , lowerCamelCase : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , lowerCamelCase : int = 1 , **lowerCamelCase : Dict , ) -> List[str]: __snake_case : Tuple = self.segmentation_processor( text=[text] , images=[image] , padding="max_length" , return_tensors="pt" ).to(self.device ) __snake_case : str = self.segmentation_model(**lowerCamelCase ) __snake_case : Dict = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() __snake_case : List[Any] = self.numpy_to_pil(lowerCamelCase )[0].resize(image.size ) # Run inpainting pipeline with the generated mask __snake_case : Tuple = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=lowerCamelCase , image=lowerCamelCase , mask_image=lowerCamelCase , height=lowerCamelCase , width=lowerCamelCase , num_inference_steps=lowerCamelCase , guidance_scale=lowerCamelCase , negative_prompt=lowerCamelCase , num_images_per_prompt=lowerCamelCase , eta=lowerCamelCase , generator=lowerCamelCase , latents=lowerCamelCase , output_type=lowerCamelCase , return_dict=lowerCamelCase , callback=lowerCamelCase , callback_steps=lowerCamelCase , )
81
0
'''simple docstring''' import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XGLMTokenizer, XGLMTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __UpperCamelCase = get_tests_dir("fixtures/test_sentencepiece.model") @require_sentencepiece @require_tokenizers class _A ( _lowerCAmelCase , unittest.TestCase ): lowercase__: List[str] = XGLMTokenizer lowercase__: Union[str, Any] = XGLMTokenizerFast lowercase__: List[Any] = True lowercase__: Optional[Any] = True def lowercase__ ( self : str ) -> int: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __snake_case : Optional[int] = XGLMTokenizer(__magic_name__ , keep_accents=__magic_name__ ) tokenizer.save_pretrained(self.tmpdirname ) def lowercase__ ( self : str ) -> List[Any]: """simple docstring""" __snake_case : Optional[int] = "<pad>" __snake_case : Optional[int] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__magic_name__ ) , __magic_name__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__magic_name__ ) , __magic_name__ ) def lowercase__ ( self : int ) -> Any: """simple docstring""" __snake_case : List[str] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(len(__magic_name__ ) , 10_08 ) def lowercase__ ( self : str ) -> Tuple: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 10_08 ) def lowercase__ ( self : Optional[int] ) -> Any: """simple docstring""" __snake_case : Any = XGLMTokenizer(__magic_name__ , keep_accents=__magic_name__ ) __snake_case : Dict = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(__magic_name__ , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__magic_name__ ) , [value + tokenizer.fairseq_offset for value in [2_85, 46, 10, 1_70, 3_82]] , ) __snake_case : Dict = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( __magic_name__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) __snake_case : Any = tokenizer.convert_tokens_to_ids(__magic_name__ ) self.assertListEqual( __magic_name__ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) __snake_case : Dict = tokenizer.convert_ids_to_tokens(__magic_name__ ) self.assertListEqual( __magic_name__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """.""", ] , ) @cached_property def lowercase__ ( self : List[Any] ) -> Optional[int]: """simple docstring""" return XGLMTokenizer.from_pretrained("""facebook/xglm-564M""" ) def lowercase__ ( self : Optional[int] ) -> Optional[Any]: """simple docstring""" with tempfile.NamedTemporaryFile() as f: shutil.copyfile(__magic_name__ , f.name ) __snake_case : int = XGLMTokenizer(f.name , keep_accents=__magic_name__ ) __snake_case : List[str] = pickle.dumps(__magic_name__ ) pickle.loads(__magic_name__ ) def lowercase__ ( self : Optional[Any] ) -> List[str]: """simple docstring""" if not self.test_rust_tokenizer: return __snake_case : Any = self.get_tokenizer() __snake_case : List[str] = self.get_rust_tokenizer() __snake_case : Optional[Any] = "I was born in 92000, and this is falsé." __snake_case : Optional[int] = tokenizer.tokenize(__magic_name__ ) __snake_case : List[str] = rust_tokenizer.tokenize(__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) __snake_case : List[Any] = tokenizer.encode(__magic_name__ , add_special_tokens=__magic_name__ ) __snake_case : Any = rust_tokenizer.encode(__magic_name__ , add_special_tokens=__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) __snake_case : int = self.get_rust_tokenizer() __snake_case : Optional[int] = tokenizer.encode(__magic_name__ ) __snake_case : List[Any] = rust_tokenizer.encode(__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) @slow def lowercase__ ( self : Optional[Any] ) -> List[str]: """simple docstring""" __snake_case : List[str] = "Hello World!" __snake_case : str = [2, 3_12_27, 44_47, 35] self.assertListEqual(__magic_name__ , self.big_tokenizer.encode(__magic_name__ ) ) @slow def lowercase__ ( self : Any ) -> List[Any]: """simple docstring""" __snake_case : Optional[int] = ( "This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will" " add words that should not exsist and be tokenized to unk, such as saoneuhaoesuth" ) # fmt: off __snake_case : Any = [2, 10_18, 67, 11, 19_88, 26_17, 56_31, 2_78, 11, 34_07, 48, 7_16_30, 2_80_85, 4, 32_34, 1_57, 13, 6, 5, 6, 4, 35_26, 7_68, 15, 6_59, 57, 2_98, 39_83, 8_64, 1_29, 21, 6, 5, 1_36_75, 3_77, 6_52, 75_80, 1_03_41, 1_55, 28_17, 4_22, 16_66, 7, 16_74, 53, 1_13, 20_22_77, 1_78_92, 33, 60, 87, 4, 32_34, 1_57, 61, 26_67, 5_23_76, 19, 88, 23, 7_35] # fmt: on self.assertListEqual(__magic_name__ , self.big_tokenizer.encode(__magic_name__ ) ) @slow def lowercase__ ( self : str ) -> int: """simple docstring""" __snake_case : List[str] = { "input_ids": [[2, 10_88_25, 11_63, 15, 8_80_10, 4_73, 1_58_98, 1_57, 1_36_72, 18_57, 3_12, 8, 23_80_21, 11_63, 53, 1_36_72, 18_57, 3_12, 8, 5_32_83, 18_23_96, 8, 1_85_66, 16, 3_67_33, 41_01, 8, 2_30, 24_40_17, 12_25_53, 7, 15, 13_25_97, 4, 2_93, 1_25_11, 76_10, 4, 34_14, 13_25_97, 9, 4, 3_23_61, 3_62, 4, 7_34, 2_85_12, 3_25_69, 18, 4, 3_23_61, 2_60_96, 1_49_82, 73, 1_87_15, 2_14_33, 23_52_61, 15, 4_92, 1_24_27, 16, 53, 1_87_15, 2_14_33, 6_54_54, 15, 2_36_59, 5_63, 16, 2_78, 5_97, 28_43, 5_95, 79_31, 18_23_96, 6_41_86, 22, 8_86, 5_95, 13_29_81, 53, 2_55_40, 34_49, 4_39_82, 3_99_01, 59_51, 8_78, 3_30, 4, 2_76_94, 8_02_69, 3_12, 53, 65_17, 1_17_80, 6_11, 2_04_08, 5], [2, 6, 13_25_97, 67, 4_28_97, 33, 5_92, 8, 16_37_29, 2_55_40, 3_61, 13_69_97, 10_95_14, 17_32_30, 7, 5_01, 60, 10_29_13, 1_96, 56_31, 2_35, 6_32_43, 4_73, 6, 23_17_57, 74, 52_77, 79_05, 53, 30_95, 3_73_17, 22, 4_54, 18_38_74, 5], [2, 2_68, 3_12_98, 4_65_30, 6, 13_29_35, 4_38_31, 7, 5_97, 32, 24, 36_88, 98_65, 5]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] } # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__magic_name__ , model_name="""facebook/xglm-564M""" , padding=__magic_name__ , )
26
class a : """simple docstring""" def __init__( self : Tuple , lowerCamelCase : list ) -> None: __snake_case : str = set_counts __snake_case : Union[str, Any] = max(lowerCamelCase ) __snake_case : List[Any] = len(lowerCamelCase ) __snake_case : Tuple = [1] * num_sets __snake_case : Dict = list(range(lowerCamelCase ) ) def __snake_case ( self : str , lowerCamelCase : int , lowerCamelCase : int ) -> bool: __snake_case : List[Any] = self.get_parent(lowerCamelCase ) __snake_case : Tuple = self.get_parent(lowerCamelCase ) if src_parent == dst_parent: return False if self.ranks[dst_parent] >= self.ranks[src_parent]: self.set_counts[dst_parent] += self.set_counts[src_parent] __snake_case : List[str] = 0 __snake_case : List[Any] = dst_parent if self.ranks[dst_parent] == self.ranks[src_parent]: self.ranks[dst_parent] += 1 __snake_case : Dict = self.set_counts[dst_parent] else: self.set_counts[src_parent] += self.set_counts[dst_parent] __snake_case : Union[str, Any] = 0 __snake_case : Optional[int] = src_parent __snake_case : Tuple = self.set_counts[src_parent] __snake_case : str = max(self.max_set , lowerCamelCase ) return True def __snake_case ( self : int , lowerCamelCase : int ) -> int: if self.parents[disj_set] == disj_set: return disj_set __snake_case : Optional[int] = self.get_parent(self.parents[disj_set] ) return self.parents[disj_set]
81
0
'''simple docstring''' import argparse import torch from transformers import BertForMaskedLM if __name__ == "__main__": UpperCAmelCase = argparse.ArgumentParser( description=( '''Extraction some layers of the full BertForMaskedLM or RObertaForMaskedLM for Transfer Learned''' ''' Distillation''' ) ) parser.add_argument('''--model_type''', default='''bert''', choices=['''bert''']) parser.add_argument('''--model_name''', default='''bert-base-uncased''', type=str) parser.add_argument('''--dump_checkpoint''', default='''serialization_dir/tf_bert-base-uncased_0247911.pth''', type=str) parser.add_argument('''--vocab_transform''', action='''store_true''') UpperCAmelCase = parser.parse_args() if args.model_type == "bert": UpperCAmelCase = BertForMaskedLM.from_pretrained(args.model_name) UpperCAmelCase = "bert" else: raise ValueError('''args.model_type should be \"bert\".''') UpperCAmelCase = model.state_dict() UpperCAmelCase = {} for w in ["word_embeddings", "position_embeddings"]: UpperCAmelCase = state_dict[F'''{prefix}.embeddings.{w}.weight'''] for w in ["weight", "bias"]: UpperCAmelCase = state_dict[F'''{prefix}.embeddings.LayerNorm.{w}'''] UpperCAmelCase = 0 for teacher_idx in [0, 2, 4, 7, 9, 11]: for w in ["weight", "bias"]: UpperCAmelCase = state_dict[ F'''{prefix}.encoder.layer.{teacher_idx}.attention.self.query.{w}''' ] UpperCAmelCase = state_dict[ F'''{prefix}.encoder.layer.{teacher_idx}.attention.self.key.{w}''' ] UpperCAmelCase = state_dict[ F'''{prefix}.encoder.layer.{teacher_idx}.attention.self.value.{w}''' ] UpperCAmelCase = state_dict[ F'''{prefix}.encoder.layer.{teacher_idx}.attention.output.dense.{w}''' ] UpperCAmelCase = state_dict[ F'''{prefix}.encoder.layer.{teacher_idx}.attention.output.LayerNorm.{w}''' ] UpperCAmelCase = state_dict[ F'''{prefix}.encoder.layer.{teacher_idx}.intermediate.dense.{w}''' ] UpperCAmelCase = state_dict[ F'''{prefix}.encoder.layer.{teacher_idx}.output.dense.{w}''' ] UpperCAmelCase = state_dict[ F'''{prefix}.encoder.layer.{teacher_idx}.output.LayerNorm.{w}''' ] std_idx += 1 UpperCAmelCase = state_dict["cls.predictions.decoder.weight"] UpperCAmelCase = state_dict["cls.predictions.bias"] if args.vocab_transform: for w in ["weight", "bias"]: UpperCAmelCase = state_dict[F'''cls.predictions.transform.dense.{w}'''] UpperCAmelCase = state_dict[F'''cls.predictions.transform.LayerNorm.{w}'''] print(F'''N layers selected for distillation: {std_idx}''') print(F'''Number of params transferred for distillation: {len(compressed_sd.keys())}''') print(F'''Save transferred checkpoint to {args.dump_checkpoint}.''') torch.save(compressed_sd, args.dump_checkpoint)
119
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class a (unittest.TestCase ): """simple docstring""" def __init__( self : int , lowerCamelCase : Any , lowerCamelCase : int=7 , lowerCamelCase : str=3 , lowerCamelCase : Optional[int]=30 , lowerCamelCase : Dict=400 , lowerCamelCase : str=True , lowerCamelCase : str=None , lowerCamelCase : Any=True , lowerCamelCase : Union[str, Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[str]=True , lowerCamelCase : Optional[int]=1 / 255 , lowerCamelCase : Any=True , ) -> str: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __snake_case : Optional[Any] = size if size is not None else {"shortest_edge": 18, "longest_edge": 1333} __snake_case : Optional[Any] = parent __snake_case : List[Any] = batch_size __snake_case : Optional[int] = num_channels __snake_case : str = min_resolution __snake_case : int = max_resolution __snake_case : int = do_resize __snake_case : Tuple = size __snake_case : Any = do_normalize __snake_case : int = image_mean __snake_case : Tuple = image_std __snake_case : Dict = do_rescale __snake_case : Optional[Any] = rescale_factor __snake_case : str = do_pad def __snake_case ( self : Any ) -> int: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def __snake_case ( self : Dict , lowerCamelCase : List[Any] , lowerCamelCase : Optional[Any]=False ) -> List[str]: if not batched: __snake_case : Dict = image_inputs[0] if isinstance(lowerCamelCase , Image.Image ): __snake_case , __snake_case : Dict = image.size else: __snake_case , __snake_case : List[str] = image.shape[1], image.shape[2] if w < h: __snake_case : Optional[int] = int(self.size["shortest_edge"] * h / w ) __snake_case : int = self.size["shortest_edge"] elif w > h: __snake_case : List[str] = self.size["shortest_edge"] __snake_case : Optional[Any] = int(self.size["shortest_edge"] * w / h ) else: __snake_case : List[Any] = self.size["shortest_edge"] __snake_case : Any = self.size["shortest_edge"] else: __snake_case : int = [] for image in image_inputs: __snake_case , __snake_case : List[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[0] )[0] __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class a (_lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ConditionalDetrImageProcessor if is_vision_available() else None def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : Optional[Any] = ConditionalDetrImageProcessingTester(self ) @property def __snake_case ( self : Any ) -> str: return self.image_processor_tester.prepare_image_processor_dict() def __snake_case ( self : Optional[Any] ) -> Optional[int]: __snake_case : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase , "image_mean" ) ) self.assertTrue(hasattr(lowerCamelCase , "image_std" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_normalize" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase , "size" ) ) def __snake_case ( self : Any ) -> Dict: __snake_case : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 1333} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) __snake_case : str = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=lowerCamelCase ) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> Dict: pass def __snake_case ( self : Tuple ) -> str: # Initialize image_processing __snake_case : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , Image.Image ) # Test not batched input __snake_case : int = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) __snake_case : Dict = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> str: # Initialize image_processing __snake_case : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __snake_case : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , numpify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , np.ndarray ) # Test not batched input __snake_case : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : List[Any] = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> List[str]: # Initialize image_processing __snake_case : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , torchify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , torch.Tensor ) # Test not batched input __snake_case : Optional[Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : List[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : int = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def __snake_case ( self : Any ) -> Optional[int]: # prepare image and target __snake_case : List[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : List[Any] = {"image_id": 39769, "annotations": target} # encode them __snake_case : List[str] = ConditionalDetrImageProcessor.from_pretrained("microsoft/conditional-detr-resnet-50" ) __snake_case : List[str] = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : Tuple = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Optional[int] = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : List[Any] = torch.tensor([58_87.96_00, 1_12_50.20_61, 48_93_53.84_38, 83_71_22.75_00, 14_79_67.51_56, 16_57_32.34_38] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : Optional[int] = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([0.55_03, 0.27_65, 0.06_04, 0.22_15] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Optional[Any] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : List[Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : Dict = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify orig_size __snake_case : int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Tuple = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) ) @slow def __snake_case ( self : str ) -> Tuple: # prepare image, target and masks_path __snake_case : Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : str = {"file_name": "000000039769.png", "image_id": 39769, "segments_info": target} __snake_case : Optional[int] = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them __snake_case : int = ConditionalDetrImageProcessor(format="coco_panoptic" ) __snake_case : str = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , masks_path=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : List[str] = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Dict = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : Any = torch.tensor([14_79_79.68_75, 16_55_27.04_69, 48_46_38.59_38, 1_12_92.93_75, 58_79.65_62, 76_34.11_47] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : str = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : Optional[Any] = torch.tensor([0.26_25, 0.54_37, 0.46_88, 0.86_25] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Tuple = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : Dict = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : int = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify masks __snake_case : List[Any] = 822873 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , lowerCamelCase ) # verify orig_size __snake_case : List[str] = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Any = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) )
81
0
class lowerCAmelCase__ : def __init__( self : List[str] ) -> List[str]: __lowerCamelCase = "" __lowerCamelCase = "" __lowerCamelCase = [] def __A ( self : int , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int ) -> int: if m == -1: return n + 1 elif n == -1: return m + 1 elif self.dp[m][n] > -1: return self.dp[m][n] else: if self.worda[m] == self.worda[n]: __lowerCamelCase = self.__min_dist_top_down_dp(m - 1 , n - 1 ) else: __lowerCamelCase = self.__min_dist_top_down_dp(SCREAMING_SNAKE_CASE__ , n - 1 ) __lowerCamelCase = self.__min_dist_top_down_dp(m - 1 , SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = self.__min_dist_top_down_dp(m - 1 , n - 1 ) __lowerCamelCase = 1 + min(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) return self.dp[m][n] def __A ( self : str , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : str ) -> int: __lowerCamelCase = worda __lowerCamelCase = worda __lowerCamelCase = [[-1 for _ in range(len(SCREAMING_SNAKE_CASE__ ) )] for _ in range(len(SCREAMING_SNAKE_CASE__ ) )] return self.__min_dist_top_down_dp(len(SCREAMING_SNAKE_CASE__ ) - 1 , len(SCREAMING_SNAKE_CASE__ ) - 1 ) def __A ( self : Optional[Any] , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : str ) -> int: __lowerCamelCase = worda __lowerCamelCase = worda __lowerCamelCase = len(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = len(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = [[0 for _ in range(n + 1 )] for _ in range(m + 1 )] for i in range(m + 1 ): for j in range(n + 1 ): if i == 0: # first string is empty __lowerCamelCase = j elif j == 0: # second string is empty __lowerCamelCase = i elif worda[i - 1] == worda[j - 1]: # last characters are equal __lowerCamelCase = self.dp[i - 1][j - 1] else: __lowerCamelCase = self.dp[i][j - 1] __lowerCamelCase = self.dp[i - 1][j] __lowerCamelCase = self.dp[i - 1][j - 1] __lowerCamelCase = 1 + min(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) return self.dp[m][n] if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : Union[str, Any] = EditDistance() print("****************** Testing Edit Distance DP Algorithm ******************") print() SCREAMING_SNAKE_CASE__ : Dict = input("Enter the first string: ").strip() SCREAMING_SNAKE_CASE__ : Any = input("Enter the second string: ").strip() print() print(F'The minimum edit distance is: {solver.min_dist_top_down(Sa, Sa)}') print(F'The minimum edit distance is: {solver.min_dist_bottom_up(Sa, Sa)}') print() print("*************** End of Testing Edit Distance DP Algorithm ***************")
298
from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from ... import AutoBackbone from ...modeling_outputs import SemanticSegmenterOutput from ...modeling_utils import PreTrainedModel from ...utils import add_start_docstrings, add_start_docstrings_to_model_forward, replace_return_docstrings from ...utils.backbone_utils import BackboneMixin from .configuration_upernet import UperNetConfig _snake_case : Optional[Any] = [ "openmmlab/upernet-convnext-tiny", # See all UperNet models at https://huggingface.co/models?filter=upernet ] # General docstring _snake_case : Dict = "UperNetConfig" class a (nn.Module ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : Union[int, Tuple[int, int]] , lowerCamelCase : Union[int, Tuple[int, int], str] = 0 , lowerCamelCase : bool = False , lowerCamelCase : Union[int, Tuple[int, int]] = 1 , ) -> None: super().__init__() __snake_case : Union[str, Any] = nn.Convad( in_channels=lowerCamelCase , out_channels=lowerCamelCase , kernel_size=lowerCamelCase , padding=lowerCamelCase , bias=lowerCamelCase , dilation=lowerCamelCase , ) __snake_case : Dict = nn.BatchNormad(lowerCamelCase ) __snake_case : List[Any] = nn.ReLU() def __snake_case ( self : List[Any] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : Dict = self.conv(lowerCamelCase ) __snake_case : int = self.batch_norm(lowerCamelCase ) __snake_case : Optional[Any] = self.activation(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : str , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : int ) -> None: super().__init__() __snake_case : Tuple = [ nn.AdaptiveAvgPoolad(lowerCamelCase ), UperNetConvModule(lowerCamelCase , lowerCamelCase , kernel_size=1 ), ] for i, layer in enumerate(self.layers ): self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : Dict , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : List[str] = input for layer in self.layers: __snake_case : Tuple = layer(lowerCamelCase ) return hidden_state class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Tuple[int, ...] , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : bool ) -> None: super().__init__() __snake_case : Dict = pool_scales __snake_case : List[str] = align_corners __snake_case : List[Any] = in_channels __snake_case : str = channels __snake_case : Optional[Any] = [] for i, pool_scale in enumerate(lowerCamelCase ): __snake_case : Union[str, Any] = UperNetPyramidPoolingBlock(pool_scale=lowerCamelCase , in_channels=lowerCamelCase , channels=lowerCamelCase ) self.blocks.append(lowerCamelCase ) self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> List[torch.Tensor]: __snake_case : Tuple = [] for ppm in self.blocks: __snake_case : Any = ppm(lowerCamelCase ) __snake_case : List[Any] = nn.functional.interpolate( lowerCamelCase , size=x.size()[2:] , mode="bilinear" , align_corners=self.align_corners ) ppm_outs.append(lowerCamelCase ) return ppm_outs class a (nn.Module ): """simple docstring""" def __init__( self : int , lowerCamelCase : str , lowerCamelCase : str ) -> List[str]: super().__init__() __snake_case : Dict = config __snake_case : List[str] = config.pool_scales # e.g. (1, 2, 3, 6) __snake_case : Tuple = in_channels __snake_case : str = config.hidden_size __snake_case : List[str] = False __snake_case : Any = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) # PSP Module __snake_case : Tuple = UperNetPyramidPoolingModule( self.pool_scales , self.in_channels[-1] , self.channels , align_corners=self.align_corners , ) __snake_case : List[str] = UperNetConvModule( self.in_channels[-1] + len(self.pool_scales ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) # FPN Module __snake_case : List[Any] = nn.ModuleList() __snake_case : Dict = nn.ModuleList() for in_channels in self.in_channels[:-1]: # skip the top layer __snake_case : Union[str, Any] = UperNetConvModule(lowerCamelCase , self.channels , kernel_size=1 ) __snake_case : Optional[int] = UperNetConvModule(self.channels , self.channels , kernel_size=3 , padding=1 ) self.lateral_convs.append(lowerCamelCase ) self.fpn_convs.append(lowerCamelCase ) __snake_case : int = UperNetConvModule( len(self.in_channels ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) def __snake_case ( self : List[str] ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> str: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : List[Any] , lowerCamelCase : Tuple ) -> Optional[int]: __snake_case : str = inputs[-1] __snake_case : int = [x] psp_outs.extend(self.psp_modules(lowerCamelCase ) ) __snake_case : Tuple = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Union[str, Any] = self.bottleneck(lowerCamelCase ) return output def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # build laterals __snake_case : Any = [lateral_conv(encoder_hidden_states[i] ) for i, lateral_conv in enumerate(self.lateral_convs )] laterals.append(self.psp_forward(lowerCamelCase ) ) # build top-down path __snake_case : Dict = len(lowerCamelCase ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Union[str, Any] = laterals[i - 1].shape[2:] __snake_case : Optional[int] = laterals[i - 1] + nn.functional.interpolate( laterals[i] , size=lowerCamelCase , mode="bilinear" , align_corners=self.align_corners ) # build outputs __snake_case : str = [self.fpn_convs[i](laterals[i] ) for i in range(used_backbone_levels - 1 )] # append psp feature fpn_outs.append(laterals[-1] ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Tuple = nn.functional.interpolate( fpn_outs[i] , size=fpn_outs[0].shape[2:] , mode="bilinear" , align_corners=self.align_corners ) __snake_case : str = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Optional[Any] = self.fpn_bottleneck(lowerCamelCase ) __snake_case : Tuple = self.classifier(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Optional[int] , lowerCamelCase : int = 2 , lowerCamelCase : int = 3 , lowerCamelCase : Union[int, Tuple[int, int]] = 1 ) -> None: super().__init__() __snake_case : List[Any] = config __snake_case : List[str] = config.auxiliary_in_channels __snake_case : List[Any] = config.auxiliary_channels __snake_case : Tuple = config.auxiliary_num_convs __snake_case : int = config.auxiliary_concat_input __snake_case : Optional[int] = in_index __snake_case : Tuple = (kernel_size // 2) * dilation __snake_case : Optional[int] = [] convs.append( UperNetConvModule( self.in_channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) for i in range(self.num_convs - 1 ): convs.append( UperNetConvModule( self.channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) if self.num_convs == 0: __snake_case : Union[str, Any] = nn.Identity() else: __snake_case : Any = nn.Sequential(*lowerCamelCase ) if self.concat_input: __snake_case : int = UperNetConvModule( self.in_channels + self.channels , self.channels , kernel_size=lowerCamelCase , padding=kernel_size // 2 ) __snake_case : Optional[Any] = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) def __snake_case ( self : Dict ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : Tuple ) -> Optional[int]: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : Optional[int] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # just take the relevant feature maps __snake_case : List[str] = encoder_hidden_states[self.in_index] __snake_case : Optional[Any] = self.convs(lowerCamelCase ) if self.concat_input: __snake_case : Optional[Any] = self.conv_cat(torch.cat([hidden_states, output] , dim=1 ) ) __snake_case : Union[str, Any] = self.classifier(lowerCamelCase ) return output class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = UperNetConfig __UpperCAmelCase : int = "pixel_values" __UpperCAmelCase : str = True def __snake_case ( self : Union[str, Any] , lowerCamelCase : List[Any] ) -> List[Any]: if isinstance(lowerCamelCase , lowerCamelCase ): module.backbone.init_weights() module.decode_head.init_weights() module.auxiliary_head.init_weights() def __snake_case ( self : Optional[Any] ) -> List[str]: self.backbone.init_weights() self.decode_head.init_weights() self.auxiliary_head.init_weights() def __snake_case ( self : List[Any] , lowerCamelCase : Any , lowerCamelCase : Optional[Any]=False ) -> Dict: if isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : Union[str, Any] = value _snake_case : Dict = R"\n Parameters:\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n config ([`UperNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" _snake_case : Tuple = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using\n [`AutoImageProcessor`]. See [`SegformerImageProcessor.__call__`] for details.\n output_attentions (`bool`, *optional*):\n Whether or not to return the attentions tensors of all attention layers in case the backbone has them. See\n `attentions` under returned tensors for more detail.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers of the backbone. See `hidden_states` under\n returned tensors for more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n" @add_start_docstrings( "UperNet framework leveraging any vision backbone e.g. for ADE20k, CityScapes." , _lowerCAmelCase , ) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Dict , lowerCamelCase : int ) -> Optional[int]: super().__init__(lowerCamelCase ) __snake_case : Any = AutoBackbone.from_config(config.backbone_config ) # Semantic segmentation head(s) __snake_case : Union[str, Any] = UperNetHead(lowerCamelCase , in_channels=self.backbone.channels ) __snake_case : Any = UperNetFCNHead(lowerCamelCase ) if config.use_auxiliary_head else None # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UPERNET_INPUTS_DOCSTRING.format("batch_size, sequence_length" ) ) @replace_return_docstrings(output_type=lowerCamelCase , config_class=_CONFIG_FOR_DOC ) def __snake_case ( self : Union[str, Any] , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , ) -> Union[tuple, SemanticSegmenterOutput]: __snake_case : Any = return_dict if return_dict is not None else self.config.use_return_dict __snake_case : int = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __snake_case : str = output_attentions if output_attentions is not None else self.config.output_attentions __snake_case : Tuple = self.backbone.forward_with_filtered_kwargs( lowerCamelCase , output_hidden_states=lowerCamelCase , output_attentions=lowerCamelCase ) __snake_case : List[Any] = outputs.feature_maps __snake_case : List[Any] = self.decode_head(lowerCamelCase ) __snake_case : List[str] = nn.functional.interpolate(lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : Optional[int] = None if self.auxiliary_head is not None: __snake_case : Dict = self.auxiliary_head(lowerCamelCase ) __snake_case : Dict = nn.functional.interpolate( lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : int = None if labels is not None: if self.config.num_labels == 1: raise ValueError("The number of labels should be greater than one" ) else: # compute weighted loss __snake_case : Any = CrossEntropyLoss(ignore_index=self.config.loss_ignore_index ) __snake_case : Union[str, Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = main_loss + self.config.auxiliary_loss_weight * auxiliary_loss if not return_dict: if output_hidden_states: __snake_case : Any = (logits,) + outputs[1:] else: __snake_case : str = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return SemanticSegmenterOutput( loss=lowerCamelCase , logits=lowerCamelCase , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
81
0
"""simple docstring""" import contextlib import csv import json import os import sqlitea import tarfile import textwrap import zipfile import pyarrow as pa import pyarrow.parquet as pq import pytest import datasets import datasets.config @pytest.fixture(scope='session' ) def lowercase__( ): lowercase_ : Optional[Any] = 10 lowercase_ : List[Any] = datasets.Features( { 'tokens': datasets.Sequence(datasets.Value('string' ) ), 'labels': datasets.Sequence(datasets.ClassLabel(names=['negative', 'positive'] ) ), 'answers': datasets.Sequence( { 'text': datasets.Value('string' ), 'answer_start': datasets.Value('int32' ), } ), 'id': datasets.Value('int64' ), } ) lowercase_ : Any = datasets.Dataset.from_dict( { 'tokens': [['foo'] * 5] * n, 'labels': [[1] * 5] * n, 'answers': [{'answer_start': [97], 'text': ['1976']}] * 10, 'id': list(range(__lowerCamelCase ) ), } , features=__lowerCamelCase , ) return dataset @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : Tuple ): lowercase_ : Optional[int] = str(tmp_path_factory.mktemp('data' ) / 'file.arrow' ) dataset.map(cache_file_name=__lowerCamelCase ) return filename # FILE_CONTENT + files __SCREAMING_SNAKE_CASE ="\\n Text data.\n Second line of data." @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Union[str, Any] ): lowercase_ : Tuple = tmp_path_factory.mktemp('data' ) / "file.txt" lowercase_ : Optional[Any] = FILE_CONTENT with open(__lowerCamelCase , 'w' ) as f: f.write(__lowerCamelCase ) return filename @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Union[str, Any] ): import bza lowercase_ : List[str] = tmp_path_factory.mktemp('data' ) / "file.txt.bz2" lowercase_ : Optional[Any] = bytes(__lowerCamelCase , 'utf-8' ) with bza.open(__lowerCamelCase , 'wb' ) as f: f.write(__lowerCamelCase ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : List[Any] ): import gzip lowercase_ : List[str] = str(tmp_path_factory.mktemp('data' ) / 'file.txt.gz' ) lowercase_ : List[str] = bytes(__lowerCamelCase , 'utf-8' ) with gzip.open(__lowerCamelCase , 'wb' ) as f: f.write(__lowerCamelCase ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Any ): if datasets.config.LZ4_AVAILABLE: import lza.frame lowercase_ : Any = tmp_path_factory.mktemp('data' ) / "file.txt.lz4" lowercase_ : str = bytes(__lowerCamelCase , 'utf-8' ) with lza.frame.open(__lowerCamelCase , 'wb' ) as f: f.write(__lowerCamelCase ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Optional[Any] , __SCREAMING_SNAKE_CASE : Optional[Any] ): if datasets.config.PY7ZR_AVAILABLE: import pyazr lowercase_ : int = tmp_path_factory.mktemp('data' ) / "file.txt.7z" with pyazr.SevenZipFile(__lowerCamelCase , 'w' ) as archive: archive.write(__lowerCamelCase , arcname=os.path.basename(__lowerCamelCase ) ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : List[Any] , __SCREAMING_SNAKE_CASE : Optional[int] ): import tarfile lowercase_ : int = tmp_path_factory.mktemp('data' ) / "file.txt.tar" with tarfile.TarFile(__lowerCamelCase , 'w' ) as f: f.add(__lowerCamelCase , arcname=os.path.basename(__lowerCamelCase ) ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Tuple ): import lzma lowercase_ : Optional[int] = tmp_path_factory.mktemp('data' ) / "file.txt.xz" lowercase_ : Any = bytes(__lowerCamelCase , 'utf-8' ) with lzma.open(__lowerCamelCase , 'wb' ) as f: f.write(__lowerCamelCase ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : str ): import zipfile lowercase_ : Optional[int] = tmp_path_factory.mktemp('data' ) / "file.txt.zip" with zipfile.ZipFile(__lowerCamelCase , 'w' ) as f: f.write(__lowerCamelCase , arcname=os.path.basename(__lowerCamelCase ) ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Any ): if datasets.config.ZSTANDARD_AVAILABLE: import zstandard as zstd lowercase_ : Optional[int] = tmp_path_factory.mktemp('data' ) / "file.txt.zst" lowercase_ : Optional[Any] = bytes(__lowerCamelCase , 'utf-8' ) with zstd.open(__lowerCamelCase , 'wb' ) as f: f.write(__lowerCamelCase ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Optional[int] ): lowercase_ : int = tmp_path_factory.mktemp('data' ) / "file.xml" lowercase_ : Dict = textwrap.dedent( '\\n <?xml version=\"1.0\" encoding=\"UTF-8\" ?>\n <tmx version=\"1.4\">\n <header segtype=\"sentence\" srclang=\"ca\" />\n <body>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 1</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 1</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 2</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 2</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 3</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 3</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 4</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 4</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 5</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 5</seg></tuv>\n </tu>\n </body>\n </tmx>' ) with open(__lowerCamelCase , 'w' ) as f: f.write(__lowerCamelCase ) return filename __SCREAMING_SNAKE_CASE =[ {"col_1": "0", "col_2": 0, "col_3": 0.0}, {"col_1": "1", "col_2": 1, "col_3": 1.0}, {"col_1": "2", "col_2": 2, "col_3": 2.0}, {"col_1": "3", "col_2": 3, "col_3": 3.0}, ] __SCREAMING_SNAKE_CASE =[ {"col_1": "4", "col_2": 4, "col_3": 4.0}, {"col_1": "5", "col_2": 5, "col_3": 5.0}, ] __SCREAMING_SNAKE_CASE ={ "col_1": ["0", "1", "2", "3"], "col_2": [0, 1, 2, 3], "col_3": [0.0, 1.0, 2.0, 3.0], } __SCREAMING_SNAKE_CASE =[ {"col_3": 0.0, "col_1": "0", "col_2": 0}, {"col_3": 1.0, "col_1": "1", "col_2": 1}, ] __SCREAMING_SNAKE_CASE =[ {"col_1": "s0", "col_2": 0, "col_3": 0.0}, {"col_1": "s1", "col_2": 1, "col_3": 1.0}, {"col_1": "s2", "col_2": 2, "col_3": 2.0}, {"col_1": "s3", "col_2": 3, "col_3": 3.0}, ] @pytest.fixture(scope='session' ) def lowercase__( ): return DATA_DICT_OF_LISTS @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Optional[Any] ): lowercase_ : Dict = datasets.Dataset.from_dict(__lowerCamelCase ) lowercase_ : int = str(tmp_path_factory.mktemp('data' ) / 'dataset.arrow' ) dataset.map(cache_file_name=__lowerCamelCase ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Dict ): lowercase_ : Tuple = str(tmp_path_factory.mktemp('data' ) / 'dataset.sqlite' ) with contextlib.closing(sqlitea.connect(__lowerCamelCase ) ) as con: lowercase_ : Dict = con.cursor() cur.execute('CREATE TABLE dataset(col_1 text, col_2 int, col_3 real)' ) for item in DATA: cur.execute('INSERT INTO dataset(col_1, col_2, col_3) VALUES (?, ?, ?)' , tuple(item.values() ) ) con.commit() return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Union[str, Any] ): lowercase_ : Any = str(tmp_path_factory.mktemp('data' ) / 'dataset.csv' ) with open(__lowerCamelCase , 'w' , newline='' ) as f: lowercase_ : Any = csv.DictWriter(__lowerCamelCase , fieldnames=['col_1', 'col_2', 'col_3'] ) writer.writeheader() for item in DATA: writer.writerow(__lowerCamelCase ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : List[Any] ): lowercase_ : Optional[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset2.csv' ) with open(__lowerCamelCase , 'w' , newline='' ) as f: lowercase_ : str = csv.DictWriter(__lowerCamelCase , fieldnames=['col_1', 'col_2', 'col_3'] ) writer.writeheader() for item in DATA: writer.writerow(__lowerCamelCase ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : List[str] ): import bza lowercase_ : List[str] = tmp_path_factory.mktemp('data' ) / "dataset.csv.bz2" with open(__lowerCamelCase , 'rb' ) as f: lowercase_ : Tuple = f.read() # data = bytes(FILE_CONTENT, "utf-8") with bza.open(__lowerCamelCase , 'wb' ) as f: f.write(__lowerCamelCase ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Any , __SCREAMING_SNAKE_CASE : Optional[int] , __SCREAMING_SNAKE_CASE : Optional[Any] ): lowercase_ : List[str] = tmp_path_factory.mktemp('data' ) / "dataset.csv.zip" with zipfile.ZipFile(__lowerCamelCase , 'w' ) as f: f.write(__lowerCamelCase , arcname=os.path.basename(__lowerCamelCase ) ) f.write(__lowerCamelCase , arcname=os.path.basename(__lowerCamelCase ) ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : Optional[Any] , __SCREAMING_SNAKE_CASE : Any ): lowercase_ : Any = tmp_path_factory.mktemp('data' ) / "dataset.csv.zip" with zipfile.ZipFile(__lowerCamelCase , 'w' ) as f: f.write(__lowerCamelCase , arcname=os.path.basename(csv_path.replace('.csv' , '.CSV' ) ) ) f.write(__lowerCamelCase , arcname=os.path.basename(csva_path.replace('.csv' , '.CSV' ) ) ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : Any , __SCREAMING_SNAKE_CASE : str ): lowercase_ : List[Any] = tmp_path_factory.mktemp('data' ) / "dataset_with_dir.csv.zip" with zipfile.ZipFile(__lowerCamelCase , 'w' ) as f: f.write(__lowerCamelCase , arcname=os.path.join('main_dir' , os.path.basename(__lowerCamelCase ) ) ) f.write(__lowerCamelCase , arcname=os.path.join('main_dir' , os.path.basename(__lowerCamelCase ) ) ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : List[Any] ): lowercase_ : Optional[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset.parquet' ) lowercase_ : Tuple = pa.schema( { 'col_1': pa.string(), 'col_2': pa.intaa(), 'col_3': pa.floataa(), } ) with open(__lowerCamelCase , 'wb' ) as f: lowercase_ : List[str] = pq.ParquetWriter(__lowerCamelCase , schema=__lowerCamelCase ) lowercase_ : Any = pa.Table.from_pydict({k: [DATA[i][k] for i in range(len(__lowerCamelCase ) )] for k in DATA[0]} , schema=__lowerCamelCase ) writer.write_table(__lowerCamelCase ) writer.close() return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : List[str] ): lowercase_ : Dict = str(tmp_path_factory.mktemp('data' ) / 'dataset.json' ) lowercase_ : str = {"data": DATA} with open(__lowerCamelCase , 'w' ) as f: json.dump(__lowerCamelCase , __lowerCamelCase ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : List[Any] ): lowercase_ : List[str] = str(tmp_path_factory.mktemp('data' ) / 'dataset.json' ) lowercase_ : Tuple = {"data": DATA_DICT_OF_LISTS} with open(__lowerCamelCase , 'w' ) as f: json.dump(__lowerCamelCase , __lowerCamelCase ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Dict ): lowercase_ : List[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset.jsonl' ) with open(__lowerCamelCase , 'w' ) as f: for item in DATA: f.write(json.dumps(__lowerCamelCase ) + '\n' ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Tuple ): lowercase_ : List[str] = str(tmp_path_factory.mktemp('data' ) / 'dataset2.jsonl' ) with open(__lowerCamelCase , 'w' ) as f: for item in DATA: f.write(json.dumps(__lowerCamelCase ) + '\n' ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Optional[Any] ): lowercase_ : List[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset_312.jsonl' ) with open(__lowerCamelCase , 'w' ) as f: for item in DATA_312: f.write(json.dumps(__lowerCamelCase ) + '\n' ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : int ): lowercase_ : List[str] = str(tmp_path_factory.mktemp('data' ) / 'dataset-str.jsonl' ) with open(__lowerCamelCase , 'w' ) as f: for item in DATA_STR: f.write(json.dumps(__lowerCamelCase ) + '\n' ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : Union[str, Any] ): import gzip lowercase_ : int = str(tmp_path_factory.mktemp('data' ) / 'dataset.txt.gz' ) with open(__lowerCamelCase , 'rb' ) as orig_file: with gzip.open(__lowerCamelCase , 'wb' ) as zipped_file: zipped_file.writelines(__lowerCamelCase ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : List[Any] , __SCREAMING_SNAKE_CASE : Optional[Any] ): import gzip lowercase_ : Optional[int] = str(tmp_path_factory.mktemp('data' ) / 'dataset.jsonl.gz' ) with open(__lowerCamelCase , 'rb' ) as orig_file: with gzip.open(__lowerCamelCase , 'wb' ) as zipped_file: zipped_file.writelines(__lowerCamelCase ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : Any ): lowercase_ : Any = tmp_path_factory.mktemp('data' ) / "dataset.jsonl.zip" with zipfile.ZipFile(__lowerCamelCase , 'w' ) as f: f.write(__lowerCamelCase , arcname=os.path.basename(__lowerCamelCase ) ) f.write(__lowerCamelCase , arcname=os.path.basename(__lowerCamelCase ) ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Optional[int] , __SCREAMING_SNAKE_CASE : List[Any] , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : Any ): lowercase_ : List[str] = tmp_path_factory.mktemp('data' ) / "dataset_nested.jsonl.zip" with zipfile.ZipFile(__lowerCamelCase , 'w' ) as f: f.write(__lowerCamelCase , arcname=os.path.join('nested' , os.path.basename(__lowerCamelCase ) ) ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : Any , __SCREAMING_SNAKE_CASE : str ): lowercase_ : Optional[int] = tmp_path_factory.mktemp('data' ) / "dataset_with_dir.jsonl.zip" with zipfile.ZipFile(__lowerCamelCase , 'w' ) as f: f.write(__lowerCamelCase , arcname=os.path.join('main_dir' , os.path.basename(__lowerCamelCase ) ) ) f.write(__lowerCamelCase , arcname=os.path.join('main_dir' , os.path.basename(__lowerCamelCase ) ) ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : str ): lowercase_ : Optional[int] = tmp_path_factory.mktemp('data' ) / "dataset.jsonl.tar" with tarfile.TarFile(__lowerCamelCase , 'w' ) as f: f.add(__lowerCamelCase , arcname=os.path.basename(__lowerCamelCase ) ) f.add(__lowerCamelCase , arcname=os.path.basename(__lowerCamelCase ) ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Any , __SCREAMING_SNAKE_CASE : Any , __SCREAMING_SNAKE_CASE : List[str] , __SCREAMING_SNAKE_CASE : List[Any] ): lowercase_ : Optional[Any] = tmp_path_factory.mktemp('data' ) / "dataset_nested.jsonl.tar" with tarfile.TarFile(__lowerCamelCase , 'w' ) as f: f.add(__lowerCamelCase , arcname=os.path.join('nested' , os.path.basename(__lowerCamelCase ) ) ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Union[str, Any] ): lowercase_ : List[Any] = ["0", "1", "2", "3"] lowercase_ : Optional[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset.txt' ) with open(__lowerCamelCase , 'w' ) as f: for item in data: f.write(item + '\n' ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : int ): lowercase_ : List[str] = ["0", "1", "2", "3"] lowercase_ : Any = str(tmp_path_factory.mktemp('data' ) / 'dataset2.txt' ) with open(__lowerCamelCase , 'w' ) as f: for item in data: f.write(item + '\n' ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Dict ): lowercase_ : Tuple = ["0", "1", "2", "3"] lowercase_ : Any = tmp_path_factory.mktemp('data' ) / "dataset.abc" with open(__lowerCamelCase , 'w' ) as f: for item in data: f.write(item + '\n' ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : Dict , __SCREAMING_SNAKE_CASE : str ): lowercase_ : Optional[int] = tmp_path_factory.mktemp('data' ) / "dataset.text.zip" with zipfile.ZipFile(__lowerCamelCase , 'w' ) as f: f.write(__lowerCamelCase , arcname=os.path.basename(__lowerCamelCase ) ) f.write(__lowerCamelCase , arcname=os.path.basename(__lowerCamelCase ) ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Optional[int] , __SCREAMING_SNAKE_CASE : List[str] , __SCREAMING_SNAKE_CASE : List[Any] ): lowercase_ : str = tmp_path_factory.mktemp('data' ) / "dataset_with_dir.text.zip" with zipfile.ZipFile(__lowerCamelCase , 'w' ) as f: f.write(__lowerCamelCase , arcname=os.path.join('main_dir' , os.path.basename(__lowerCamelCase ) ) ) f.write(__lowerCamelCase , arcname=os.path.join('main_dir' , os.path.basename(__lowerCamelCase ) ) ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : str ): lowercase_ : Optional[int] = tmp_path_factory.mktemp('data' ) / "dataset.ext.zip" with zipfile.ZipFile(__lowerCamelCase , 'w' ) as f: f.write(__lowerCamelCase , arcname=os.path.basename('unsupported.ext' ) ) f.write(__lowerCamelCase , arcname=os.path.basename('unsupported_2.ext' ) ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Any ): lowercase_ : Any = "\n".join(['First', 'Second\u2029with Unicode new line', 'Third'] ) lowercase_ : List[str] = str(tmp_path_factory.mktemp('data' ) / 'dataset_with_unicode_new_lines.txt' ) with open(__lowerCamelCase , 'w' , encoding='utf-8' ) as f: f.write(__lowerCamelCase ) return path @pytest.fixture(scope='session' ) def lowercase__( ): return os.path.join('tests' , 'features' , 'data' , 'test_image_rgb.jpg' ) @pytest.fixture(scope='session' ) def lowercase__( ): return os.path.join('tests' , 'features' , 'data' , 'test_audio_44100.wav' ) @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : Dict ): lowercase_ : Tuple = tmp_path_factory.mktemp('data' ) / "dataset.img.zip" with zipfile.ZipFile(__lowerCamelCase , 'w' ) as f: f.write(__lowerCamelCase , arcname=os.path.basename(__lowerCamelCase ) ) f.write(__lowerCamelCase , arcname=os.path.basename(__lowerCamelCase ).replace('.jpg' , '2.jpg' ) ) return path @pytest.fixture(scope='session' ) def lowercase__( __SCREAMING_SNAKE_CASE : List[Any] ): lowercase_ : Any = tmp_path_factory.mktemp('data_dir' ) (data_dir / "subdir").mkdir() with open(data_dir / 'subdir' / 'train.txt' , 'w' ) as f: f.write('foo\n' * 10 ) with open(data_dir / 'subdir' / 'test.txt' , 'w' ) as f: f.write('bar\n' * 10 ) # hidden file with open(data_dir / 'subdir' / '.test.txt' , 'w' ) as f: f.write('bar\n' * 10 ) # hidden directory (data_dir / ".subdir").mkdir() with open(data_dir / '.subdir' / 'train.txt' , 'w' ) as f: f.write('foo\n' * 10 ) with open(data_dir / '.subdir' / 'test.txt' , 'w' ) as f: f.write('bar\n' * 10 ) return data_dir
425
from __future__ import annotations def lowerCAmelCase_ ( __lowerCamelCase ): if not nums: return 0 __snake_case : Any = nums[0] __snake_case : str = 0 for num in nums[1:]: __snake_case , __snake_case : List[str] = ( max_excluding + num, max(__lowerCamelCase , __lowerCamelCase ), ) return max(__lowerCamelCase , __lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
81
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ : Tuple = logging.get_logger(__name__) snake_case_ : List[Any] = { "microsoft/cvt-13": "https://huggingface.co/microsoft/cvt-13/resolve/main/config.json", # See all Cvt models at https://huggingface.co/models?filter=cvt } class lowercase__ ( _lowerCAmelCase ): '''simple docstring''' _snake_case = "cvt" def __init__( self , lowerCamelCase__=3 , lowerCamelCase__=[7, 3, 3] , lowerCamelCase__=[4, 2, 2] , lowerCamelCase__=[2, 1, 1] , lowerCamelCase__=[6_4, 1_9_2, 3_8_4] , lowerCamelCase__=[1, 3, 6] , lowerCamelCase__=[1, 2, 1_0] , lowerCamelCase__=[4.0, 4.0, 4.0] , lowerCamelCase__=[0.0, 0.0, 0.0] , lowerCamelCase__=[0.0, 0.0, 0.0] , lowerCamelCase__=[0.0, 0.0, 0.1] , lowerCamelCase__=[True, True, True] , lowerCamelCase__=[False, False, True] , lowerCamelCase__=["dw_bn", "dw_bn", "dw_bn"] , lowerCamelCase__=[3, 3, 3] , lowerCamelCase__=[1, 1, 1] , lowerCamelCase__=[2, 2, 2] , lowerCamelCase__=[1, 1, 1] , lowerCamelCase__=[1, 1, 1] , lowerCamelCase__=0.02 , lowerCamelCase__=1e-1_2 , **lowerCamelCase__ , ): '''simple docstring''' super().__init__(**lowerCamelCase__ ) UpperCamelCase = num_channels UpperCamelCase = patch_sizes UpperCamelCase = patch_stride UpperCamelCase = patch_padding UpperCamelCase = embed_dim UpperCamelCase = num_heads UpperCamelCase = depth UpperCamelCase = mlp_ratio UpperCamelCase = attention_drop_rate UpperCamelCase = drop_rate UpperCamelCase = drop_path_rate UpperCamelCase = qkv_bias UpperCamelCase = cls_token UpperCamelCase = qkv_projection_method UpperCamelCase = kernel_qkv UpperCamelCase = padding_kv UpperCamelCase = stride_kv UpperCamelCase = padding_q UpperCamelCase = stride_q UpperCamelCase = initializer_range UpperCamelCase = layer_norm_eps
212
from __future__ import annotations from typing import Any def lowerCAmelCase_ ( __lowerCamelCase ): create_state_space_tree(__lowerCamelCase , [] , 0 ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): if index == len(__lowerCamelCase ): print(__lowerCamelCase ) return create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.pop() if __name__ == "__main__": _snake_case : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["A", "B", "C"]) generate_all_subsequences(seq)
81
0
import unittest from transformers import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING, is_vision_available, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class __magic_name__ : """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( *snake_case :List[str] , **snake_case :Optional[int] ): '''simple docstring''' pass @is_pipeline_test @require_vision @require_torch class __magic_name__ ( unittest.TestCase ): """simple docstring""" __UpperCamelCase = MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING def SCREAMING_SNAKE_CASE ( self :Optional[Any] , snake_case :List[Any] , snake_case :Optional[int] , snake_case :Any ): '''simple docstring''' A_ : List[str] = pipeline( "zero-shot-object-detection" , model="hf-internal-testing/tiny-random-owlvit-object-detection" ) A_ : Dict = [ { "image": "./tests/fixtures/tests_samples/COCO/000000039769.png", "candidate_labels": ["cat", "remote", "couch"], } ] return object_detector, examples def SCREAMING_SNAKE_CASE ( self :Tuple , snake_case :Tuple , snake_case :Optional[Any] ): '''simple docstring''' A_ : Dict = object_detector(examples[0] , threshold=0.0 ) A_ : Optional[int] = len(snake_case ) self.assertGreater(snake_case , 0 ) self.assertEqual( snake_case , [ { "score": ANY(snake_case ), "label": ANY(snake_case ), "box": {"xmin": ANY(snake_case ), "ymin": ANY(snake_case ), "xmax": ANY(snake_case ), "ymax": ANY(snake_case )}, } for i in range(snake_case ) ] , ) @require_tf @unittest.skip("Zero Shot Object Detection not implemented in TF" ) def SCREAMING_SNAKE_CASE ( self :Optional[int] ): '''simple docstring''' pass @require_torch def SCREAMING_SNAKE_CASE ( self :Any ): '''simple docstring''' A_ : List[Any] = pipeline( "zero-shot-object-detection" , model="hf-internal-testing/tiny-random-owlvit-object-detection" ) A_ : Optional[Any] = object_detector( "./tests/fixtures/tests_samples/COCO/000000039769.png" , candidate_labels=["cat", "remote", "couch"] , threshold=0.64 , ) self.assertEqual( nested_simplify(snake_case , decimals=4 ) , [ {"score": 0.7235, "label": "cat", "box": {"xmin": 204, "ymin": 167, "xmax": 232, "ymax": 190}}, {"score": 0.7218, "label": "remote", "box": {"xmin": 204, "ymin": 167, "xmax": 232, "ymax": 190}}, {"score": 0.7184, "label": "couch", "box": {"xmin": 204, "ymin": 167, "xmax": 232, "ymax": 190}}, {"score": 0.6748, "label": "remote", "box": {"xmin": 571, "ymin": 83, "xmax": 598, "ymax": 103}}, {"score": 0.6656, "label": "cat", "box": {"xmin": 571, "ymin": 83, "xmax": 598, "ymax": 103}}, {"score": 0.6614, "label": "couch", "box": {"xmin": 571, "ymin": 83, "xmax": 598, "ymax": 103}}, {"score": 0.6456, "label": "remote", "box": {"xmin": 494, "ymin": 105, "xmax": 521, "ymax": 127}}, {"score": 0.642, "label": "remote", "box": {"xmin": 67, "ymin": 274, "xmax": 93, "ymax": 297}}, {"score": 0.6419, "label": "cat", "box": {"xmin": 494, "ymin": 105, "xmax": 521, "ymax": 127}}, ] , ) A_ : str = object_detector( [ { "image": "./tests/fixtures/tests_samples/COCO/000000039769.png", "candidate_labels": ["cat", "remote", "couch"], } ] , threshold=0.64 , ) self.assertEqual( nested_simplify(snake_case , decimals=4 ) , [ [ {"score": 0.7235, "label": "cat", "box": {"xmin": 204, "ymin": 167, "xmax": 232, "ymax": 190}}, {"score": 0.7218, "label": "remote", "box": {"xmin": 204, "ymin": 167, "xmax": 232, "ymax": 190}}, {"score": 0.7184, "label": "couch", "box": {"xmin": 204, "ymin": 167, "xmax": 232, "ymax": 190}}, {"score": 0.6748, "label": "remote", "box": {"xmin": 571, "ymin": 83, "xmax": 598, "ymax": 103}}, {"score": 0.6656, "label": "cat", "box": {"xmin": 571, "ymin": 83, "xmax": 598, "ymax": 103}}, {"score": 0.6614, "label": "couch", "box": {"xmin": 571, "ymin": 83, "xmax": 598, "ymax": 103}}, {"score": 0.6456, "label": "remote", "box": {"xmin": 494, "ymin": 105, "xmax": 521, "ymax": 127}}, {"score": 0.642, "label": "remote", "box": {"xmin": 67, "ymin": 274, "xmax": 93, "ymax": 297}}, {"score": 0.6419, "label": "cat", "box": {"xmin": 494, "ymin": 105, "xmax": 521, "ymax": 127}}, ] ] , ) @require_torch @slow def SCREAMING_SNAKE_CASE ( self :Union[str, Any] ): '''simple docstring''' A_ : int = pipeline("zero-shot-object-detection" ) A_ : Optional[Any] = object_detector( "http://images.cocodataset.org/val2017/000000039769.jpg" , candidate_labels=["cat", "remote", "couch"] , ) self.assertEqual( nested_simplify(snake_case , decimals=4 ) , [ {"score": 0.2868, "label": "cat", "box": {"xmin": 324, "ymin": 20, "xmax": 640, "ymax": 373}}, {"score": 0.277, "label": "remote", "box": {"xmin": 40, "ymin": 72, "xmax": 177, "ymax": 115}}, {"score": 0.2537, "label": "cat", "box": {"xmin": 1, "ymin": 55, "xmax": 315, "ymax": 472}}, {"score": 0.1474, "label": "remote", "box": {"xmin": 335, "ymin": 74, "xmax": 371, "ymax": 187}}, {"score": 0.1208, "label": "couch", "box": {"xmin": 4, "ymin": 0, "xmax": 642, "ymax": 476}}, ] , ) A_ : Optional[int] = object_detector( [ { "image": "http://images.cocodataset.org/val2017/000000039769.jpg", "candidate_labels": ["cat", "remote", "couch"], }, { "image": "http://images.cocodataset.org/val2017/000000039769.jpg", "candidate_labels": ["cat", "remote", "couch"], }, ] , ) self.assertEqual( nested_simplify(snake_case , decimals=4 ) , [ [ {"score": 0.2868, "label": "cat", "box": {"xmin": 324, "ymin": 20, "xmax": 640, "ymax": 373}}, {"score": 0.277, "label": "remote", "box": {"xmin": 40, "ymin": 72, "xmax": 177, "ymax": 115}}, {"score": 0.2537, "label": "cat", "box": {"xmin": 1, "ymin": 55, "xmax": 315, "ymax": 472}}, {"score": 0.1474, "label": "remote", "box": {"xmin": 335, "ymin": 74, "xmax": 371, "ymax": 187}}, {"score": 0.1208, "label": "couch", "box": {"xmin": 4, "ymin": 0, "xmax": 642, "ymax": 476}}, ], [ {"score": 0.2868, "label": "cat", "box": {"xmin": 324, "ymin": 20, "xmax": 640, "ymax": 373}}, {"score": 0.277, "label": "remote", "box": {"xmin": 40, "ymin": 72, "xmax": 177, "ymax": 115}}, {"score": 0.2537, "label": "cat", "box": {"xmin": 1, "ymin": 55, "xmax": 315, "ymax": 472}}, {"score": 0.1474, "label": "remote", "box": {"xmin": 335, "ymin": 74, "xmax": 371, "ymax": 187}}, {"score": 0.1208, "label": "couch", "box": {"xmin": 4, "ymin": 0, "xmax": 642, "ymax": 476}}, ], ] , ) @require_tf @unittest.skip("Zero Shot Object Detection not implemented in TF" ) def SCREAMING_SNAKE_CASE ( self :Optional[Any] ): '''simple docstring''' pass @require_torch @slow def SCREAMING_SNAKE_CASE ( self :Optional[int] ): '''simple docstring''' A_ : List[Any] = 0.2 A_ : Any = pipeline("zero-shot-object-detection" ) A_ : Optional[Any] = object_detector( "http://images.cocodataset.org/val2017/000000039769.jpg" , candidate_labels=["cat", "remote", "couch"] , threshold=snake_case , ) self.assertEqual( nested_simplify(snake_case , decimals=4 ) , [ {"score": 0.2868, "label": "cat", "box": {"xmin": 324, "ymin": 20, "xmax": 640, "ymax": 373}}, {"score": 0.277, "label": "remote", "box": {"xmin": 40, "ymin": 72, "xmax": 177, "ymax": 115}}, {"score": 0.2537, "label": "cat", "box": {"xmin": 1, "ymin": 55, "xmax": 315, "ymax": 472}}, ] , ) @require_torch @slow def SCREAMING_SNAKE_CASE ( self :int ): '''simple docstring''' A_ : str = 2 A_ : str = pipeline("zero-shot-object-detection" ) A_ : int = object_detector( "http://images.cocodataset.org/val2017/000000039769.jpg" , candidate_labels=["cat", "remote", "couch"] , top_k=snake_case , ) self.assertEqual( nested_simplify(snake_case , decimals=4 ) , [ {"score": 0.2868, "label": "cat", "box": {"xmin": 324, "ymin": 20, "xmax": 640, "ymax": 373}}, {"score": 0.277, "label": "remote", "box": {"xmin": 40, "ymin": 72, "xmax": 177, "ymax": 115}}, ] , )
454
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case , __snake_case : List[Any] = image.size __snake_case , __snake_case : Tuple = (x - x % 3_2 for x in (w, h)) # resize to integer multiple of 32 __snake_case : str = image.resize((w, h) , resample=PIL_INTERPOLATION["lanczos"] ) __snake_case : int = np.array(__lowerCamelCase ).astype(np.floataa ) / 2_5_5.0 __snake_case : Union[str, Any] = image[None].transpose(0 , 3 , 1 , 2 ) __snake_case : Union[str, Any] = torch.from_numpy(__lowerCamelCase ) return 2.0 * image - 1.0 class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : VQModel , lowerCamelCase : UNetaDModel , lowerCamelCase : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ) -> Union[str, Any]: super().__init__() self.register_modules(vqvae=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase ) @torch.no_grad() def __call__( self : List[str] , lowerCamelCase : Union[torch.Tensor, PIL.Image.Image] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : Optional[int] = 100 , lowerCamelCase : Optional[float] = 0.0 , lowerCamelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , ) -> Union[Tuple, ImagePipelineOutput]: if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : Any = 1 elif isinstance(lowerCamelCase , torch.Tensor ): __snake_case : Any = image.shape[0] else: raise ValueError(F'`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(lowerCamelCase )}' ) if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : List[Any] = preprocess(lowerCamelCase ) __snake_case , __snake_case : int = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image __snake_case : str = (batch_size, self.unet.config.in_channels // 2, height, width) __snake_case : str = next(self.unet.parameters() ).dtype __snake_case : Tuple = randn_tensor(lowerCamelCase , generator=lowerCamelCase , device=self.device , dtype=lowerCamelCase ) __snake_case : List[Any] = image.to(device=self.device , dtype=lowerCamelCase ) # set timesteps and move to the correct device self.scheduler.set_timesteps(lowerCamelCase , device=self.device ) __snake_case : str = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler __snake_case : Dict = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] __snake_case : Union[str, Any] = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __snake_case : int = {} if accepts_eta: __snake_case : List[str] = eta for t in self.progress_bar(lowerCamelCase ): # concat latents and low resolution image in the channel dimension. __snake_case : Union[str, Any] = torch.cat([latents, image] , dim=1 ) __snake_case : Optional[Any] = self.scheduler.scale_model_input(lowerCamelCase , lowerCamelCase ) # predict the noise residual __snake_case : int = self.unet(lowerCamelCase , lowerCamelCase ).sample # compute the previous noisy sample x_t -> x_t-1 __snake_case : Union[str, Any] = self.scheduler.step(lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase ).prev_sample # decode the image latents with the VQVAE __snake_case : List[Any] = self.vqvae.decode(lowerCamelCase ).sample __snake_case : Dict = torch.clamp(lowerCamelCase , -1.0 , 1.0 ) __snake_case : Any = image / 2 + 0.5 __snake_case : Tuple = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __snake_case : Tuple = self.numpy_to_pil(lowerCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCamelCase )
81
0
"""simple docstring""" import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import MaMaaaTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from transformers.utils import is_sentencepiece_available if is_sentencepiece_available(): from transformers.models.mam_aaa.tokenization_mam_aaa import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin if is_sentencepiece_available(): lowerCamelCase__ = get_tests_dir("fixtures/test_sentencepiece.model") if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right lowerCamelCase__ = 128_022 lowerCamelCase__ = 128_028 @require_sentencepiece class _UpperCamelCase ( _lowerCAmelCase , unittest.TestCase): __lowerCamelCase = MaMaaaTokenizer __lowerCamelCase = False __lowerCamelCase = False __lowerCamelCase = True def A (self ): """simple docstring""" super().setUp() A__ = ["</s>", "<unk>", "▁This", "▁is", "▁a", "▁t", "est", "\u0120", "<pad>"] A__ = dict(zip(lowerCamelCase__ , range(len(lowerCamelCase__ ) ) ) ) A__ = Path(self.tmpdirname ) save_json(lowerCamelCase__ , save_dir / VOCAB_FILES_NAMES["""vocab_file"""] ) if not (save_dir / VOCAB_FILES_NAMES["spm_file"]).exists(): copyfile(lowerCamelCase__ , save_dir / VOCAB_FILES_NAMES["""spm_file"""] ) A__ = MaMaaaTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def A (self , **lowerCamelCase__ ): """simple docstring""" return MaMaaaTokenizer.from_pretrained(self.tmpdirname , **lowerCamelCase__ ) def A (self , lowerCamelCase__ ): """simple docstring""" return ( "This is a test", "This is a test", ) def A (self ): """simple docstring""" A__ = "</s>" A__ = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase__ ) , lowerCamelCase__ ) def A (self ): """simple docstring""" A__ = self.get_tokenizer() A__ = list(tokenizer.get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """</s>""" ) self.assertEqual(vocab_keys[1] , """<unk>""" ) self.assertEqual(vocab_keys[-1] , """<s>""" ) self.assertEqual(len(lowerCamelCase__ ) , tokenizer.vocab_size + len(tokenizer.get_added_vocab() ) ) @unittest.skip("""Skip this test while all models are still to be uploaded.""" ) def A (self ): """simple docstring""" pass def A (self ): """simple docstring""" A__ = self.get_tokenizer() A__ = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(lowerCamelCase__ , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) , [2, 3, 4, 5, 6] , ) A__ = tokenizer.convert_ids_to_tokens([2, 3, 4, 5, 6] ) self.assertListEqual(lowerCamelCase__ , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) A__ = tokenizer.convert_tokens_to_string(lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ , """This is a test""" ) @slow def A (self ): """simple docstring""" # fmt: off A__ = {"input_ids": [[1_2_8_0_2_2, 1_1_0_1_0_8, 3_9_7, 1_1, 3_8_2_7_2, 2_2_4_7, 1_2_4_8_1_1, 2_8_5, 1_8_1_0_5, 1_5_8_6, 2_0_7, 7, 3_9_5_3_4, 4_4_2_8, 3_9_7, 1_0_1_9, 1_8_1_0_5, 1_5_8_6, 2_0_7, 7, 4_1_3_3_7, 1_6_7_8_6, 2_4_1, 7, 2_0_2_1_4, 1_7, 1_2_5_6_9_0, 1_0_3_9_8, 7, 4_4_3_7_8, 5_8_0_6_9, 6_8_3_4_2, 7_7_9_8, 7_3_4_3, 1_1, 2_9_9, 3_3_3_1_0, 4, 1_5_8, 3_7_3_5_0, 9_4_0_7_7, 4_5_6_9, 2_9_9, 3_3_3_1_0, 9_0, 4, 5_2_8_4_0, 2_9_0, 4, 3_1_2_7_0, 1_1_2, 2_9_9, 6_8_2, 4, 5_2_8_4_0, 3_9_9_5_3, 1_4_0_7_9, 1_9_3, 5_2_5_1_9, 9_0_8_9_4, 1_7_8_9_4, 1_2_0_6_9_7, 1_1, 4_0_4_4_5, 5_5_1, 1_7, 1_0_1_9, 5_2_5_1_9, 9_0_8_9_4, 1_7_7_5_6, 9_6_3, 1_1, 4_0_4_4_5, 4_8_0, 1_7, 9_7_9_2, 1_1_2_0, 5_1_7_3, 1_3_9_3, 6_2_4_0, 1_6_7_8_6, 2_4_1, 1_2_0_9_9_6, 2_8, 1_2_4_5, 1_3_9_3, 1_1_8_2_4_0, 1_1_1_2_3, 1_0_1_9, 9_3_6_1_2, 2_6_9_1, 1_0_6_1_8, 9_8_0_5_8, 1_2_0_4_0_9, 1_9_2_8, 2_7_9, 4, 4_0_6_8_3, 3_6_7, 1_7_8, 2_0_7, 1_0_1_9, 1_0_3, 1_0_3_1_2_1, 5_0_6, 6_5_2_9_6, 5, 2], [1_2_8_0_2_2, 2_1_2_1_7, 3_6_7, 1_1_7, 1_2_5_4_5_0, 1_2_8, 7_1_9, 7, 7_3_0_8, 4_0, 9_3_6_1_2, 1_2_6_6_9, 1_1_1_6, 1_6_7_0_4, 7_1, 1_7_7_8_5, 3_6_9_9, 1_5_5_9_2, 3_5, 1_4_4, 9_5_8_4, 2_4_1, 1_1_9_4_3, 7_1_3, 9_5_0, 7_9_9, 2_2_4_7, 8_8_4_2_7, 1_5_0, 1_4_9, 1_1_8_8_1_3, 1_2_0_7_0_6, 1_0_1_9, 1_0_6_9_0_6, 8_1_5_1_8, 2_8, 1_2_2_4, 2_2_7_9_9, 3_9_7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1_2_8_0_2_2, 1_6_5_8, 1_2_3_3_1_1, 5_1_5_5, 5_5_7_8, 4_7_2_2, 2_7_9, 1_4_9_4_7, 2_3_6_6, 1_1_2_0, 1_1_9_7, 1_4, 1_3_4_8, 9_2_3_2, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase__ , model_name="""facebook/m2m100_418M""" , revision="""c168bae485c864188cf9aa0e4108b0b6934dc91e""" , ) @require_torch @require_sentencepiece @require_tokenizers class _UpperCamelCase ( unittest.TestCase): __lowerCamelCase = "facebook/m2m100_418M" __lowerCamelCase = [ "In my opinion, there are two levels of response from the French government.", "NSA Affair Emphasizes Complete Lack of Debate on Intelligence", ] __lowerCamelCase = [ "Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.", "L'affaire NSA souligne l'absence totale de débat sur le renseignement", ] # fmt: off __lowerCamelCase = [EN_CODE, 5_9_3, 1_9_4_9, 1_1_5_7_8_1, 4, 7_1_5_8_6, 4_2_3_4, 6_0_6_3_3, 1_2_6_2_3_3, 4_3_2, 1_2_3_8_0_8, 1_5_5_9_2, 1_1_9_7, 1_1_7_1_3_2, 1_2_0_6_1_8, 5, 2] @classmethod def A (cls ): """simple docstring""" A__ = MaMaaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang="""en""" , tgt_lang="""fr""" ) A__ = 1 return cls def A (self ): """simple docstring""" self.assertEqual(self.tokenizer.get_lang_id("""ar""" ) , 1_2_8_0_0_6 ) self.assertEqual(self.tokenizer.get_lang_id("""en""" ) , 1_2_8_0_2_2 ) self.assertEqual(self.tokenizer.get_lang_id("""ro""" ) , 1_2_8_0_7_6 ) self.assertEqual(self.tokenizer.get_lang_id("""mr""" ) , 1_2_8_0_6_3 ) def A (self ): """simple docstring""" A__ = self.tokenizer.get_vocab() self.assertEqual(len(lowerCamelCase__ ) , self.tokenizer.vocab_size ) self.assertEqual(vocab["""<unk>"""] , 3 ) self.assertIn(self.tokenizer.get_lang_token("""en""" ) , lowerCamelCase__ ) def A (self ): """simple docstring""" A__ = "en" A__ = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , lowerCamelCase__ ) def A (self ): """simple docstring""" self.assertIn(lowerCamelCase__ , self.tokenizer.all_special_ids ) # fmt: off A__ = [FR_CODE, 5_3_6_4, 8_2, 8_6_4_2, 4, 2_9_4, 4_7, 8, 1_4_0_2_8, 1_3_6, 3_2_8_6, 9_7_0_6, 6, 9_0_7_9_7, 6, 1_4_4_0_1_2, 1_6_2, 8_8_1_2_8, 3_0_0_6_1, 5, 2] # fmt: on A__ = self.tokenizer.decode(lowerCamelCase__ , skip_special_tokens=lowerCamelCase__ ) A__ = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) self.assertNotIn(self.tokenizer.eos_token , lowerCamelCase__ ) def A (self ): """simple docstring""" A__ = tempfile.mkdtemp() A__ = self.tokenizer.lang_token_to_id self.tokenizer.save_pretrained(lowerCamelCase__ ) A__ = MaMaaaTokenizer.from_pretrained(lowerCamelCase__ ) self.assertDictEqual(new_tok.lang_token_to_id , lowerCamelCase__ ) @require_torch def A (self ): """simple docstring""" A__ = "en" A__ = "fr" A__ = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=lowerCamelCase__ , return_tensors="""pt""" ) A__ = shift_tokens_right( batch["""labels"""] , self.tokenizer.pad_token_id , self.tokenizer.eos_token_id ) for k in batch: A__ = batch[k].tolist() # batch = {k: v.tolist() for k,v in batch.items()} # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 # batch.decoder_inputs_ids[0][0] == assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == FR_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2] == [2, FR_CODE] @require_torch def A (self ): """simple docstring""" A__ = "mr" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("""mr""" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) A__ = "zh" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("""zh""" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) @require_torch def A (self ): """simple docstring""" A__ = "mr" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("""mr""" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) A__ = "zh" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("""zh""" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) @require_torch def A (self ): """simple docstring""" A__ = self.tokenizer._build_translation_inputs("""A test""" , return_tensors="""pt""" , src_lang="""en""" , tgt_lang="""ar""" ) self.assertEqual( nested_simplify(lowerCamelCase__ ) , { # en_XX, A, test, EOS """input_ids""": [[1_2_8_0_2_2, 5_8, 4_1_8_3, 2]], """attention_mask""": [[1, 1, 1, 1]], # ar_AR """forced_bos_token_id""": 1_2_8_0_0_6, } , )
574
import gc import unittest import torch from parameterized import parameterized from diffusers import AutoencoderKL from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = AutoencoderKL __UpperCAmelCase : Optional[Any] = "sample" __UpperCAmelCase : Optional[int] = 1e-2 @property def __snake_case ( self : Dict ) -> Optional[Any]: __snake_case : Optional[Any] = 4 __snake_case : Tuple = 3 __snake_case : List[str] = (32, 32) __snake_case : str = floats_tensor((batch_size, num_channels) + sizes ).to(lowerCamelCase ) return {"sample": image} @property def __snake_case ( self : Union[str, Any] ) -> Tuple: return (3, 32, 32) @property def __snake_case ( self : int ) -> int: return (3, 32, 32) def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Optional[Any] = { "block_out_channels": [32, 64], "in_channels": 3, "out_channels": 3, "down_block_types": ["DownEncoderBlock2D", "DownEncoderBlock2D"], "up_block_types": ["UpDecoderBlock2D", "UpDecoderBlock2D"], "latent_channels": 4, } __snake_case : Any = self.dummy_input return init_dict, inputs_dict def __snake_case ( self : str ) -> Dict: pass def __snake_case ( self : Tuple ) -> List[str]: pass @unittest.skipIf(torch_device == "mps" , "Gradient checkpointing skipped on MPS" ) def __snake_case ( self : Any ) -> Optional[Any]: # enable deterministic behavior for gradient checkpointing __snake_case , __snake_case : int = self.prepare_init_args_and_inputs_for_common() __snake_case : str = self.model_class(**lowerCamelCase ) model.to(lowerCamelCase ) assert not model.is_gradient_checkpointing and model.training __snake_case : str = model(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model.zero_grad() __snake_case : Any = torch.randn_like(lowerCamelCase ) __snake_case : str = (out - labels).mean() loss.backward() # re-instantiate the model now enabling gradient checkpointing __snake_case : Optional[int] = self.model_class(**lowerCamelCase ) # clone model model_a.load_state_dict(model.state_dict() ) model_a.to(lowerCamelCase ) model_a.enable_gradient_checkpointing() assert model_a.is_gradient_checkpointing and model_a.training __snake_case : int = model_a(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model_a.zero_grad() __snake_case : Union[str, Any] = (out_a - labels).mean() loss_a.backward() # compare the output and parameters gradients self.assertTrue((loss - loss_a).abs() < 1E-5 ) __snake_case : Optional[int] = dict(model.named_parameters() ) __snake_case : List[Any] = dict(model_a.named_parameters() ) for name, param in named_params.items(): self.assertTrue(torch_all_close(param.grad.data , named_params_a[name].grad.data , atol=5E-5 ) ) def __snake_case ( self : List[Any] ) -> Optional[int]: __snake_case , __snake_case : Optional[Any] = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" , output_loading_info=lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(lowerCamelCase ) __snake_case : Optional[Any] = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def __snake_case ( self : Optional[Any] ) -> Union[str, Any]: __snake_case : Tuple = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" ) __snake_case : Dict = model.to(lowerCamelCase ) model.eval() if torch_device == "mps": __snake_case : int = torch.manual_seed(0 ) else: __snake_case : str = torch.Generator(device=lowerCamelCase ).manual_seed(0 ) __snake_case : List[str] = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) __snake_case : Union[str, Any] = image.to(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , sample_posterior=lowerCamelCase , generator=lowerCamelCase ).sample __snake_case : List[Any] = output[0, -1, -3:, -3:].flatten().cpu() # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. if torch_device == "mps": __snake_case : Union[str, Any] = torch.tensor( [ -4.0078E-01, -3.8323E-04, -1.2681E-01, -1.1462E-01, 2.0095E-01, 1.0893E-01, -8.8247E-02, -3.0361E-01, -9.8644E-03, ] ) elif torch_device == "cpu": __snake_case : Tuple = torch.tensor( [-0.13_52, 0.08_78, 0.04_19, -0.08_18, -0.10_69, 0.06_88, -0.14_58, -0.44_46, -0.00_26] ) else: __snake_case : List[str] = torch.tensor( [-0.24_21, 0.46_42, 0.25_07, -0.04_38, 0.06_82, 0.31_60, -0.20_18, -0.07_27, 0.24_85] ) self.assertTrue(torch_all_close(lowerCamelCase , lowerCamelCase , rtol=1E-2 ) ) @slow class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : int , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] ) -> List[str]: return F'gaussian_noise_s={seed}_shape={"_".join([str(lowerCamelCase ) for s in shape] )}.npy' def __snake_case ( self : List[Any] ) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __snake_case ( self : Tuple , lowerCamelCase : List[Any]=0 , lowerCamelCase : Tuple=(4, 3, 512, 512) , lowerCamelCase : Optional[int]=False ) -> str: __snake_case : List[Any] = torch.floataa if fpaa else torch.floataa __snake_case : Tuple = torch.from_numpy(load_hf_numpy(self.get_file_format(lowerCamelCase , lowerCamelCase ) ) ).to(lowerCamelCase ).to(lowerCamelCase ) return image def __snake_case ( self : Optional[Any] , lowerCamelCase : int="CompVis/stable-diffusion-v1-4" , lowerCamelCase : int=False ) -> int: __snake_case : str = "fp16" if fpaa else None __snake_case : int = torch.floataa if fpaa else torch.floataa __snake_case : int = AutoencoderKL.from_pretrained( lowerCamelCase , subfolder="vae" , torch_dtype=lowerCamelCase , revision=lowerCamelCase , ) model.to(lowerCamelCase ).eval() return model def __snake_case ( self : str , lowerCamelCase : int=0 ) -> Optional[Any]: if torch_device == "mps": return torch.manual_seed(lowerCamelCase ) return torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) @parameterized.expand( [ # fmt: off [33, [-0.16_03, 0.98_78, -0.04_95, -0.07_90, -0.27_09, 0.83_75, -0.20_60, -0.08_24], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_76, 0.11_68, 0.13_32, -0.48_40, -0.25_08, -0.07_91, -0.04_93, -0.40_89], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[str] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] ) -> List[Any]: __snake_case : Optional[Any] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) __snake_case : Tuple = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : List[Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : int = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [33, [-0.05_13, 0.02_89, 1.37_99, 0.21_66, -0.25_73, -0.08_71, 0.51_03, -0.09_99]], [47, [-0.41_28, -0.13_20, -0.37_04, 0.19_65, -0.41_16, -0.23_32, -0.33_40, 0.22_47]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : Any , lowerCamelCase : List[str] , lowerCamelCase : List[str] ) -> Tuple: __snake_case : Any = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Any = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.16_09, 0.98_66, -0.04_87, -0.07_77, -0.27_16, 0.83_68, -0.20_55, -0.08_14], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_77, 0.11_47, 0.13_33, -0.48_41, -0.25_06, -0.08_05, -0.04_91, -0.40_85], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Any , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) with torch.no_grad(): __snake_case : int = model(lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Union[str, Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : List[str] = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [13, [-0.20_51, -0.18_03, -0.23_11, -0.21_14, -0.32_92, -0.35_74, -0.29_53, -0.33_23]], [37, [-0.26_32, -0.26_25, -0.21_99, -0.27_41, -0.45_39, -0.49_90, -0.37_20, -0.49_25]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : List[str] , lowerCamelCase : Tuple , lowerCamelCase : Any ) -> Optional[Any]: __snake_case : List[str] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : str = sample[-1, -2:, :2, -2:].flatten().cpu() __snake_case : Optional[int] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-3 ) @parameterized.expand( [ # fmt: off [27, [-0.03_69, 0.02_07, -0.07_76, -0.06_82, -0.17_47, -0.19_30, -0.14_65, -0.20_39]], [16, [-0.16_28, -0.21_34, -0.27_47, -0.26_42, -0.37_74, -0.44_04, -0.36_87, -0.42_77]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : str , lowerCamelCase : Optional[int] , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[str] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : Union[str, Any] = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Optional[Any] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=5E-3 ) @parameterized.expand([(13,), (16,), (27,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> Tuple: __snake_case : Dict = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : Any = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Any = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-1 ) @parameterized.expand([(13,), (16,), (37,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : List[Any] , lowerCamelCase : Any ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : Union[str, Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : List[Any] = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Dict = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.30_01, 0.09_18, -2.69_84, -3.97_20, -3.20_99, -5.03_53, 1.73_38, -0.20_65, 3.42_67]], [47, [-1.50_30, -4.38_71, -6.03_55, -9.11_57, -1.66_61, -2.78_53, 2.16_07, -5.08_23, 2.56_33]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Dict ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : int = self.get_sd_image(lowerCamelCase ) __snake_case : int = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model.encode(lowerCamelCase ).latent_dist __snake_case : Dict = dist.sample(generator=lowerCamelCase ) assert list(sample.shape ) == [image.shape[0], 4] + [i // 8 for i in image.shape[2:]] __snake_case : List[str] = sample[0, -1, -3:, -3:].flatten().cpu() __snake_case : Dict = torch.tensor(lowerCamelCase ) __snake_case : Dict = 3E-3 if torch_device != "mps" else 1E-2 assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=lowerCamelCase )
81
0