code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
'''simple docstring''' from numpy import exp, pi, sqrt def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase = 0.0 , UpperCAmelCase = 1.0 ): return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
152
import warnings from ...utils import logging from .image_processing_flava import FlavaImageProcessor _snake_case : Optional[int] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[Any] , *lowerCamelCase : Any , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class FlavaFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use FlavaImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowerCAmelCase_ = logging.get_logger(__name__) class A (_lowerCAmelCase ): _SCREAMING_SNAKE_CASE = ["pixel_values"] def __init__( self , lowercase_ = True , lowercase_ = None , lowercase_ = None , lowercase_ = PILImageResampling.BILINEAR , lowercase_ = True , lowercase_ = 1 / 255 , lowercase_ = True , lowercase_ = None , lowercase_ = None , **lowercase_ , ) -> None: '''simple docstring''' super().__init__(**lowercase_ ) _snake_case : List[Any] = size if size is not None else {"shortest_edge": 384} _snake_case : List[str] = get_size_dict(lowercase_ , default_to_square=lowercase_ ) _snake_case : Optional[int] = do_resize _snake_case : List[Any] = size # Default value set here for backwards compatibility where the value in config is None _snake_case : List[str] = crop_pct if crop_pct is not None else 224 / 256 _snake_case : List[str] = resample _snake_case : Union[str, Any] = do_rescale _snake_case : List[str] = rescale_factor _snake_case : int = do_normalize _snake_case : Union[str, Any] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _snake_case : Dict = image_std if image_std is not None else IMAGENET_STANDARD_STD def __a ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ = PILImageResampling.BICUBIC , lowercase_ = None , **lowercase_ , ) -> np.ndarray: '''simple docstring''' _snake_case : Optional[int] = get_size_dict(lowercase_ , default_to_square=lowercase_ ) if "shortest_edge" not in size: raise ValueError(F'''Size dictionary must contain \'shortest_edge\' key. Got {size.keys()}''' ) _snake_case : List[Any] = size["shortest_edge"] if shortest_edge < 384: # maintain same ratio, resizing shortest edge to shortest_edge/crop_pct _snake_case : str = int(shortest_edge / crop_pct ) _snake_case : List[Any] = get_resize_output_image_size(lowercase_ , size=lowercase_ , default_to_square=lowercase_ ) _snake_case : Optional[Any] = resize(image=lowercase_ , size=lowercase_ , resample=lowercase_ , data_format=lowercase_ , **lowercase_ ) # then crop to (shortest_edge, shortest_edge) return center_crop(image=lowercase_ , size=(shortest_edge, shortest_edge) , data_format=lowercase_ , **lowercase_ ) else: # warping (no cropping) when evaluated at 384 or larger return resize( lowercase_ , size=(shortest_edge, shortest_edge) , resample=lowercase_ , data_format=lowercase_ , **lowercase_ ) def __a ( self , lowercase_ , lowercase_ , lowercase_ = None , **lowercase_ , ) -> Union[str, Any]: '''simple docstring''' return rescale(lowercase_ , scale=lowercase_ , data_format=lowercase_ , **lowercase_ ) def __a ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ = None , **lowercase_ , ) -> np.ndarray: '''simple docstring''' return normalize(lowercase_ , mean=lowercase_ , std=lowercase_ , data_format=lowercase_ , **lowercase_ ) def __a ( self , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = ChannelDimension.FIRST , **lowercase_ , ) -> PIL.Image.Image: '''simple docstring''' _snake_case : Union[str, Any] = do_resize if do_resize is not None else self.do_resize _snake_case : Optional[int] = crop_pct if crop_pct is not None else self.crop_pct _snake_case : Dict = resample if resample is not None else self.resample _snake_case : Optional[Any] = do_rescale if do_rescale is not None else self.do_rescale _snake_case : str = rescale_factor if rescale_factor is not None else self.rescale_factor _snake_case : List[Any] = do_normalize if do_normalize is not None else self.do_normalize _snake_case : List[Any] = image_mean if image_mean is not None else self.image_mean _snake_case : str = image_std if image_std is not None else self.image_std _snake_case : str = size if size is not None else self.size _snake_case : List[str] = get_size_dict(lowercase_ , default_to_square=lowercase_ ) _snake_case : Any = make_list_of_images(lowercase_ ) if not valid_images(lowercase_ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_resize and size["shortest_edge"] < 384 and crop_pct is None: raise ValueError('''crop_pct must be specified if size < 384.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # All transformations expect numpy arrays. _snake_case : Tuple = [to_numpy_array(lowercase_ ) for image in images] if do_resize: _snake_case : Optional[int] = [self.resize(image=lowercase_ , size=lowercase_ , crop_pct=lowercase_ , resample=lowercase_ ) for image in images] if do_rescale: _snake_case : Any = [self.rescale(image=lowercase_ , scale=lowercase_ ) for image in images] if do_normalize: _snake_case : List[Any] = [self.normalize(image=lowercase_ , mean=lowercase_ , std=lowercase_ ) for image in images] _snake_case : List[Any] = [to_channel_dimension_format(lowercase_ , lowercase_ ) for image in images] _snake_case : int = {"pixel_values": images} return BatchFeature(data=lowercase_ , tensor_type=lowercase_ )
326
import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : Tuple ) -> Optional[Any]: __snake_case : Dict = tempfile.mkdtemp() __snake_case : Any = SamImageProcessor() __snake_case : Optional[int] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : Optional[Any] , **lowerCamelCase : Optional[int] ) -> Optional[Any]: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[Any] ) -> Dict: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : int ) -> List[Any]: __snake_case : List[str] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : int = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : List[Any] ) -> Dict: __snake_case : int = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Dict = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Optional[Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : List[str] ) -> Tuple: __snake_case : int = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Optional[int] = self.prepare_image_inputs() __snake_case : List[str] = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : Dict = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_torch def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Tuple = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[str] = [torch.ones((1, 3, 5, 5) )] __snake_case : Tuple = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : int = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , torch.tensor(lowerCamelCase ) , torch.tensor(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : List[str] = [np.ones((1, 3, 5, 5) )] __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : str = [[1, 0], [0, 1]] with self.assertRaises(lowerCamelCase ): __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) @require_vision @require_tf class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[Any] ) -> Union[str, Any]: __snake_case : int = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : str , **lowerCamelCase : Any ) -> Tuple: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> Any: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : str ) -> List[Any]: __snake_case : int = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : Dict = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : int ) -> List[str]: __snake_case : List[Any] = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Optional[Any] = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Tuple = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : Union[str, Any] ) -> List[Any]: __snake_case : str = self.get_image_processor() __snake_case : Union[str, Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : int = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : List[str] = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_tf def __snake_case ( self : Any ) -> Optional[int]: __snake_case : List[str] = self.get_image_processor() __snake_case : Dict = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Union[str, Any] = [tf.ones((1, 3, 5, 5) )] __snake_case : List[Any] = [[1764, 2646]] __snake_case : Dict = [[683, 1024]] __snake_case : List[str] = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , tf.convert_to_tensor(lowerCamelCase ) , tf.convert_to_tensor(lowerCamelCase ) , return_tensors="tf" , ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : Union[str, Any] = [np.ones((1, 3, 5, 5) )] __snake_case : List[str] = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Tuple = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): __snake_case : Dict = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[str] ) -> str: __snake_case : Optional[int] = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : List[str] , **lowerCamelCase : Any ) -> Dict: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> List[Any]: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : str = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : List[Any] = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : str = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[Any] = np.random.randint(0 , 2 , size=(1, 3, 5, 5) ).astype(np.floataa ) __snake_case : Dict = [tf.convert_to_tensor(lowerCamelCase )] __snake_case : List[Any] = [torch.tensor(lowerCamelCase )] __snake_case : Optional[Any] = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : Union[str, Any] = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) __snake_case : Dict = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="pt" ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def __snake_case ( self : List[Any] ) -> List[str]: __snake_case : Any = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : Any = image_processor(lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Optional[Any] = processor(images=lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Tuple = image_processor(lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() __snake_case : List[Any] = processor(images=lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) )
81
0
"""simple docstring""" import unittest from typing import Dict, List, Optional, Union import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BridgeTowerImageProcessor class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): """simple docstring""" def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ = True , lowerCAmelCase__ = None , lowerCAmelCase__ = 3_2 , lowerCAmelCase__ = True , lowerCAmelCase__ = 1 / 2_5_5 , lowerCAmelCase__ = True , lowerCAmelCase__ = True , lowerCAmelCase__ = [0.48_14_54_66, 0.4_57_82_75, 0.40_82_10_73] , lowerCAmelCase__ = [0.26_86_29_54, 0.26_13_02_58, 0.27_57_77_11] , lowerCAmelCase__ = True , lowerCAmelCase__=7 , lowerCAmelCase__=3_0 , lowerCAmelCase__=4_0_0 , lowerCAmelCase__=3 , ): __SCREAMING_SNAKE_CASE = parent __SCREAMING_SNAKE_CASE = do_resize __SCREAMING_SNAKE_CASE = size if size is not None else {"shortest_edge": 2_8_8} __SCREAMING_SNAKE_CASE = size_divisor __SCREAMING_SNAKE_CASE = do_rescale __SCREAMING_SNAKE_CASE = rescale_factor __SCREAMING_SNAKE_CASE = do_normalize __SCREAMING_SNAKE_CASE = do_center_crop __SCREAMING_SNAKE_CASE = image_mean __SCREAMING_SNAKE_CASE = image_std __SCREAMING_SNAKE_CASE = do_pad __SCREAMING_SNAKE_CASE = batch_size __SCREAMING_SNAKE_CASE = num_channels __SCREAMING_SNAKE_CASE = min_resolution __SCREAMING_SNAKE_CASE = max_resolution def snake_case_ ( self): return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "size_divisor": self.size_divisor, } def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__=False): if not batched: __SCREAMING_SNAKE_CASE = self.size["shortest_edge"] __SCREAMING_SNAKE_CASE = image_inputs[0] if isinstance(lowerCAmelCase__ , Image.Image): __SCREAMING_SNAKE_CASE = image.size else: __SCREAMING_SNAKE_CASE = image.shape[1], image.shape[2] __SCREAMING_SNAKE_CASE = size / min(lowerCAmelCase__ , lowerCAmelCase__) if h < w: __SCREAMING_SNAKE_CASE = size, scale * w else: __SCREAMING_SNAKE_CASE = scale * h, size __SCREAMING_SNAKE_CASE = int((1_3_3_3 / 8_0_0) * size) if max(lowerCAmelCase__ , lowerCAmelCase__) > max_size: __SCREAMING_SNAKE_CASE = max_size / max(lowerCAmelCase__ , lowerCAmelCase__) __SCREAMING_SNAKE_CASE = newh * scale __SCREAMING_SNAKE_CASE = neww * scale __SCREAMING_SNAKE_CASE = int(newh + 0.5), int(neww + 0.5) __SCREAMING_SNAKE_CASE = ( newh // self.size_divisor * self.size_divisor, neww // self.size_divisor * self.size_divisor, ) else: __SCREAMING_SNAKE_CASE = [] for image in image_inputs: __SCREAMING_SNAKE_CASE = self.get_expected_values([image]) expected_values.append((expected_height, expected_width)) __SCREAMING_SNAKE_CASE = max(lowerCAmelCase__ , key=lambda lowerCAmelCase__: item[0])[0] __SCREAMING_SNAKE_CASE = max(lowerCAmelCase__ , key=lambda lowerCAmelCase__: item[1])[1] return expected_height, expected_width @require_torch @require_vision class SCREAMING_SNAKE_CASE_ ( _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __lowercase : Any = BridgeTowerImageProcessor if is_vision_available() else None def snake_case_ ( self): __SCREAMING_SNAKE_CASE = BridgeTowerImageProcessingTester(self) @property def snake_case_ ( self): return self.image_processor_tester.prepare_image_processor_dict() def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(lowerCAmelCase__ , """image_mean""")) self.assertTrue(hasattr(lowerCAmelCase__ , """image_std""")) self.assertTrue(hasattr(lowerCAmelCase__ , """do_normalize""")) self.assertTrue(hasattr(lowerCAmelCase__ , """do_resize""")) self.assertTrue(hasattr(lowerCAmelCase__ , """size""")) self.assertTrue(hasattr(lowerCAmelCase__ , """size_divisor""")) def snake_case_ ( self): pass def snake_case_ ( self): # Initialize image processor __SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict) # create random PIL images __SCREAMING_SNAKE_CASE = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , Image.Image) # Test not batched input __SCREAMING_SNAKE_CASE = image_processing(image_inputs[0] , return_tensors="""pt""").pixel_values __SCREAMING_SNAKE_CASE = self.image_processor_tester.get_expected_values(lowerCAmelCase__) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __SCREAMING_SNAKE_CASE = image_processing(lowerCAmelCase__ , return_tensors="""pt""").pixel_values __SCREAMING_SNAKE_CASE = self.image_processor_tester.get_expected_values(lowerCAmelCase__ , batched=lowerCAmelCase__) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def snake_case_ ( self): # Initialize image processor __SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors __SCREAMING_SNAKE_CASE = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ , numpify=lowerCAmelCase__) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , np.ndarray) # Test not batched input __SCREAMING_SNAKE_CASE = image_processing(image_inputs[0] , return_tensors="""pt""").pixel_values __SCREAMING_SNAKE_CASE = self.image_processor_tester.get_expected_values(lowerCAmelCase__) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __SCREAMING_SNAKE_CASE = image_processing(lowerCAmelCase__ , return_tensors="""pt""").pixel_values __SCREAMING_SNAKE_CASE = self.image_processor_tester.get_expected_values(lowerCAmelCase__ , batched=lowerCAmelCase__) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def snake_case_ ( self): # Initialize image processor __SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors __SCREAMING_SNAKE_CASE = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ , torchify=lowerCAmelCase__) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , torch.Tensor) # Test not batched input __SCREAMING_SNAKE_CASE = image_processing(image_inputs[0] , return_tensors="""pt""").pixel_values __SCREAMING_SNAKE_CASE = self.image_processor_tester.get_expected_values(lowerCAmelCase__) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __SCREAMING_SNAKE_CASE = image_processing(lowerCAmelCase__ , return_tensors="""pt""").pixel_values __SCREAMING_SNAKE_CASE = self.image_processor_tester.get_expected_values(lowerCAmelCase__ , batched=lowerCAmelCase__) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , )
155
import warnings from ...utils import logging from .image_processing_deit import DeiTImageProcessor _snake_case : List[str] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[str] , *lowerCamelCase : str , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class DeiTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use DeiTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
def lowerCAmelCase_ (lowercase__ : Tuple , lowercase__ : List[str] ) -> Dict: '''simple docstring''' if a < 0 or b < 0: raise ValueError('''the value of both inputs must be positive''' ) lowerCAmelCase__ = str(bin(__lowerCamelCase ) )[2:] # remove the leading "0b" lowerCAmelCase__ = str(bin(__lowerCamelCase ) )[2:] # remove the leading "0b" lowerCAmelCase__ = max(len(__lowerCamelCase ) , len(__lowerCamelCase ) ) return "0b" + "".join( str(int(char_a != char_b ) ) for char_a, char_b in zip(a_binary.zfill(__lowerCamelCase ) , b_binary.zfill(__lowerCamelCase ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
668
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case : Union[str, Any] = { "configuration_owlvit": [ "OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "OwlViTConfig", "OwlViTOnnxConfig", "OwlViTTextConfig", "OwlViTVisionConfig", ], "processing_owlvit": ["OwlViTProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[Any] = ["OwlViTFeatureExtractor"] _snake_case : Optional[int] = ["OwlViTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "OwlViTModel", "OwlViTPreTrainedModel", "OwlViTTextModel", "OwlViTVisionModel", "OwlViTForObjectDetection", ] if TYPE_CHECKING: from .configuration_owlvit import ( OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, OwlViTConfig, OwlViTOnnxConfig, OwlViTTextConfig, OwlViTVisionConfig, ) from .processing_owlvit import OwlViTProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_owlvit import OwlViTFeatureExtractor from .image_processing_owlvit import OwlViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_owlvit import ( OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST, OwlViTForObjectDetection, OwlViTModel, OwlViTPreTrainedModel, OwlViTTextModel, OwlViTVisionModel, ) else: import sys _snake_case : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
0
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = {"vocab_file": "spm_char.model"} __UpperCamelCase = { "vocab_file": { "microsoft/speecht5_asr": "https://huggingface.co/microsoft/speecht5_asr/resolve/main/spm_char.model", "microsoft/speecht5_tts": "https://huggingface.co/microsoft/speecht5_tts/resolve/main/spm_char.model", "microsoft/speecht5_vc": "https://huggingface.co/microsoft/speecht5_vc/resolve/main/spm_char.model", } } __UpperCamelCase = { "microsoft/speecht5_asr": 1024, "microsoft/speecht5_tts": 1024, "microsoft/speecht5_vc": 1024, } class _A ( _lowerCAmelCase ): lowercase__: str = VOCAB_FILES_NAMES lowercase__: Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP lowercase__: Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__: Dict = ["input_ids", "attention_mask"] def __init__( self : Optional[Any] , __magic_name__ : Optional[int] , __magic_name__ : str="<s>" , __magic_name__ : Union[str, Any]="</s>" , __magic_name__ : Any="<unk>" , __magic_name__ : Union[str, Any]="<pad>" , __magic_name__ : Optional[Dict[str, Any]] = None , **__magic_name__ : Any , ) -> None: """simple docstring""" __snake_case : Optional[int] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=__magic_name__ , eos_token=__magic_name__ , unk_token=__magic_name__ , pad_token=__magic_name__ , sp_model_kwargs=self.sp_model_kwargs , **__magic_name__ , ) __snake_case : Dict = vocab_file __snake_case : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__magic_name__ ) @property def lowercase__ ( self : int ) -> Tuple: """simple docstring""" return self.sp_model.get_piece_size() def lowercase__ ( self : List[str] ) -> Any: """simple docstring""" __snake_case : List[str] = {self.convert_ids_to_tokens(__magic_name__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : int ) -> Union[str, Any]: """simple docstring""" __snake_case : Optional[int] = self.__dict__.copy() __snake_case : Optional[Any] = None return state def __setstate__( self : int , __magic_name__ : str ) -> Tuple: """simple docstring""" __snake_case : Tuple = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): __snake_case : Tuple = {} __snake_case : Dict = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def lowercase__ ( self : List[str] , __magic_name__ : str ) -> List[str]: """simple docstring""" return self.sp_model.encode(__magic_name__ , out_type=__magic_name__ ) def lowercase__ ( self : Union[str, Any] , __magic_name__ : Union[str, Any] ) -> Optional[int]: """simple docstring""" return self.sp_model.piece_to_id(__magic_name__ ) def lowercase__ ( self : Tuple , __magic_name__ : Dict ) -> Tuple: """simple docstring""" __snake_case : List[Any] = self.sp_model.IdToPiece(__magic_name__ ) return token def lowercase__ ( self : int , __magic_name__ : Any ) -> Tuple: """simple docstring""" __snake_case : Optional[Any] = [] __snake_case : Tuple = "" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(__magic_name__ ) + token __snake_case : Tuple = [] else: current_sub_tokens.append(__magic_name__ ) out_string += self.sp_model.decode(__magic_name__ ) return out_string.strip() def lowercase__ ( self : Any , __magic_name__ : Optional[Any] , __magic_name__ : Optional[Any]=None ) -> List[int]: """simple docstring""" if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def lowercase__ ( self : Optional[Any] , __magic_name__ : List[int] , __magic_name__ : Optional[List[int]] = None , __magic_name__ : bool = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__magic_name__ , token_ids_a=__magic_name__ , already_has_special_tokens=__magic_name__ ) __snake_case : str = [1] if token_ids_a is None: return ([0] * len(__magic_name__ )) + suffix_ones return ([0] * len(__magic_name__ )) + ([0] * len(__magic_name__ )) + suffix_ones def lowercase__ ( self : Any , __magic_name__ : str , __magic_name__ : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(__magic_name__ ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return __snake_case : Optional[int] = os.path.join( __magic_name__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__magic_name__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __magic_name__ ) elif not os.path.isfile(self.vocab_file ): with open(__magic_name__ , """wb""" ) as fi: __snake_case : int = self.sp_model.serialized_model_proto() fi.write(__magic_name__ ) return (out_vocab_file,)
26
from ....configuration_utils import PretrainedConfig from ....utils import logging _snake_case : int = logging.get_logger(__name__) _snake_case : str = { "Visual-Attention-Network/van-base": ( "https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json" ), } class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[Any] = "van" def __init__( self : Optional[int] , lowerCamelCase : Any=224 , lowerCamelCase : str=3 , lowerCamelCase : Any=[7, 3, 3, 3] , lowerCamelCase : Dict=[4, 2, 2, 2] , lowerCamelCase : List[Any]=[64, 128, 320, 512] , lowerCamelCase : str=[3, 3, 12, 3] , lowerCamelCase : Dict=[8, 8, 4, 4] , lowerCamelCase : Any="gelu" , lowerCamelCase : Optional[int]=0.02 , lowerCamelCase : Tuple=1E-6 , lowerCamelCase : Optional[int]=1E-2 , lowerCamelCase : int=0.0 , lowerCamelCase : Optional[Any]=0.0 , **lowerCamelCase : Optional[int] , ) -> int: super().__init__(**lowerCamelCase ) __snake_case : Union[str, Any] = image_size __snake_case : Any = num_channels __snake_case : Any = patch_sizes __snake_case : List[Any] = strides __snake_case : str = hidden_sizes __snake_case : Dict = depths __snake_case : Optional[int] = mlp_ratios __snake_case : Dict = hidden_act __snake_case : Union[str, Any] = initializer_range __snake_case : List[str] = layer_norm_eps __snake_case : Optional[int] = layer_scale_init_value __snake_case : List[Any] = drop_path_rate __snake_case : int = dropout_rate
81
0
'''simple docstring''' from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase = { "configuration_informer": [ "INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "InformerConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ "INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "InformerForPrediction", "InformerModel", "InformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys UpperCAmelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
119
import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer _snake_case : Union[str, Any] = logging.getLogger(__name__) def lowerCAmelCase_ ( ): __snake_case : int = argparse.ArgumentParser( description="Prepare TFRecord shards from pre-tokenized samples of the wikitext dataset." ) parser.add_argument( "--dataset_name" , type=__lowerCamelCase , default="wikitext" , help="Name of the training. Explore datasets at: hf.co/datasets." , ) parser.add_argument( "--dataset_config" , type=__lowerCamelCase , default="wikitext-103-raw-v1" , help="Configuration name of the dataset." ) parser.add_argument( "--tokenizer_name_or_path" , type=__lowerCamelCase , default="sayakpaul/unigram-tokenizer-wikitext" , help="Tokenizer identifier. Can be a local filepath or a Hub identifier." , ) parser.add_argument( "--shard_size" , type=__lowerCamelCase , default=1_0_0_0 , help="Number of entries to go in a single shard." , ) parser.add_argument("--split" , type=__lowerCamelCase , default="train" , choices=["train", "test", "validation"] ) parser.add_argument( "--limit" , default=__lowerCamelCase , type=__lowerCamelCase , help="Limit the number of shards (used for debugging)." , ) parser.add_argument( "--max_length" , type=__lowerCamelCase , default=5_1_2 , help="Maximum sequence length. For training on TPUs, it helps to have a maximum" " sequence length that is a multiple of 8." , ) parser.add_argument( "--output_dir" , default="tf-tpu" , type=__lowerCamelCase , help="Output directory where the TFRecord shards will be saved. If the" " path is appended with `gs://` ('gs://tf-tpu', for example) then the TFRecord" " shards will be directly saved to a Google Cloud Storage bucket." , ) __snake_case : List[str] = parser.parse_args() return args def lowerCAmelCase_ ( __lowerCamelCase ): def fn(__lowerCamelCase ): return tokenizer(examples["text"] ) return fn def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = [] for i in range(len(tokenized_data["input_ids"] ) ): __snake_case : Tuple = { "input_ids": tf.train.Feature(intaa_list=tf.train.IntaaList(value=tokenized_data["input_ids"][i] ) ), "attention_mask": tf.train.Feature( intaa_list=tf.train.IntaaList(value=tokenized_data["attention_mask"][i] ) ), } __snake_case : List[Any] = tf.train.Features(feature=__lowerCamelCase ) __snake_case : str = tf.train.Example(features=__lowerCamelCase ) __snake_case : List[str] = example.SerializeToString() records.append(__lowerCamelCase ) return records def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Optional[int] = datasets.load_dataset(args.dataset_name , args.dataset_config , split=args.split ) if args.limit is not None: __snake_case : Optional[Any] = min(len(__lowerCamelCase ) , args.limit ) __snake_case : Dict = dataset.select(range(__lowerCamelCase ) ) print(F'Limiting the dataset to {args.limit} entries.' ) __snake_case : Dict = AutoTokenizer.from_pretrained(args.tokenizer_name_or_path ) # Handle output directory creation. # For serializing into a Google Cloud Storage Bucket, one needs to first # create a bucket. if "gs" not in args.output_dir: if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) __snake_case : Dict = os.path.join(args.output_dir , args.split ) if not os.path.exists(__lowerCamelCase ): os.makedirs(__lowerCamelCase ) else: __snake_case : str = os.path.join(args.output_dir , args.split ) # Tokenize the whole dataset at once. __snake_case : Any = tokenize_function(__lowerCamelCase ) __snake_case : Optional[Any] = dataset.map(__lowerCamelCase , batched=__lowerCamelCase , num_proc=4 , remove_columns=["text"] ) # We need to concatenate all our texts together, and then split the result # into chunks of a fixed size, which we will call block_size. To do this, we # will use the map method again, with the option batched=True. When we use batched=True, # the function we pass to map() will be passed multiple inputs at once, allowing us # to group them into more or fewer examples than we had in the input. # This allows us to create our new fixed-length samples. The advantage of this # method is that we don't lose a whole lot of content from the dataset compared to the # case where we simply tokenize with a pre-defined max_length. def group_texts(__lowerCamelCase ): # Concatenate all texts. __snake_case : List[str] = {k: sum(examples[k] , [] ) for k in examples.keys()} __snake_case : List[Any] = len(concatenated_examples[list(examples.keys() )[0]] ) # We drop the small remainder, though you could add padding instead if the model supports it # In this, as in all things, we advise you to follow your heart 🫀 __snake_case : Any = (total_length // args.max_length) * args.max_length # Split by chunks of max_len. __snake_case : int = { k: [t[i : i + args.max_length] for i in range(0 , __lowerCamelCase , args.max_length )] for k, t in concatenated_examples.items() } return result __snake_case : Any = dataset_tokenized.map(__lowerCamelCase , batched=__lowerCamelCase , batch_size=1_0_0_0 , num_proc=4 ) __snake_case : Optional[Any] = 0 __snake_case : Optional[Any] = 0 for shard in range(0 , len(__lowerCamelCase ) , args.shard_size ): __snake_case : List[str] = grouped_dataset[shard : shard + args.shard_size] __snake_case : Any = len(dataset_snapshot["input_ids"] ) __snake_case : List[Any] = os.path.join(__lowerCamelCase , F'dataset-{shard_count}-{records_containing}.tfrecord' ) __snake_case : Optional[Any] = get_serialized_examples(__lowerCamelCase ) with tf.io.TFRecordWriter(__lowerCamelCase ) as out_file: for i in range(len(__lowerCamelCase ) ): __snake_case : Union[str, Any] = serialized_examples[i] out_file.write(__lowerCamelCase ) print("Wrote file {} containing {} records".format(__lowerCamelCase , __lowerCamelCase ) ) shard_count += 1 total_records += records_containing with open(F'split-{args.split}-records-count.txt' , "w" ) as f: print(F'Total {args.split} records: {total_records}' , file=__lowerCamelCase ) if __name__ == "__main__": _snake_case : List[Any] = parse_args() main(args)
81
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available SCREAMING_SNAKE_CASE__ : Optional[int] = { "configuration_poolformer": [ "POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "PoolFormerConfig", "PoolFormerOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : str = ["PoolFormerFeatureExtractor"] SCREAMING_SNAKE_CASE__ : List[Any] = ["PoolFormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : str = [ "POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "PoolFormerForImageClassification", "PoolFormerModel", "PoolFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__ : str = _LazyModule(__name__, globals()["__file__"], _import_structure)
298
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) _snake_case : List[Any] = "\\n Text data.\n Second line of data." _snake_case : Tuple = "file" @pytest.fixture(scope="session" ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / (FILE_PATH + ".zstd") __snake_case : Optional[Any] = bytes(__lowerCamelCase , "utf-8" ) with zstd.open(__lowerCamelCase , "wb" ) as f: f.write(__lowerCamelCase ) return path @pytest.fixture def lowerCAmelCase_ ( __lowerCamelCase ): with open(os.path.join(tmpfs.local_root_dir , __lowerCamelCase ) , "w" ) as f: f.write(__lowerCamelCase ) return FILE_PATH @pytest.mark.parametrize("compression_format" , ["gzip", "xz", "zstd"] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Optional[int] = {"gzip": gz_file, "xz": xz_file, "zstd": zstd_path} __snake_case : str = input_paths[compression_format] __snake_case : Optional[Any] = tmp_path / "cache" __snake_case : Optional[int] = DownloadConfig(cache_dir=__lowerCamelCase , extract_compressed_file=__lowerCamelCase ) __snake_case : Union[str, Any] = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) with open(__lowerCamelCase ) as f: __snake_case : Dict = f.read() with open(__lowerCamelCase ) as f: __snake_case : Tuple = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("default_extracted" , [True, False] ) @pytest.mark.parametrize("default_cache_dir" , [True, False] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Tuple = "custom_cache" __snake_case : List[str] = "custom_extracted_dir" __snake_case : Any = tmp_path / "custom_extracted_path" if default_extracted: __snake_case : List[Any] = ("downloads" if default_cache_dir else custom_cache_dir, "extracted") else: monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_DIR" , __lowerCamelCase ) monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_PATH" , str(__lowerCamelCase ) ) __snake_case : Optional[Any] = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) __snake_case : Optional[int] = xz_file __snake_case : Optional[int] = ( DownloadConfig(extract_compressed_file=__lowerCamelCase ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=__lowerCamelCase ) ) __snake_case : str = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) assert Path(__lowerCamelCase ).parent.parts[-2:] == expected def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : Optional[Any] = str(Path(__lowerCamelCase ).resolve() ) assert cached_path(__lowerCamelCase ) == text_file # relative path __snake_case : Any = str(Path(__lowerCamelCase ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(__lowerCamelCase ) == text_file def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : List[Any] = str(tmp_path.resolve() / "__missing_file__.txt" ) with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) # relative path __snake_case : Optional[int] = "./__missing_file__.txt" with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : str = get_from_cache(F'tmp://{tmpfs_file}' ) with open(__lowerCamelCase ) as f: __snake_case : Union[str, Any] = f.read() assert output_file_content == FILE_CONTENT @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( ): with pytest.raises(__lowerCamelCase ): cached_path("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[Any] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): http_get("https://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): http_head("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[str] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): ftp_get("ftp://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): ftp_head("ftp://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): fsspec_get("s3://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): fsspec_head("s3://huggingface.co" )
81
0
"""simple docstring""" import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __SCREAMING_SNAKE_CASE =get_tests_dir("fixtures/test_sentencepiece.model") @require_sentencepiece @require_tokenizers class UpperCamelCase ( _lowerCAmelCase , unittest.TestCase ): lowercase = XLMRobertaTokenizer lowercase = XLMRobertaTokenizerFast lowercase = True lowercase = True def _UpperCAmelCase ( self ) -> Union[str, Any]: '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing lowercase_ : Optional[int] = XLMRobertaTokenizer(__UpperCamelCase ,keep_accents=__UpperCamelCase ) tokenizer.save_pretrained(self.tmpdirname ) def _UpperCAmelCase ( self ) -> int: '''simple docstring''' lowercase_ : Optional[int] = "<pad>" lowercase_ : Any = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__UpperCamelCase ) ,__UpperCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__UpperCamelCase ) ,__UpperCamelCase ) def _UpperCAmelCase ( self ) -> int: '''simple docstring''' lowercase_ : Optional[int] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,'<s>' ) self.assertEqual(vocab_keys[1] ,'<pad>' ) self.assertEqual(vocab_keys[-1] ,'<mask>' ) self.assertEqual(len(__UpperCamelCase ) ,1002 ) def _UpperCAmelCase ( self ) -> str: '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size ,1002 ) def _UpperCAmelCase ( self ) -> List[Any]: '''simple docstring''' lowercase_ : Union[str, Any] = XLMRobertaTokenizer(__UpperCamelCase ,keep_accents=__UpperCamelCase ) lowercase_ : int = tokenizer.tokenize('This is a test' ) self.assertListEqual(__UpperCamelCase ,['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__UpperCamelCase ) ,[value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] ,) lowercase_ : List[Any] = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( __UpperCamelCase ,[ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] ,) lowercase_ : int = tokenizer.convert_tokens_to_ids(__UpperCamelCase ) self.assertListEqual( __UpperCamelCase ,[ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] ,) lowercase_ : Optional[int] = tokenizer.convert_ids_to_tokens(__UpperCamelCase ) self.assertListEqual( __UpperCamelCase ,[ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] ,) def _UpperCAmelCase ( self ) -> int: '''simple docstring''' if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return lowercase_ : List[str] = (self.rust_tokenizer_class, "hf-internal-testing/tiny-xlm-roberta", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowercase_ : Optional[int] = self.rust_tokenizer_class.from_pretrained(__UpperCamelCase ,**__UpperCamelCase ) lowercase_ : Optional[Any] = self.tokenizer_class.from_pretrained(__UpperCamelCase ,**__UpperCamelCase ) lowercase_ : str = tempfile.mkdtemp() lowercase_ : List[str] = tokenizer_r.save_pretrained(__UpperCamelCase ) lowercase_ : Any = tokenizer_p.save_pretrained(__UpperCamelCase ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('tokenizer.json' in f for f in tokenizer_r_files ) ) lowercase_ : Tuple = tuple(f for f in tokenizer_r_files if 'tokenizer.json' not in f ) self.assertSequenceEqual(__UpperCamelCase ,__UpperCamelCase ) # Checks everything loads correctly in the same way lowercase_ : Any = tokenizer_r.from_pretrained(__UpperCamelCase ) lowercase_ : Optional[Any] = tokenizer_p.from_pretrained(__UpperCamelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__UpperCamelCase ,__UpperCamelCase ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(__UpperCamelCase ) # Save tokenizer rust, legacy_format=True lowercase_ : str = tempfile.mkdtemp() lowercase_ : int = tokenizer_r.save_pretrained(__UpperCamelCase ,legacy_format=__UpperCamelCase ) lowercase_ : Any = tokenizer_p.save_pretrained(__UpperCamelCase ) # Checks it save with the same files self.assertSequenceEqual(__UpperCamelCase ,__UpperCamelCase ) # Checks everything loads correctly in the same way lowercase_ : Optional[int] = tokenizer_r.from_pretrained(__UpperCamelCase ) lowercase_ : Optional[int] = tokenizer_p.from_pretrained(__UpperCamelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__UpperCamelCase ,__UpperCamelCase ) ) shutil.rmtree(__UpperCamelCase ) # Save tokenizer rust, legacy_format=False lowercase_ : Tuple = tempfile.mkdtemp() lowercase_ : Union[str, Any] = tokenizer_r.save_pretrained(__UpperCamelCase ,legacy_format=__UpperCamelCase ) lowercase_ : Optional[Any] = tokenizer_p.save_pretrained(__UpperCamelCase ) # Checks it saved the tokenizer.json file self.assertTrue(any('tokenizer.json' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way lowercase_ : Tuple = tokenizer_r.from_pretrained(__UpperCamelCase ) lowercase_ : Tuple = tokenizer_p.from_pretrained(__UpperCamelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__UpperCamelCase ,__UpperCamelCase ) ) shutil.rmtree(__UpperCamelCase ) @cached_property def _UpperCAmelCase ( self ) -> int: '''simple docstring''' return XLMRobertaTokenizer.from_pretrained('xlm-roberta-base' ) def _UpperCAmelCase ( self ) -> Dict: '''simple docstring''' with tempfile.NamedTemporaryFile() as f: shutil.copyfile(__UpperCamelCase ,f.name ) lowercase_ : int = XLMRobertaTokenizer(f.name ,keep_accents=__UpperCamelCase ) lowercase_ : Union[str, Any] = pickle.dumps(__UpperCamelCase ) pickle.loads(__UpperCamelCase ) def _UpperCAmelCase ( self ) -> Tuple: '''simple docstring''' if not self.test_rust_tokenizer: return lowercase_ : Dict = self.get_tokenizer() lowercase_ : Any = self.get_rust_tokenizer() lowercase_ : Dict = "I was born in 92000, and this is falsé." lowercase_ : Dict = tokenizer.tokenize(__UpperCamelCase ) lowercase_ : Any = rust_tokenizer.tokenize(__UpperCamelCase ) self.assertListEqual(__UpperCamelCase ,__UpperCamelCase ) lowercase_ : Union[str, Any] = tokenizer.encode(__UpperCamelCase ,add_special_tokens=__UpperCamelCase ) lowercase_ : str = rust_tokenizer.encode(__UpperCamelCase ,add_special_tokens=__UpperCamelCase ) self.assertListEqual(__UpperCamelCase ,__UpperCamelCase ) lowercase_ : Any = self.get_rust_tokenizer() lowercase_ : Dict = tokenizer.encode(__UpperCamelCase ) lowercase_ : int = rust_tokenizer.encode(__UpperCamelCase ) self.assertListEqual(__UpperCamelCase ,__UpperCamelCase ) @slow def _UpperCAmelCase ( self ) -> List[str]: '''simple docstring''' lowercase_ : List[str] = "Hello World!" lowercase_ : Any = [0, 3_5378, 6661, 38, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(__UpperCamelCase ,self.big_tokenizer.encode(__UpperCamelCase ) ) @slow def _UpperCAmelCase ( self ) -> Optional[Any]: '''simple docstring''' lowercase_ : Tuple = ( "This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will" " add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth" ) lowercase_ : str = [ 0, 3293, 83, 10, 4552, 4989, 7986, 678, 10, 5915, 111, 17_9459, 12_4850, 4, 6044, 237, 12, 6, 5, 6, 4, 6780, 705, 15, 1388, 44, 378, 1_0114, 711, 152, 20, 6, 5, 2_2376, 642, 1221, 1_5190, 3_4153, 450, 5608, 959, 1119, 5_7702, 136, 186, 47, 1098, 2_9367, 47, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 6044, 237, 6284, 5_0901, 528, 31, 90, 34, 927, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(__UpperCamelCase ,self.big_tokenizer.encode(__UpperCamelCase ) ) @slow def _UpperCAmelCase ( self ) -> Optional[int]: '''simple docstring''' lowercase_ : Union[str, Any] = {"input_ids": [[0, 1_1062, 8_2772, 7, 15, 8_2772, 538, 5_1529, 237, 1_7198, 1290, 206, 9, 21_5175, 1314, 136, 1_7198, 1290, 206, 9, 5_6359, 42, 12_2009, 9, 1_6466, 16, 8_7344, 4537, 9, 4717, 7_8381, 6, 15_9958, 7, 15, 2_4480, 618, 4, 527, 2_2693, 5428, 4, 2777, 2_4480, 9874, 4, 4_3523, 594, 4, 803, 1_8392, 3_3189, 18, 4, 4_3523, 2_4447, 1_2399, 100, 2_4955, 8_3658, 9626, 14_4057, 15, 839, 2_2335, 16, 136, 2_4955, 8_3658, 8_3479, 15, 3_9102, 724, 16, 678, 645, 2789, 1328, 4589, 42, 12_2009, 11_5774, 23, 805, 1328, 4_6876, 7, 136, 5_3894, 1940, 4_2227, 4_1159, 1_7721, 823, 425, 4, 2_7512, 9_8722, 206, 136, 5531, 4970, 919, 1_7336, 5, 2], [0, 2_0080, 618, 83, 8_2775, 47, 479, 9, 1517, 73, 5_3894, 333, 8_0581, 11_0117, 1_8811, 5256, 1295, 51, 15_2526, 297, 7986, 390, 12_4416, 538, 3_5431, 214, 98, 1_5044, 2_5737, 136, 7108, 4_3701, 23, 756, 13_5355, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 581, 6_3773, 11_9455, 6, 14_7797, 8_8203, 7, 645, 70, 21, 3285, 1_0269, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__UpperCamelCase ,model_name='xlm-roberta-base' ,revision='d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3' ,)
425
_snake_case : Optional[int] = {"a": ["c", "b"], "b": ["d", "e"], "c": [], "d": [], "e": []} _snake_case : Dict = ["a", "b", "c", "d", "e"] def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : List[str] = start # add current to visited visited.append(__lowerCamelCase ) __snake_case : List[Any] = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: __snake_case : Tuple = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # if all neighbors visited add current to sort sort.append(__lowerCamelCase ) # if all vertices haven't been visited select a new one to visit if len(__lowerCamelCase ) != len(__lowerCamelCase ): for vertice in vertices: if vertice not in visited: __snake_case : int = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # return sort return sort if __name__ == "__main__": _snake_case : List[Any] = topological_sort("a", [], []) print(sort)
81
0
'''simple docstring''' import os import pytest import yaml from datasets.features.features import Features, Value from datasets.info import DatasetInfo, DatasetInfosDict @pytest.mark.parametrize( '''files''', [ ['''full:README.md''', '''dataset_infos.json'''], ['''empty:README.md''', '''dataset_infos.json'''], ['''dataset_infos.json'''], ['''full:README.md'''], ], ) def __snake_case ( _UpperCAmelCase : str, _UpperCAmelCase : Union[str, Any]): UpperCamelCase = tmp_path_factory.mktemp('''dset_infos_dir''') if "full:README.md" in files: with open(dataset_infos_dir / '''README.md''', '''w''') as f: f.write('''---\ndataset_info:\n dataset_size: 42\n---''') if "empty:README.md" in files: with open(dataset_infos_dir / '''README.md''', '''w''') as f: f.write('''''') # we want to support dataset_infos.json for backward compatibility if "dataset_infos.json" in files: with open(dataset_infos_dir / '''dataset_infos.json''', '''w''') as f: f.write('''{\"default\": {\"dataset_size\": 42}}''') UpperCamelCase = DatasetInfosDict.from_directory(__lowerCamelCase) assert dataset_infos assert dataset_infos["default"].dataset_size == 42 @pytest.mark.parametrize( '''dataset_info''', [ DatasetInfo(), DatasetInfo( description='''foo''', features=Features({'''a''': Value('''int32''')}), builder_name='''builder''', config_name='''config''', version='''1.0.0''', splits=[{'''name''': '''train'''}], download_size=42, ), ], ) def __snake_case ( _UpperCAmelCase : Any, _UpperCAmelCase : str): UpperCamelCase = str(__lowerCamelCase) dataset_info.write_to_directory(__lowerCamelCase) UpperCamelCase = DatasetInfo.from_directory(__lowerCamelCase) assert dataset_info == reloaded assert os.path.exists(os.path.join(__lowerCamelCase, '''dataset_info.json''')) def __snake_case ( ): UpperCamelCase = DatasetInfo( description='''foo''', citation='''bar''', homepage='''https://foo.bar''', license='''CC0''', features=Features({'''a''': Value('''int32''')}), post_processed={}, supervised_keys=(), task_templates=[], builder_name='''builder''', config_name='''config''', version='''1.0.0''', splits=[{'''name''': '''train''', '''num_examples''': 42}], download_checksums={}, download_size=1337, post_processing_size=442, dataset_size=1234, size_in_bytes=1337 + 442 + 1234, ) UpperCamelCase = dataset_info._to_yaml_dict() assert sorted(__lowerCamelCase) == sorted(DatasetInfo._INCLUDED_INFO_IN_YAML) for key in DatasetInfo._INCLUDED_INFO_IN_YAML: assert key in dataset_info_yaml_dict assert isinstance(dataset_info_yaml_dict[key], (list, dict, int, str)) UpperCamelCase = yaml.safe_dump(__lowerCamelCase) UpperCamelCase = yaml.safe_load(__lowerCamelCase) assert dataset_info_yaml_dict == reloaded def __snake_case ( ): UpperCamelCase = DatasetInfo() UpperCamelCase = dataset_info._to_yaml_dict() assert dataset_info_yaml_dict == {} @pytest.mark.parametrize( '''dataset_infos_dict''', [ DatasetInfosDict(), DatasetInfosDict({'''default''': DatasetInfo()}), DatasetInfosDict({'''my_config_name''': DatasetInfo()}), DatasetInfosDict( { '''default''': DatasetInfo( description='''foo''', features=Features({'''a''': Value('''int32''')}), builder_name='''builder''', config_name='''config''', version='''1.0.0''', splits=[{'''name''': '''train'''}], download_size=42, ) }), DatasetInfosDict( { '''v1''': DatasetInfo(dataset_size=42), '''v2''': DatasetInfo(dataset_size=1337), }), ], ) def __snake_case ( _UpperCAmelCase : Optional[int], _UpperCAmelCase : List[Any]): UpperCamelCase = str(__lowerCamelCase) dataset_infos_dict.write_to_directory(__lowerCamelCase) UpperCamelCase = DatasetInfosDict.from_directory(__lowerCamelCase) # the config_name of the dataset_infos_dict take over the attribute for config_name, dataset_info in dataset_infos_dict.items(): UpperCamelCase = config_name # the yaml representation doesn't include fields like description or citation # so we just test that we can recover what we can from the yaml UpperCamelCase = DatasetInfo._from_yaml_dict(dataset_info._to_yaml_dict()) assert dataset_infos_dict == reloaded if dataset_infos_dict: assert os.path.exists(os.path.join(__lowerCamelCase, '''README.md'''))
212
def lowerCAmelCase_ ( ): return [ a * b * (1_0_0_0 - a - b) for a in range(1 , 9_9_9 ) for b in range(__lowerCamelCase , 9_9_9 ) if (a * a + b * b == (1_0_0_0 - a - b) ** 2) ][0] if __name__ == "__main__": print(f'''{solution() = }''')
81
0
import os import torch from ..logging import get_logger from .constants import FSDP_PYTORCH_VERSION, MODEL_NAME, OPTIMIZER_NAME from .versions import is_torch_version if is_torch_version('''>=''', FSDP_PYTORCH_VERSION): import torch.distributed.checkpoint as dist_cp from torch.distributed.checkpoint.default_planner import DefaultLoadPlanner, DefaultSavePlanner from torch.distributed.checkpoint.optimizer import load_sharded_optimizer_state_dict from torch.distributed.fsdp.fully_sharded_data_parallel import FullyShardedDataParallel as FSDP from torch.distributed.fsdp.fully_sharded_data_parallel import StateDictType _lowerCAmelCase : Any = get_logger(__name__) def __snake_case ( _lowerCAmelCase : List[str] , _lowerCAmelCase : List[str] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Dict , _lowerCAmelCase : List[str]=0 ) -> str: os.makedirs(__lowerCamelCase , exist_ok=__lowerCamelCase ) with FSDP.state_dict_type( __lowerCamelCase , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): A_ : Optional[Any] = model.state_dict() if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: A_ : Tuple = f"{MODEL_NAME}.bin" if model_index == 0 else f"{MODEL_NAME}_{model_index}.bin" A_ : List[str] = os.path.join(__lowerCamelCase , __lowerCamelCase ) if accelerator.process_index == 0: logger.info(f"Saving model to {output_model_file}" ) torch.save(__lowerCamelCase , __lowerCamelCase ) logger.info(f"Model saved to {output_model_file}" ) elif fsdp_plugin.state_dict_type == StateDictType.LOCAL_STATE_DICT: A_ : Optional[int] = ( f"{MODEL_NAME}_rank{accelerator.process_index}.bin" if model_index == 0 else f"{MODEL_NAME}_{model_index}_rank{accelerator.process_index}.bin" ) A_ : Tuple = os.path.join(__lowerCamelCase , __lowerCamelCase ) logger.info(f"Saving model to {output_model_file}" ) torch.save(__lowerCamelCase , __lowerCamelCase ) logger.info(f"Model saved to {output_model_file}" ) elif fsdp_plugin.state_dict_type == StateDictType.SHARDED_STATE_DICT: A_ : List[Any] = os.path.join(__lowerCamelCase , f"{MODEL_NAME}_{model_index}" ) os.makedirs(__lowerCamelCase , exist_ok=__lowerCamelCase ) logger.info(f"Saving model to {ckpt_dir}" ) A_ : Optional[int] = {"model": state_dict} dist_cp.save_state_dict( state_dict=__lowerCamelCase , storage_writer=dist_cp.FileSystemWriter(__lowerCamelCase ) , planner=DefaultSavePlanner() , ) logger.info(f"Model saved to {ckpt_dir}" ) def __snake_case ( _lowerCAmelCase : Any , _lowerCAmelCase : Dict , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any , _lowerCAmelCase : List[str]=0 ) -> List[Any]: accelerator.wait_for_everyone() with FSDP.state_dict_type( __lowerCamelCase , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: if type(__lowerCamelCase ) != FSDP and accelerator.process_index != 0: if not fsdp_plugin.sync_module_states: raise ValueError( "Set the `sync_module_states` flag to `True` so that model states are synced across processes when " "initializing FSDP object" ) return A_ : List[Any] = f"{MODEL_NAME}.bin" if model_index == 0 else f"{MODEL_NAME}_{model_index}.bin" A_ : Dict = os.path.join(__lowerCamelCase , __lowerCamelCase ) logger.info(f"Loading model from {input_model_file}" ) A_ : Dict = torch.load(__lowerCamelCase ) logger.info(f"Model loaded from {input_model_file}" ) elif fsdp_plugin.state_dict_type == StateDictType.LOCAL_STATE_DICT: A_ : Any = ( f"{MODEL_NAME}_rank{accelerator.process_index}.bin" if model_index == 0 else f"{MODEL_NAME}_{model_index}_rank{accelerator.process_index}.bin" ) A_ : List[str] = os.path.join(__lowerCamelCase , __lowerCamelCase ) logger.info(f"Loading model from {input_model_file}" ) A_ : Any = torch.load(__lowerCamelCase ) logger.info(f"Model loaded from {input_model_file}" ) elif fsdp_plugin.state_dict_type == StateDictType.SHARDED_STATE_DICT: A_ : Optional[int] = ( os.path.join(__lowerCamelCase , f"{MODEL_NAME}_{model_index}" ) if f"{MODEL_NAME}" not in input_dir else input_dir ) logger.info(f"Loading model from {ckpt_dir}" ) A_ : Optional[int] = {"model": model.state_dict()} dist_cp.load_state_dict( state_dict=__lowerCamelCase , storage_reader=dist_cp.FileSystemReader(__lowerCamelCase ) , planner=DefaultLoadPlanner() , ) A_ : int = state_dict["model"] logger.info(f"Model loaded from {ckpt_dir}" ) model.load_state_dict(__lowerCamelCase ) def __snake_case ( _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int]=0 ) -> List[Any]: os.makedirs(__lowerCamelCase , exist_ok=__lowerCamelCase ) with FSDP.state_dict_type( __lowerCamelCase , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): A_ : Dict = FSDP.optim_state_dict(__lowerCamelCase , __lowerCamelCase ) if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: if accelerator.process_index == 0: A_ : List[str] = ( f"{OPTIMIZER_NAME}.bin" if optimizer_index == 0 else f"{OPTIMIZER_NAME}_{optimizer_index}.bin" ) A_ : List[str] = os.path.join(__lowerCamelCase , __lowerCamelCase ) logger.info(f"Saving Optimizer state to {output_optimizer_file}" ) torch.save(__lowerCamelCase , __lowerCamelCase ) logger.info(f"Optimizer state saved in {output_optimizer_file}" ) else: A_ : Any = os.path.join(__lowerCamelCase , f"{OPTIMIZER_NAME}_{optimizer_index}" ) os.makedirs(__lowerCamelCase , exist_ok=__lowerCamelCase ) logger.info(f"Saving Optimizer state to {ckpt_dir}" ) dist_cp.save_state_dict( state_dict={"optimizer": optim_state} , storage_writer=dist_cp.FileSystemWriter(__lowerCamelCase ) , planner=DefaultSavePlanner() , ) logger.info(f"Optimizer state saved in {ckpt_dir}" ) def __snake_case ( _lowerCAmelCase : int , _lowerCAmelCase : Dict , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Tuple=0 ) -> List[Any]: accelerator.wait_for_everyone() with FSDP.state_dict_type( __lowerCamelCase , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: A_ : Union[str, Any] = None # below check should work but currently it isn't working (mostly opytorch issue), # in the meantime disabling it at the cost of excess memory usage # if accelerator.process_index == 0 or not fsdp_plugin.optim_state_dict_config.rank0_only: A_ : Dict = ( f"{OPTIMIZER_NAME}.bin" if optimizer_index == 0 else f"{OPTIMIZER_NAME}_{optimizer_index}.bin" ) A_ : List[Any] = os.path.join(__lowerCamelCase , __lowerCamelCase ) logger.info(f"Loading Optimizer state from {input_optimizer_file}" ) A_ : Dict = torch.load(__lowerCamelCase ) logger.info(f"Optimizer state loaded from {input_optimizer_file}" ) else: A_ : Optional[int] = ( os.path.join(__lowerCamelCase , f"{OPTIMIZER_NAME}_{optimizer_index}" ) if f"{OPTIMIZER_NAME}" not in input_dir else input_dir ) logger.info(f"Loading Optimizer from {ckpt_dir}" ) A_ : str = load_sharded_optimizer_state_dict( model_state_dict=model.state_dict() , optimizer_key="optimizer" , storage_reader=dist_cp.FileSystemReader(__lowerCamelCase ) , ) A_ : List[Any] = optim_state["optimizer"] logger.info(f"Optimizer loaded from {ckpt_dir}" ) A_ : Tuple = FSDP.optim_state_dict_to_load(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) optimizer.load_state_dict(__lowerCamelCase )
454
from math import sqrt def lowerCAmelCase_ ( __lowerCamelCase = 1_0_0_0_0_0_0 ): __snake_case : int = 0 __snake_case : int = 0 __snake_case : int while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(__lowerCamelCase , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(f'''{solution() = }''')
81
0
"""simple docstring""" import numpy as np import torch from torch.utils.data import Dataset from utils import logger class _UpperCamelCase ( _lowerCAmelCase): def __init__(self , lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" A__ = params A__ = np.array(lowerCamelCase__ ) A__ = np.array([len(lowerCamelCase__ ) for t in data] ) self.check() self.remove_long_sequences() self.remove_empty_sequences() self.remove_unknown_sequences() self.check() self.print_statistics() def __getitem__(self , lowerCamelCase__ ): """simple docstring""" return (self.token_ids[index], self.lengths[index]) def __len__(self ): """simple docstring""" return len(self.lengths ) def A (self ): """simple docstring""" assert len(self.token_ids ) == len(self.lengths ) assert all(self.lengths[i] == len(self.token_ids[i] ) for i in range(len(self.lengths ) ) ) def A (self ): """simple docstring""" A__ = self.params.max_model_input_size A__ = self.lengths > max_len logger.info(F"""Splitting {sum(lowerCamelCase__ )} too long sequences.""" ) def divide_chunks(lowerCamelCase__ , lowerCamelCase__ ): return [l[i : i + n] for i in range(0 , len(lowerCamelCase__ ) , lowerCamelCase__ )] A__ = [] A__ = [] if self.params.mlm: A__ = self.params.special_tok_ids["cls_token"], self.params.special_tok_ids["sep_token"] else: A__ = self.params.special_tok_ids["bos_token"], self.params.special_tok_ids["eos_token"] for seq_, len_ in zip(self.token_ids , self.lengths ): assert (seq_[0] == cls_id) and (seq_[-1] == sep_id), seq_ if len_ <= max_len: new_tok_ids.append(seq_ ) new_lengths.append(len_ ) else: A__ = [] for sub_s in divide_chunks(seq_ , max_len - 2 ): if sub_s[0] != cls_id: A__ = np.insert(lowerCamelCase__ , 0 , lowerCamelCase__ ) if sub_s[-1] != sep_id: A__ = np.insert(lowerCamelCase__ , len(lowerCamelCase__ ) , lowerCamelCase__ ) assert len(lowerCamelCase__ ) <= max_len assert (sub_s[0] == cls_id) and (sub_s[-1] == sep_id), sub_s sub_seqs.append(lowerCamelCase__ ) new_tok_ids.extend(lowerCamelCase__ ) new_lengths.extend([len(lowerCamelCase__ ) for l in sub_seqs] ) A__ = np.array(lowerCamelCase__ ) A__ = np.array(lowerCamelCase__ ) def A (self ): """simple docstring""" A__ = len(self ) A__ = self.lengths > 1_1 A__ = self.token_ids[indices] A__ = self.lengths[indices] A__ = len(self ) logger.info(F"""Remove {init_size - new_size} too short (<=11 tokens) sequences.""" ) def A (self ): """simple docstring""" if "unk_token" not in self.params.special_tok_ids: return else: A__ = self.params.special_tok_ids["unk_token"] A__ = len(self ) A__ = np.array([np.count_nonzero(a == unk_token_id ) for a in self.token_ids] ) A__ = (unk_occs / self.lengths) < 0.5 A__ = self.token_ids[indices] A__ = self.lengths[indices] A__ = len(self ) logger.info(F"""Remove {init_size - new_size} sequences with a high level of unknown tokens (50%).""" ) def A (self ): """simple docstring""" if not self.params.is_master: return logger.info(F"""{len(self )} sequences""" ) # data_len = sum(self.lengths) # nb_unique_tokens = len(Counter(list(chain(*self.token_ids)))) # logger.info(f'{data_len} tokens ({nb_unique_tokens} unique)') # unk_idx = self.params.special_tok_ids['unk_token'] # nb_unknown = sum([(t==unk_idx).sum() for t in self.token_ids]) # logger.info(f'{nb_unknown} unknown tokens (covering {100*nb_unknown/data_len:.2f}% of the data)') def A (self , lowerCamelCase__ ): """simple docstring""" A__ = [t[0] for t in batch] A__ = [t[1] for t in batch] assert len(lowerCamelCase__ ) == len(lowerCamelCase__ ) # Max for paddings A__ = max(lowerCamelCase__ ) # Pad token ids if self.params.mlm: A__ = self.params.special_tok_ids["pad_token"] else: A__ = self.params.special_tok_ids["unk_token"] A__ = [list(t.astype(lowerCamelCase__ ) ) + [pad_idx] * (max_seq_len_ - len(lowerCamelCase__ )) for t in token_ids] assert len(tk_ ) == len(lowerCamelCase__ ) assert all(len(lowerCamelCase__ ) == max_seq_len_ for t in tk_ ) A__ = torch.tensor(tk_ ) # (bs, max_seq_len_) A__ = torch.tensor(lowerCamelCase__ ) # (bs) return tk_t, lg_t
574
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class a (_lowerCAmelCase ): """simple docstring""" def __snake_case ( self : str ) -> str: __snake_case : Tuple = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCamelCase , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "neck_hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "num_attention_heads" ) ) class a : """simple docstring""" def __init__( self : Optional[int] , lowerCamelCase : List[str] , lowerCamelCase : Tuple=13 , lowerCamelCase : str=32 , lowerCamelCase : Dict=2 , lowerCamelCase : List[str]=3 , lowerCamelCase : Any=640 , lowerCamelCase : Optional[Any]=4 , lowerCamelCase : Tuple="silu" , lowerCamelCase : int=3 , lowerCamelCase : Dict=32 , lowerCamelCase : str=0.1 , lowerCamelCase : Optional[int]=0.1 , lowerCamelCase : Optional[Any]=0.1 , lowerCamelCase : Dict=0.02 , lowerCamelCase : Union[str, Any]=True , lowerCamelCase : Optional[int]=True , lowerCamelCase : Union[str, Any]=10 , lowerCamelCase : int=None , ) -> str: __snake_case : Optional[Any] = parent __snake_case : Optional[Any] = batch_size __snake_case : Any = image_size __snake_case : List[Any] = patch_size __snake_case : Any = num_channels __snake_case : Union[str, Any] = last_hidden_size __snake_case : Any = num_attention_heads __snake_case : Any = hidden_act __snake_case : Tuple = conv_kernel_size __snake_case : Any = output_stride __snake_case : Any = hidden_dropout_prob __snake_case : List[Any] = attention_probs_dropout_prob __snake_case : Optional[Any] = classifier_dropout_prob __snake_case : Union[str, Any] = use_labels __snake_case : Optional[int] = is_training __snake_case : Dict = num_labels __snake_case : Any = initializer_range __snake_case : Optional[int] = scope def __snake_case ( self : str ) -> Union[str, Any]: __snake_case : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : List[Any] = None __snake_case : Optional[int] = None if self.use_labels: __snake_case : Optional[int] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : Any = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __snake_case : Optional[Any] = self.get_config() return config, pixel_values, labels, pixel_labels def __snake_case ( self : Any ) -> Union[str, Any]: return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __snake_case ( self : Any , lowerCamelCase : Tuple , lowerCamelCase : int , lowerCamelCase : Tuple , lowerCamelCase : Optional[int] ) -> Dict: __snake_case : List[Any] = MobileViTModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[str] = model(lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[Any] , lowerCamelCase : List[str] , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] , lowerCamelCase : Tuple ) -> List[str]: __snake_case : str = self.num_labels __snake_case : List[Any] = MobileViTForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : Dict , lowerCamelCase : Any , lowerCamelCase : Any , lowerCamelCase : Optional[Any] , lowerCamelCase : Dict ) -> Dict: __snake_case : Union[str, Any] = self.num_labels __snake_case : Optional[int] = MobileViTForSemanticSegmentation(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Tuple = model(lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[int] ) -> List[Any]: __snake_case : Optional[Any] = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case , __snake_case : Union[str, Any] = config_and_inputs __snake_case : Dict = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) __UpperCAmelCase : Optional[Any] = ( { "feature-extraction": MobileViTModel, "image-classification": MobileViTForImageClassification, "image-segmentation": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) __UpperCAmelCase : List[str] = False __UpperCAmelCase : int = False __UpperCAmelCase : Optional[int] = False __UpperCAmelCase : Optional[int] = False def __snake_case ( self : Optional[int] ) -> Dict: __snake_case : Tuple = MobileViTModelTester(self ) __snake_case : Any = MobileViTConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase ) def __snake_case ( self : Optional[int] ) -> Dict: self.config_tester.run_common_tests() @unittest.skip(reason="MobileViT does not use inputs_embeds" ) def __snake_case ( self : Dict ) -> Any: pass @unittest.skip(reason="MobileViT does not support input and output embeddings" ) def __snake_case ( self : Dict ) -> List[Any]: pass @unittest.skip(reason="MobileViT does not output attentions" ) def __snake_case ( self : int ) -> Dict: pass def __snake_case ( self : int ) -> Union[str, Any]: __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Optional[Any] = model_class(lowerCamelCase ) __snake_case : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : int = [*signature.parameters.keys()] __snake_case : List[str] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __snake_case ( self : int ) -> Tuple: pass def __snake_case ( self : Any ) -> Tuple: __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def __snake_case ( self : Any ) -> str: def check_hidden_states_output(lowerCamelCase : Union[str, Any] , lowerCamelCase : Dict , lowerCamelCase : Any ): __snake_case : int = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : int = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : Union[str, Any] = outputs.hidden_states __snake_case : int = 5 self.assertEqual(len(lowerCamelCase ) , lowerCamelCase ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. __snake_case : List[Any] = 2 for i in range(len(lowerCamelCase ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) __snake_case , __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : str = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : List[Any] = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Any ) -> Any: __snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) def __snake_case ( self : List[str] ) -> List[str]: __snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowerCamelCase ) @slow def __snake_case ( self : List[str] ) -> Any: for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : List[str] = MobileViTModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def lowerCAmelCase_ ( ): __snake_case : Optional[int] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a (unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self : str ) -> Dict: return MobileViTImageProcessor.from_pretrained("apple/mobilevit-xx-small" ) if is_vision_available() else None @slow def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : Tuple = MobileViTForImageClassification.from_pretrained("apple/mobilevit-xx-small" ).to(lowerCamelCase ) __snake_case : Optional[Any] = self.default_image_processor __snake_case : Union[str, Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Dict = model(**lowerCamelCase ) # verify the logits __snake_case : Union[str, Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([-1.93_64, -1.23_27, -0.46_53] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : str ) -> Optional[int]: __snake_case : Optional[int] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : str = model.to(lowerCamelCase ) __snake_case : int = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Optional[int] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : List[str] = model(**lowerCamelCase ) __snake_case : Union[str, Any] = outputs.logits # verify the logits __snake_case : Tuple = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , lowerCamelCase ) __snake_case : Union[str, Any] = torch.tensor( [ [[6.97_13, 6.97_86, 7.24_22], [7.28_93, 7.28_25, 7.44_46], [7.65_80, 7.87_97, 7.94_20]], [[-10.68_69, -10.32_50, -10.34_71], [-10.42_28, -9.98_68, -9.71_32], [-11.04_05, -11.02_21, -10.73_18]], [[-3.30_89, -2.85_39, -2.67_40], [-3.27_06, -2.56_21, -2.51_08], [-3.25_34, -2.66_15, -2.66_51]], ] , device=lowerCamelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : Union[str, Any] ) -> Optional[int]: __snake_case : Optional[Any] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Tuple = model.to(lowerCamelCase ) __snake_case : Dict = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : List[Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Any = model(**lowerCamelCase ) __snake_case : Dict = outputs.logits.detach().cpu() __snake_case : Any = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase , target_sizes=[(50, 60)] ) __snake_case : int = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , lowerCamelCase ) __snake_case : List[str] = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase ) __snake_case : Optional[int] = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , lowerCamelCase )
81
0
'''simple docstring''' import argparse import os import torch from transformers import ( XLNetConfig, XLNetForQuestionAnswering, XLNetForSequenceClassification, XLNetLMHeadModel, load_tf_weights_in_xlnet, ) from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging __a: int = { "cola": 2, "mnli": 3, "mrpc": 2, "sst-2": 2, "sts-b": 1, "qqp": 2, "qnli": 2, "rte": 2, "wnli": 2, } logging.set_verbosity_info() def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=None ): # Initialise PyTorch model lowercase__ : Any = XLNetConfig.from_json_file(__lowerCamelCase ) lowercase__ : Optional[int] = finetuning_task.lower() if finetuning_task is not None else "" if finetuning_task in GLUE_TASKS_NUM_LABELS: print(F"""Building PyTorch XLNetForSequenceClassification model from configuration: {config}""" ) lowercase__ : List[Any] = finetuning_task lowercase__ : Any = GLUE_TASKS_NUM_LABELS[finetuning_task] lowercase__ : int = XLNetForSequenceClassification(__lowerCamelCase ) elif "squad" in finetuning_task: lowercase__ : Optional[Any] = finetuning_task lowercase__ : Dict = XLNetForQuestionAnswering(__lowerCamelCase ) else: lowercase__ : Tuple = XLNetLMHeadModel(__lowerCamelCase ) # Load weights from tf checkpoint load_tf_weights_in_xlnet(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # Save pytorch-model lowercase__ : List[Any] = os.path.join(__lowerCamelCase , __lowerCamelCase ) lowercase__ : Optional[Any] = os.path.join(__lowerCamelCase , __lowerCamelCase ) print(F"""Save PyTorch model to {os.path.abspath(__lowerCamelCase )}""" ) torch.save(model.state_dict() , __lowerCamelCase ) print(F"""Save configuration file to {os.path.abspath(__lowerCamelCase )}""" ) with open(__lowerCamelCase , '''w''' , encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": __a: str = argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--xlnet_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained XLNet model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the folder to store the PyTorch model or dataset/vocab.""", ) parser.add_argument( """--finetuning_task""", default=None, type=str, help="""Name of a task on which the XLNet TensorFlow model was fine-tuned""", ) __a: Optional[int] = parser.parse_args() print(args) convert_xlnet_checkpoint_to_pytorch( args.tf_checkpoint_path, args.xlnet_config_file, args.pytorch_dump_folder_path, args.finetuning_task )
152
import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor _snake_case : int = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[int] , *lowerCamelCase : List[Any] , **lowerCamelCase : int ) -> None: warnings.warn( "The class ImageGPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use ImageGPTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowerCAmelCase_ = { "configuration_m2m_100": ["M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP", "M2M100Config", "M2M100OnnxConfig"], "tokenization_m2m_100": ["M2M100Tokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ "M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST", "M2M100ForConditionalGeneration", "M2M100Model", "M2M100PreTrainedModel", ] if TYPE_CHECKING: from .configuration_mam_aaa import M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP, MaMaaaConfig, MaMaaaOnnxConfig from .tokenization_mam_aaa import MaMaaaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mam_aaa import ( M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST, MaMaaaForConditionalGeneration, MaMaaaModel, MaMaaaPreTrainedModel, ) else: import sys lowerCAmelCase_ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
326
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _snake_case : Optional[Any] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[str] = ["pixel_values"] def __init__( self : List[Any] , lowerCamelCase : bool = True , lowerCamelCase : Union[int, float] = 1 / 255 , lowerCamelCase : bool = True , lowerCamelCase : int = 8 , **lowerCamelCase : Tuple , ) -> None: super().__init__(**lowerCamelCase ) __snake_case : Dict = do_rescale __snake_case : Dict = rescale_factor __snake_case : Optional[Any] = do_pad __snake_case : Tuple = pad_size def __snake_case ( self : Dict , lowerCamelCase : np.ndarray , lowerCamelCase : float , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase : Optional[int] ) -> np.ndarray: return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def __snake_case ( self : Optional[Any] , lowerCamelCase : np.ndarray , lowerCamelCase : int , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None ) -> Tuple: __snake_case , __snake_case : List[str] = get_image_size(lowerCamelCase ) __snake_case : Optional[Any] = (old_height // size + 1) * size - old_height __snake_case : List[Any] = (old_width // size + 1) * size - old_width return pad(lowerCamelCase , ((0, pad_height), (0, pad_width)) , mode="symmetric" , data_format=lowerCamelCase ) def __snake_case ( self : Tuple , lowerCamelCase : ImageInput , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[float] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[int] = None , lowerCamelCase : Optional[Union[str, TensorType]] = None , lowerCamelCase : Union[str, ChannelDimension] = ChannelDimension.FIRST , **lowerCamelCase : Union[str, Any] , ) -> List[str]: __snake_case : int = do_rescale if do_rescale is not None else self.do_rescale __snake_case : Dict = rescale_factor if rescale_factor is not None else self.rescale_factor __snake_case : str = do_pad if do_pad is not None else self.do_pad __snake_case : Any = pad_size if pad_size is not None else self.pad_size __snake_case : int = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) # All transformations expect numpy arrays. __snake_case : str = [to_numpy_array(lowerCamelCase ) for image in images] if do_rescale: __snake_case : Optional[int] = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_pad: __snake_case : Optional[Any] = [self.pad(lowerCamelCase , size=lowerCamelCase ) for image in images] __snake_case : int = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __snake_case : Union[str, Any] = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
81
0
"""simple docstring""" import inspect import unittest import warnings from math import ceil, floor from transformers import LevitConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_MAPPING, LevitForImageClassification, LevitForImageClassificationWithTeacher, LevitModel, ) from transformers.models.levit.modeling_levit import LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class SCREAMING_SNAKE_CASE_ ( _lowerCAmelCase ): """simple docstring""" def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.config_class(**self.inputs_dict) self.parent.assertTrue(hasattr(lowerCAmelCase__ , """hidden_sizes""")) self.parent.assertTrue(hasattr(lowerCAmelCase__ , """num_attention_heads""")) class SCREAMING_SNAKE_CASE_ : """simple docstring""" def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=1_3 , lowerCAmelCase__=6_4 , lowerCAmelCase__=3 , lowerCAmelCase__=3 , lowerCAmelCase__=2 , lowerCAmelCase__=1 , lowerCAmelCase__=1_6 , lowerCAmelCase__=[1_2_8, 2_5_6, 3_8_4] , lowerCAmelCase__=[4, 6, 8] , lowerCAmelCase__=[2, 3, 4] , lowerCAmelCase__=[1_6, 1_6, 1_6] , lowerCAmelCase__=0 , lowerCAmelCase__=[2, 2, 2] , lowerCAmelCase__=[2, 2, 2] , lowerCAmelCase__=0.02 , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=2 , ): __SCREAMING_SNAKE_CASE = parent __SCREAMING_SNAKE_CASE = batch_size __SCREAMING_SNAKE_CASE = image_size __SCREAMING_SNAKE_CASE = num_channels __SCREAMING_SNAKE_CASE = kernel_size __SCREAMING_SNAKE_CASE = stride __SCREAMING_SNAKE_CASE = padding __SCREAMING_SNAKE_CASE = hidden_sizes __SCREAMING_SNAKE_CASE = num_attention_heads __SCREAMING_SNAKE_CASE = depths __SCREAMING_SNAKE_CASE = key_dim __SCREAMING_SNAKE_CASE = drop_path_rate __SCREAMING_SNAKE_CASE = patch_size __SCREAMING_SNAKE_CASE = attention_ratio __SCREAMING_SNAKE_CASE = mlp_ratio __SCREAMING_SNAKE_CASE = initializer_range __SCREAMING_SNAKE_CASE = [ ["Subsample", key_dim[0], hidden_sizes[0] // key_dim[0], 4, 2, 2], ["Subsample", key_dim[0], hidden_sizes[1] // key_dim[0], 4, 2, 2], ] __SCREAMING_SNAKE_CASE = is_training __SCREAMING_SNAKE_CASE = use_labels __SCREAMING_SNAKE_CASE = num_labels __SCREAMING_SNAKE_CASE = initializer_range def snake_case_ ( self): __SCREAMING_SNAKE_CASE = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) __SCREAMING_SNAKE_CASE = None if self.use_labels: __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.num_labels) __SCREAMING_SNAKE_CASE = self.get_config() return config, pixel_values, labels def snake_case_ ( self): return LevitConfig( image_size=self.image_size , num_channels=self.num_channels , kernel_size=self.kernel_size , stride=self.stride , padding=self.padding , patch_size=self.patch_size , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , depths=self.depths , key_dim=self.key_dim , drop_path_rate=self.drop_path_rate , mlp_ratio=self.mlp_ratio , attention_ratio=self.attention_ratio , initializer_range=self.initializer_range , down_ops=self.down_ops , ) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = LevitModel(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = (self.image_size, self.image_size) __SCREAMING_SNAKE_CASE = image_size[0], image_size[1] for _ in range(4): __SCREAMING_SNAKE_CASE = floor(((height + 2 * self.padding - self.kernel_size) / self.stride) + 1) __SCREAMING_SNAKE_CASE = floor(((width + 2 * self.padding - self.kernel_size) / self.stride) + 1) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, ceil(height / 4) * ceil(width / 4), self.hidden_sizes[-1]) , ) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = self.num_labels __SCREAMING_SNAKE_CASE = LevitForImageClassification(lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , labels=lowerCAmelCase__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() __SCREAMING_SNAKE_CASE = config_and_inputs __SCREAMING_SNAKE_CASE = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE_ ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __lowercase : str = ( (LevitModel, LevitForImageClassification, LevitForImageClassificationWithTeacher) if is_torch_available() else () ) __lowercase : Optional[int] = ( { "feature-extraction": LevitModel, "image-classification": (LevitForImageClassification, LevitForImageClassificationWithTeacher), } if is_torch_available() else {} ) __lowercase : Dict = False __lowercase : Optional[Any] = False __lowercase : int = False __lowercase : Union[str, Any] = False __lowercase : Tuple = False def snake_case_ ( self): __SCREAMING_SNAKE_CASE = LevitModelTester(self) __SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=lowerCAmelCase__ , has_text_modality=lowerCAmelCase__ , hidden_size=3_7) def snake_case_ ( self): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def snake_case_ ( self): return @unittest.skip(reason="""Levit does not use inputs_embeds""") def snake_case_ ( self): pass @unittest.skip(reason="""Levit does not support input and output embeddings""") def snake_case_ ( self): pass @unittest.skip(reason="""Levit does not output attentions""") def snake_case_ ( self): pass def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __SCREAMING_SNAKE_CASE = model_class(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic __SCREAMING_SNAKE_CASE = [*signature.parameters.keys()] __SCREAMING_SNAKE_CASE = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCAmelCase__) def snake_case_ ( self): def check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = model_class(lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(**self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__)) __SCREAMING_SNAKE_CASE = outputs.hidden_states __SCREAMING_SNAKE_CASE = len(self.model_tester.depths) + 1 self.assertEqual(len(lowerCAmelCase__) , lowerCAmelCase__) __SCREAMING_SNAKE_CASE = (self.model_tester.image_size, self.model_tester.image_size) __SCREAMING_SNAKE_CASE = image_size[0], image_size[1] for _ in range(4): __SCREAMING_SNAKE_CASE = floor( ( (height + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1) __SCREAMING_SNAKE_CASE = floor( ( (width + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-2:]) , [ height * width, self.model_tester.hidden_sizes[0], ] , ) __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __SCREAMING_SNAKE_CASE = True check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __SCREAMING_SNAKE_CASE = True check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__) @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""") def snake_case_ ( self): pass def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=False): __SCREAMING_SNAKE_CASE = super()._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ , return_labels=lowerCAmelCase__) if return_labels: if model_class.__name__ == "LevitForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase__) def snake_case_ ( self): if not self.model_tester.is_training: return __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() __SCREAMING_SNAKE_CASE = True for model_class in self.all_model_classes: # LevitForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(lowerCAmelCase__) or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue __SCREAMING_SNAKE_CASE = model_class(lowerCAmelCase__) model.to(lowerCAmelCase__) model.train() __SCREAMING_SNAKE_CASE = self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ , return_labels=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = model(**lowerCAmelCase__).loss loss.backward() def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = True for model_class in self.all_model_classes: if model_class in get_values(lowerCAmelCase__) or not model_class.supports_gradient_checkpointing: continue # LevitForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "LevitForImageClassificationWithTeacher": continue __SCREAMING_SNAKE_CASE = model_class(lowerCAmelCase__) model.gradient_checkpointing_enable() model.to(lowerCAmelCase__) model.train() __SCREAMING_SNAKE_CASE = self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ , return_labels=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = model(**lowerCAmelCase__).loss loss.backward() def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() __SCREAMING_SNAKE_CASE = [ {"title": "multi_label_classification", "num_labels": 2, "dtype": torch.float}, {"title": "single_label_classification", "num_labels": 1, "dtype": torch.long}, {"title": "regression", "num_labels": 1, "dtype": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(lowerCAmelCase__), ] or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=f"Testing {model_class} with {problem_type['title']}"): __SCREAMING_SNAKE_CASE = problem_type["title"] __SCREAMING_SNAKE_CASE = problem_type["num_labels"] __SCREAMING_SNAKE_CASE = model_class(lowerCAmelCase__) model.to(lowerCAmelCase__) model.train() __SCREAMING_SNAKE_CASE = self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ , return_labels=lowerCAmelCase__) if problem_type["num_labels"] > 1: __SCREAMING_SNAKE_CASE = inputs["labels"].unsqueeze(1).repeat(1 , problem_type["""num_labels"""]) __SCREAMING_SNAKE_CASE = inputs["labels"].to(problem_type["""dtype"""]) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=lowerCAmelCase__) as warning_list: __SCREAMING_SNAKE_CASE = model(**lowerCAmelCase__).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message): raise ValueError( f"Something is going wrong in the regression problem: intercepted {w.message}") loss.backward() @slow def snake_case_ ( self): for model_name in LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __SCREAMING_SNAKE_CASE = LevitModel.from_pretrained(lowerCAmelCase__) self.assertIsNotNone(lowerCAmelCase__) def _lowerCAmelCase ( ): __SCREAMING_SNAKE_CASE = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): """simple docstring""" @cached_property def snake_case_ ( self): return LevitImageProcessor.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0]) @slow def snake_case_ ( self): __SCREAMING_SNAKE_CASE = LevitForImageClassificationWithTeacher.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0]).to( lowerCAmelCase__) __SCREAMING_SNAKE_CASE = self.default_image_processor __SCREAMING_SNAKE_CASE = prepare_img() __SCREAMING_SNAKE_CASE = image_processor(images=lowerCAmelCase__ , return_tensors="""pt""").to(lowerCAmelCase__) # forward pass with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(**lowerCAmelCase__) # verify the logits __SCREAMING_SNAKE_CASE = torch.Size((1, 1_0_0_0)) self.assertEqual(outputs.logits.shape , lowerCAmelCase__) __SCREAMING_SNAKE_CASE = torch.tensor([1.04_48, -0.37_45, -1.83_17]).to(lowerCAmelCase__) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCAmelCase__ , atol=1E-4))
155
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case : Any = { "configuration_informer": [ "INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "InformerConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "InformerForPrediction", "InformerModel", "InformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys _snake_case : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
0
import gc import unittest import numpy as np import torch from diffusers import AutoencoderKL, DDIMScheduler, DiTPipeline, DPMSolverMultistepScheduler, TransformeraDModel from diffusers.utils import is_xformers_available, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( CLASS_CONDITIONED_IMAGE_GENERATION_BATCH_PARAMS, CLASS_CONDITIONED_IMAGE_GENERATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class lowerCAmelCase_ ( _lowerCAmelCase , unittest.TestCase ): UpperCamelCase_ :str = DiTPipeline UpperCamelCase_ :Union[str, Any] = CLASS_CONDITIONED_IMAGE_GENERATION_PARAMS UpperCamelCase_ :List[Any] = PipelineTesterMixin.required_optional_params - { "latents", "num_images_per_prompt", "callback", "callback_steps", } UpperCamelCase_ :Optional[int] = CLASS_CONDITIONED_IMAGE_GENERATION_BATCH_PARAMS UpperCamelCase_ :Tuple = False def __snake_case ( self : int ): torch.manual_seed(0 ) lowerCAmelCase__ = TransformeraDModel( sample_size=16 , num_layers=2 , patch_size=4 , attention_head_dim=8 , num_attention_heads=2 , in_channels=4 , out_channels=8 , attention_bias=SCREAMING_SNAKE_CASE_ , activation_fn='''gelu-approximate''' , num_embeds_ada_norm=1_000 , norm_type='''ada_norm_zero''' , norm_elementwise_affine=SCREAMING_SNAKE_CASE_ , ) lowerCAmelCase__ = AutoencoderKL() lowerCAmelCase__ = DDIMScheduler() lowerCAmelCase__ = {"transformer": transformer.eval(), "vae": vae.eval(), "scheduler": scheduler} return components def __snake_case ( self : List[str] , SCREAMING_SNAKE_CASE_ : Any , SCREAMING_SNAKE_CASE_ : List[Any]=0 ): if str(SCREAMING_SNAKE_CASE_ ).startswith('''mps''' ): lowerCAmelCase__ = torch.manual_seed(SCREAMING_SNAKE_CASE_ ) else: lowerCAmelCase__ = torch.Generator(device=SCREAMING_SNAKE_CASE_ ).manual_seed(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ = { "class_labels": [1], "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs def __snake_case ( self : List[Any] ): lowerCAmelCase__ = "cpu" lowerCAmelCase__ = self.get_dummy_components() lowerCAmelCase__ = self.pipeline_class(**SCREAMING_SNAKE_CASE_ ) pipe.to(SCREAMING_SNAKE_CASE_ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ = self.get_dummy_inputs(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ = pipe(**SCREAMING_SNAKE_CASE_ ).images lowerCAmelCase__ = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 16, 16, 3) ) lowerCAmelCase__ = np.array([0.2_946, 0.6_601, 0.4_329, 0.3_296, 0.4_144, 0.5_319, 0.7_273, 0.5_013, 0.4_457] ) lowerCAmelCase__ = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(SCREAMING_SNAKE_CASE_ , 1e-3 ) def __snake_case ( self : Any ): self._test_inference_batch_single_identical(relax_max_difference=SCREAMING_SNAKE_CASE_ , expected_max_diff=1e-3 ) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def __snake_case ( self : Optional[Any] ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) @require_torch_gpu @slow class lowerCAmelCase_ ( unittest.TestCase ): def __snake_case ( self : Optional[Any] ): super().tearDown() gc.collect() torch.cuda.empty_cache() def __snake_case ( self : List[Any] ): lowerCAmelCase__ = torch.manual_seed(0 ) lowerCAmelCase__ = DiTPipeline.from_pretrained('''facebook/DiT-XL-2-256''' ) pipe.to('''cuda''' ) lowerCAmelCase__ = ["vase", "umbrella", "white shark", "white wolf"] lowerCAmelCase__ = pipe.get_label_ids(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ = pipe(SCREAMING_SNAKE_CASE_ , generator=SCREAMING_SNAKE_CASE_ , num_inference_steps=40 , output_type='''np''' ).images for word, image in zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowerCAmelCase__ = load_numpy( f'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/dit/{word}.npy' ) assert np.abs((expected_image - image).max() ) < 1e-2 def __snake_case ( self : Dict ): lowerCAmelCase__ = DiTPipeline.from_pretrained('''facebook/DiT-XL-2-512''' ) lowerCAmelCase__ = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.to('''cuda''' ) lowerCAmelCase__ = ["vase", "umbrella"] lowerCAmelCase__ = pipe.get_label_ids(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ = torch.manual_seed(0 ) lowerCAmelCase__ = pipe(SCREAMING_SNAKE_CASE_ , generator=SCREAMING_SNAKE_CASE_ , num_inference_steps=25 , output_type='''np''' ).images for word, image in zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowerCAmelCase__ = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' f'/dit/{word}_512.npy' ) assert np.abs((expected_image - image).max() ) < 1e-1
668
from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging _snake_case : Any = logging.get_logger(__name__) # pylint: disable=invalid-name class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[Any] , lowerCamelCase : CLIPSegForImageSegmentation , lowerCamelCase : CLIPSegProcessor , lowerCamelCase : AutoencoderKL , lowerCamelCase : CLIPTextModel , lowerCamelCase : CLIPTokenizer , lowerCamelCase : UNetaDConditionModel , lowerCamelCase : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , lowerCamelCase : StableDiffusionSafetyChecker , lowerCamelCase : CLIPImageProcessor , ) -> Tuple: super().__init__() if hasattr(scheduler.config , "steps_offset" ) and scheduler.config.steps_offset != 1: __snake_case : Tuple = ( F'The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`' F' should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure ' "to update the config accordingly as leaving `steps_offset` might led to incorrect results" " in future versions. If you have downloaded this checkpoint from the Hugging Face Hub," " it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`" " file" ) deprecate("steps_offset!=1" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : Any = dict(scheduler.config ) __snake_case : List[Any] = 1 __snake_case : Tuple = FrozenDict(lowerCamelCase ) if hasattr(scheduler.config , "skip_prk_steps" ) and scheduler.config.skip_prk_steps is False: __snake_case : List[str] = ( F'The configuration file of this scheduler: {scheduler} has not set the configuration' " `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make" " sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to" " incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face" " Hub, it would be very nice if you could open a Pull request for the" " `scheduler/scheduler_config.json` file" ) deprecate("skip_prk_steps not set" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : List[str] = dict(scheduler.config ) __snake_case : List[str] = True __snake_case : Any = FrozenDict(lowerCamelCase ) if safety_checker is None: logger.warning( F'You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure' " that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered" " results in services or applications open to the public. Both the diffusers team and Hugging Face" " strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling" " it only for use-cases that involve analyzing network behavior or auditing its results. For more" " information, please have a look at https://github.com/huggingface/diffusers/pull/254 ." ) self.register_modules( segmentation_model=lowerCamelCase , segmentation_processor=lowerCamelCase , vae=lowerCamelCase , text_encoder=lowerCamelCase , tokenizer=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase , safety_checker=lowerCamelCase , feature_extractor=lowerCamelCase , ) def __snake_case ( self : Dict , lowerCamelCase : Optional[Union[str, int]] = "auto" ) -> Tuple: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __snake_case : Any = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowerCamelCase ) def __snake_case ( self : List[Any] ) -> Any: self.enable_attention_slicing(lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> str: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) __snake_case : Optional[int] = torch.device("cuda" ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(lowerCamelCase , lowerCamelCase ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __snake_case ( self : int ) -> Any: if self.device != torch.device("meta" ) or not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowerCamelCase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self : List[Any] , lowerCamelCase : Union[str, List[str]] , lowerCamelCase : Union[torch.FloatTensor, PIL.Image.Image] , lowerCamelCase : str , lowerCamelCase : int = 512 , lowerCamelCase : int = 512 , lowerCamelCase : int = 50 , lowerCamelCase : float = 7.5 , lowerCamelCase : Optional[Union[str, List[str]]] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : float = 0.0 , lowerCamelCase : Optional[torch.Generator] = None , lowerCamelCase : Optional[torch.FloatTensor] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , lowerCamelCase : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , lowerCamelCase : int = 1 , **lowerCamelCase : Dict , ) -> List[str]: __snake_case : Tuple = self.segmentation_processor( text=[text] , images=[image] , padding="max_length" , return_tensors="pt" ).to(self.device ) __snake_case : str = self.segmentation_model(**lowerCamelCase ) __snake_case : Dict = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() __snake_case : List[Any] = self.numpy_to_pil(lowerCamelCase )[0].resize(image.size ) # Run inpainting pipeline with the generated mask __snake_case : Tuple = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=lowerCamelCase , image=lowerCamelCase , mask_image=lowerCamelCase , height=lowerCamelCase , width=lowerCamelCase , num_inference_steps=lowerCamelCase , guidance_scale=lowerCamelCase , negative_prompt=lowerCamelCase , num_images_per_prompt=lowerCamelCase , eta=lowerCamelCase , generator=lowerCamelCase , latents=lowerCamelCase , output_type=lowerCamelCase , return_dict=lowerCamelCase , callback=lowerCamelCase , callback_steps=lowerCamelCase , )
81
0
'''simple docstring''' import requests __UpperCamelCase = "YOUR API KEY" def _a ( _lowerCamelCase , _lowerCamelCase = giphy_api_key ) -> Optional[int]: """simple docstring""" __snake_case : Optional[int] = "+".join(query.split() ) __snake_case : List[Any] = F'''https://api.giphy.com/v1/gifs/search?q={formatted_query}&api_key={api_key}''' __snake_case : Any = requests.get(__lowerCamelCase ).json()["data"] return [gif["url"] for gif in gifs] if __name__ == "__main__": print("\n".join(get_gifs("space ship")))
26
class a : """simple docstring""" def __init__( self : Tuple , lowerCamelCase : list ) -> None: __snake_case : str = set_counts __snake_case : Union[str, Any] = max(lowerCamelCase ) __snake_case : List[Any] = len(lowerCamelCase ) __snake_case : Tuple = [1] * num_sets __snake_case : Dict = list(range(lowerCamelCase ) ) def __snake_case ( self : str , lowerCamelCase : int , lowerCamelCase : int ) -> bool: __snake_case : List[Any] = self.get_parent(lowerCamelCase ) __snake_case : Tuple = self.get_parent(lowerCamelCase ) if src_parent == dst_parent: return False if self.ranks[dst_parent] >= self.ranks[src_parent]: self.set_counts[dst_parent] += self.set_counts[src_parent] __snake_case : List[str] = 0 __snake_case : List[Any] = dst_parent if self.ranks[dst_parent] == self.ranks[src_parent]: self.ranks[dst_parent] += 1 __snake_case : Dict = self.set_counts[dst_parent] else: self.set_counts[src_parent] += self.set_counts[dst_parent] __snake_case : Union[str, Any] = 0 __snake_case : Optional[int] = src_parent __snake_case : Tuple = self.set_counts[src_parent] __snake_case : str = max(self.max_set , lowerCamelCase ) return True def __snake_case ( self : int , lowerCamelCase : int ) -> int: if self.parents[disj_set] == disj_set: return disj_set __snake_case : Optional[int] = self.get_parent(self.parents[disj_set] ) return self.parents[disj_set]
81
0
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = { "microsoft/git-base": "https://huggingface.co/microsoft/git-base/resolve/main/config.json", } class lowerCAmelCase ( _lowerCAmelCase ): lowerCAmelCase_ = "git_vision_model" def __init__( self : Tuple , __lowercase : str=768 , __lowercase : Optional[int]=3072 , __lowercase : List[Any]=12 , __lowercase : int=12 , __lowercase : Optional[Any]=3 , __lowercase : Tuple=224 , __lowercase : List[Any]=16 , __lowercase : Union[str, Any]="quick_gelu" , __lowercase : List[str]=1E-5 , __lowercase : str=0.0 , __lowercase : Union[str, Any]=0.0_2 , **__lowercase : Union[str, Any] , ): """simple docstring""" super().__init__(**__lowercase ) __lowercase =hidden_size __lowercase =intermediate_size __lowercase =num_hidden_layers __lowercase =num_attention_heads __lowercase =num_channels __lowercase =patch_size __lowercase =image_size __lowercase =initializer_range __lowercase =attention_dropout __lowercase =layer_norm_eps __lowercase =hidden_act @classmethod def snake_case ( cls : Optional[Any] , __lowercase : Union[str, os.PathLike] , **__lowercase : int ): """simple docstring""" cls._set_token_in_kwargs(__lowercase ) __lowercase =cls.get_config_dict(__lowercase , **__lowercase ) # get the vision config dict if we are loading from GITConfig if config_dict.get('model_type' ) == "git": __lowercase =config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__lowercase , **__lowercase ) class lowerCAmelCase ( _lowerCAmelCase ): lowerCAmelCase_ = "git" def __init__( self : Tuple , __lowercase : int=None , __lowercase : List[str]=30522 , __lowercase : Tuple=768 , __lowercase : Any=6 , __lowercase : Any=12 , __lowercase : str=3072 , __lowercase : List[str]="gelu" , __lowercase : Optional[Any]=0.1 , __lowercase : List[str]=0.1 , __lowercase : Dict=1024 , __lowercase : List[str]=0.0_2 , __lowercase : Optional[int]=1E-12 , __lowercase : Optional[int]=0 , __lowercase : int="absolute" , __lowercase : Optional[Any]=True , __lowercase : Union[str, Any]=False , __lowercase : List[str]=101 , __lowercase : Dict=102 , __lowercase : List[str]=None , **__lowercase : int , ): """simple docstring""" super().__init__(bos_token_id=__lowercase , eos_token_id=__lowercase , pad_token_id=__lowercase , **__lowercase ) if vision_config is None: __lowercase ={} logger.info('vision_config is None. initializing the GitVisionConfig with default values.' ) __lowercase =GitVisionConfig(**__lowercase ) __lowercase =vocab_size __lowercase =hidden_size __lowercase =num_hidden_layers __lowercase =num_attention_heads __lowercase =hidden_act __lowercase =intermediate_size __lowercase =hidden_dropout_prob __lowercase =attention_probs_dropout_prob __lowercase =max_position_embeddings __lowercase =initializer_range __lowercase =layer_norm_eps __lowercase =position_embedding_type __lowercase =use_cache __lowercase =tie_word_embeddings __lowercase =num_image_with_embedding __lowercase =bos_token_id __lowercase =eos_token_id def snake_case ( self : List[Any] ): """simple docstring""" __lowercase =copy.deepcopy(self.__dict__ ) __lowercase =self.vision_config.to_dict() __lowercase =self.__class__.model_type return output
119
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class a (unittest.TestCase ): """simple docstring""" def __init__( self : int , lowerCamelCase : Any , lowerCamelCase : int=7 , lowerCamelCase : str=3 , lowerCamelCase : Optional[int]=30 , lowerCamelCase : Dict=400 , lowerCamelCase : str=True , lowerCamelCase : str=None , lowerCamelCase : Any=True , lowerCamelCase : Union[str, Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[str]=True , lowerCamelCase : Optional[int]=1 / 255 , lowerCamelCase : Any=True , ) -> str: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __snake_case : Optional[Any] = size if size is not None else {"shortest_edge": 18, "longest_edge": 1333} __snake_case : Optional[Any] = parent __snake_case : List[Any] = batch_size __snake_case : Optional[int] = num_channels __snake_case : str = min_resolution __snake_case : int = max_resolution __snake_case : int = do_resize __snake_case : Tuple = size __snake_case : Any = do_normalize __snake_case : int = image_mean __snake_case : Tuple = image_std __snake_case : Dict = do_rescale __snake_case : Optional[Any] = rescale_factor __snake_case : str = do_pad def __snake_case ( self : Any ) -> int: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def __snake_case ( self : Dict , lowerCamelCase : List[Any] , lowerCamelCase : Optional[Any]=False ) -> List[str]: if not batched: __snake_case : Dict = image_inputs[0] if isinstance(lowerCamelCase , Image.Image ): __snake_case , __snake_case : Dict = image.size else: __snake_case , __snake_case : List[str] = image.shape[1], image.shape[2] if w < h: __snake_case : Optional[int] = int(self.size["shortest_edge"] * h / w ) __snake_case : int = self.size["shortest_edge"] elif w > h: __snake_case : List[str] = self.size["shortest_edge"] __snake_case : Optional[Any] = int(self.size["shortest_edge"] * w / h ) else: __snake_case : List[Any] = self.size["shortest_edge"] __snake_case : Any = self.size["shortest_edge"] else: __snake_case : int = [] for image in image_inputs: __snake_case , __snake_case : List[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[0] )[0] __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class a (_lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ConditionalDetrImageProcessor if is_vision_available() else None def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : Optional[Any] = ConditionalDetrImageProcessingTester(self ) @property def __snake_case ( self : Any ) -> str: return self.image_processor_tester.prepare_image_processor_dict() def __snake_case ( self : Optional[Any] ) -> Optional[int]: __snake_case : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase , "image_mean" ) ) self.assertTrue(hasattr(lowerCamelCase , "image_std" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_normalize" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase , "size" ) ) def __snake_case ( self : Any ) -> Dict: __snake_case : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 1333} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) __snake_case : str = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=lowerCamelCase ) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> Dict: pass def __snake_case ( self : Tuple ) -> str: # Initialize image_processing __snake_case : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , Image.Image ) # Test not batched input __snake_case : int = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) __snake_case : Dict = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> str: # Initialize image_processing __snake_case : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __snake_case : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , numpify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , np.ndarray ) # Test not batched input __snake_case : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : List[Any] = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> List[str]: # Initialize image_processing __snake_case : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , torchify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , torch.Tensor ) # Test not batched input __snake_case : Optional[Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : List[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : int = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def __snake_case ( self : Any ) -> Optional[int]: # prepare image and target __snake_case : List[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : List[Any] = {"image_id": 39769, "annotations": target} # encode them __snake_case : List[str] = ConditionalDetrImageProcessor.from_pretrained("microsoft/conditional-detr-resnet-50" ) __snake_case : List[str] = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : Tuple = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Optional[int] = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : List[Any] = torch.tensor([58_87.96_00, 1_12_50.20_61, 48_93_53.84_38, 83_71_22.75_00, 14_79_67.51_56, 16_57_32.34_38] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : Optional[int] = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([0.55_03, 0.27_65, 0.06_04, 0.22_15] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Optional[Any] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : List[Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : Dict = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify orig_size __snake_case : int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Tuple = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) ) @slow def __snake_case ( self : str ) -> Tuple: # prepare image, target and masks_path __snake_case : Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : str = {"file_name": "000000039769.png", "image_id": 39769, "segments_info": target} __snake_case : Optional[int] = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them __snake_case : int = ConditionalDetrImageProcessor(format="coco_panoptic" ) __snake_case : str = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , masks_path=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : List[str] = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Dict = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : Any = torch.tensor([14_79_79.68_75, 16_55_27.04_69, 48_46_38.59_38, 1_12_92.93_75, 58_79.65_62, 76_34.11_47] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : str = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : Optional[Any] = torch.tensor([0.26_25, 0.54_37, 0.46_88, 0.86_25] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Tuple = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : Dict = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : int = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify masks __snake_case : List[Any] = 822873 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , lowerCamelCase ) # verify orig_size __snake_case : List[str] = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Any = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) )
81
0
SCREAMING_SNAKE_CASE__ : Optional[int] = {"a": ["c", "b"], "b": ["d", "e"], "c": [], "d": [], "e": []} SCREAMING_SNAKE_CASE__ : Dict = ["a", "b", "c", "d", "e"] def __magic_name__ ( __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : Tuple ) -> str: __lowerCamelCase = start # add current to visited visited.append(__lowerCamelCase ) __lowerCamelCase = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: __lowerCamelCase = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # if all neighbors visited add current to sort sort.append(__lowerCamelCase ) # if all vertices haven't been visited select a new one to visit if len(__lowerCamelCase ) != len(__lowerCamelCase ): for vertice in vertices: if vertice not in visited: __lowerCamelCase = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # return sort return sort if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : List[Any] = topological_sort("a", [], []) print(sort)
298
from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from ... import AutoBackbone from ...modeling_outputs import SemanticSegmenterOutput from ...modeling_utils import PreTrainedModel from ...utils import add_start_docstrings, add_start_docstrings_to_model_forward, replace_return_docstrings from ...utils.backbone_utils import BackboneMixin from .configuration_upernet import UperNetConfig _snake_case : Optional[Any] = [ "openmmlab/upernet-convnext-tiny", # See all UperNet models at https://huggingface.co/models?filter=upernet ] # General docstring _snake_case : Dict = "UperNetConfig" class a (nn.Module ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : Union[int, Tuple[int, int]] , lowerCamelCase : Union[int, Tuple[int, int], str] = 0 , lowerCamelCase : bool = False , lowerCamelCase : Union[int, Tuple[int, int]] = 1 , ) -> None: super().__init__() __snake_case : Union[str, Any] = nn.Convad( in_channels=lowerCamelCase , out_channels=lowerCamelCase , kernel_size=lowerCamelCase , padding=lowerCamelCase , bias=lowerCamelCase , dilation=lowerCamelCase , ) __snake_case : Dict = nn.BatchNormad(lowerCamelCase ) __snake_case : List[Any] = nn.ReLU() def __snake_case ( self : List[Any] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : Dict = self.conv(lowerCamelCase ) __snake_case : int = self.batch_norm(lowerCamelCase ) __snake_case : Optional[Any] = self.activation(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : str , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : int ) -> None: super().__init__() __snake_case : Tuple = [ nn.AdaptiveAvgPoolad(lowerCamelCase ), UperNetConvModule(lowerCamelCase , lowerCamelCase , kernel_size=1 ), ] for i, layer in enumerate(self.layers ): self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : Dict , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : List[str] = input for layer in self.layers: __snake_case : Tuple = layer(lowerCamelCase ) return hidden_state class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Tuple[int, ...] , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : bool ) -> None: super().__init__() __snake_case : Dict = pool_scales __snake_case : List[str] = align_corners __snake_case : List[Any] = in_channels __snake_case : str = channels __snake_case : Optional[Any] = [] for i, pool_scale in enumerate(lowerCamelCase ): __snake_case : Union[str, Any] = UperNetPyramidPoolingBlock(pool_scale=lowerCamelCase , in_channels=lowerCamelCase , channels=lowerCamelCase ) self.blocks.append(lowerCamelCase ) self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> List[torch.Tensor]: __snake_case : Tuple = [] for ppm in self.blocks: __snake_case : Any = ppm(lowerCamelCase ) __snake_case : List[Any] = nn.functional.interpolate( lowerCamelCase , size=x.size()[2:] , mode="bilinear" , align_corners=self.align_corners ) ppm_outs.append(lowerCamelCase ) return ppm_outs class a (nn.Module ): """simple docstring""" def __init__( self : int , lowerCamelCase : str , lowerCamelCase : str ) -> List[str]: super().__init__() __snake_case : Dict = config __snake_case : List[str] = config.pool_scales # e.g. (1, 2, 3, 6) __snake_case : Tuple = in_channels __snake_case : str = config.hidden_size __snake_case : List[str] = False __snake_case : Any = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) # PSP Module __snake_case : Tuple = UperNetPyramidPoolingModule( self.pool_scales , self.in_channels[-1] , self.channels , align_corners=self.align_corners , ) __snake_case : List[str] = UperNetConvModule( self.in_channels[-1] + len(self.pool_scales ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) # FPN Module __snake_case : List[Any] = nn.ModuleList() __snake_case : Dict = nn.ModuleList() for in_channels in self.in_channels[:-1]: # skip the top layer __snake_case : Union[str, Any] = UperNetConvModule(lowerCamelCase , self.channels , kernel_size=1 ) __snake_case : Optional[int] = UperNetConvModule(self.channels , self.channels , kernel_size=3 , padding=1 ) self.lateral_convs.append(lowerCamelCase ) self.fpn_convs.append(lowerCamelCase ) __snake_case : int = UperNetConvModule( len(self.in_channels ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) def __snake_case ( self : List[str] ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> str: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : List[Any] , lowerCamelCase : Tuple ) -> Optional[int]: __snake_case : str = inputs[-1] __snake_case : int = [x] psp_outs.extend(self.psp_modules(lowerCamelCase ) ) __snake_case : Tuple = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Union[str, Any] = self.bottleneck(lowerCamelCase ) return output def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # build laterals __snake_case : Any = [lateral_conv(encoder_hidden_states[i] ) for i, lateral_conv in enumerate(self.lateral_convs )] laterals.append(self.psp_forward(lowerCamelCase ) ) # build top-down path __snake_case : Dict = len(lowerCamelCase ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Union[str, Any] = laterals[i - 1].shape[2:] __snake_case : Optional[int] = laterals[i - 1] + nn.functional.interpolate( laterals[i] , size=lowerCamelCase , mode="bilinear" , align_corners=self.align_corners ) # build outputs __snake_case : str = [self.fpn_convs[i](laterals[i] ) for i in range(used_backbone_levels - 1 )] # append psp feature fpn_outs.append(laterals[-1] ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Tuple = nn.functional.interpolate( fpn_outs[i] , size=fpn_outs[0].shape[2:] , mode="bilinear" , align_corners=self.align_corners ) __snake_case : str = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Optional[Any] = self.fpn_bottleneck(lowerCamelCase ) __snake_case : Tuple = self.classifier(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Optional[int] , lowerCamelCase : int = 2 , lowerCamelCase : int = 3 , lowerCamelCase : Union[int, Tuple[int, int]] = 1 ) -> None: super().__init__() __snake_case : List[Any] = config __snake_case : List[str] = config.auxiliary_in_channels __snake_case : List[Any] = config.auxiliary_channels __snake_case : Tuple = config.auxiliary_num_convs __snake_case : int = config.auxiliary_concat_input __snake_case : Optional[int] = in_index __snake_case : Tuple = (kernel_size // 2) * dilation __snake_case : Optional[int] = [] convs.append( UperNetConvModule( self.in_channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) for i in range(self.num_convs - 1 ): convs.append( UperNetConvModule( self.channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) if self.num_convs == 0: __snake_case : Union[str, Any] = nn.Identity() else: __snake_case : Any = nn.Sequential(*lowerCamelCase ) if self.concat_input: __snake_case : int = UperNetConvModule( self.in_channels + self.channels , self.channels , kernel_size=lowerCamelCase , padding=kernel_size // 2 ) __snake_case : Optional[Any] = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) def __snake_case ( self : Dict ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : Tuple ) -> Optional[int]: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : Optional[int] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # just take the relevant feature maps __snake_case : List[str] = encoder_hidden_states[self.in_index] __snake_case : Optional[Any] = self.convs(lowerCamelCase ) if self.concat_input: __snake_case : Optional[Any] = self.conv_cat(torch.cat([hidden_states, output] , dim=1 ) ) __snake_case : Union[str, Any] = self.classifier(lowerCamelCase ) return output class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = UperNetConfig __UpperCAmelCase : int = "pixel_values" __UpperCAmelCase : str = True def __snake_case ( self : Union[str, Any] , lowerCamelCase : List[Any] ) -> List[Any]: if isinstance(lowerCamelCase , lowerCamelCase ): module.backbone.init_weights() module.decode_head.init_weights() module.auxiliary_head.init_weights() def __snake_case ( self : Optional[Any] ) -> List[str]: self.backbone.init_weights() self.decode_head.init_weights() self.auxiliary_head.init_weights() def __snake_case ( self : List[Any] , lowerCamelCase : Any , lowerCamelCase : Optional[Any]=False ) -> Dict: if isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : Union[str, Any] = value _snake_case : Dict = R"\n Parameters:\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n config ([`UperNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" _snake_case : Tuple = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using\n [`AutoImageProcessor`]. See [`SegformerImageProcessor.__call__`] for details.\n output_attentions (`bool`, *optional*):\n Whether or not to return the attentions tensors of all attention layers in case the backbone has them. See\n `attentions` under returned tensors for more detail.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers of the backbone. See `hidden_states` under\n returned tensors for more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n" @add_start_docstrings( "UperNet framework leveraging any vision backbone e.g. for ADE20k, CityScapes." , _lowerCAmelCase , ) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Dict , lowerCamelCase : int ) -> Optional[int]: super().__init__(lowerCamelCase ) __snake_case : Any = AutoBackbone.from_config(config.backbone_config ) # Semantic segmentation head(s) __snake_case : Union[str, Any] = UperNetHead(lowerCamelCase , in_channels=self.backbone.channels ) __snake_case : Any = UperNetFCNHead(lowerCamelCase ) if config.use_auxiliary_head else None # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UPERNET_INPUTS_DOCSTRING.format("batch_size, sequence_length" ) ) @replace_return_docstrings(output_type=lowerCamelCase , config_class=_CONFIG_FOR_DOC ) def __snake_case ( self : Union[str, Any] , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , ) -> Union[tuple, SemanticSegmenterOutput]: __snake_case : Any = return_dict if return_dict is not None else self.config.use_return_dict __snake_case : int = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __snake_case : str = output_attentions if output_attentions is not None else self.config.output_attentions __snake_case : Tuple = self.backbone.forward_with_filtered_kwargs( lowerCamelCase , output_hidden_states=lowerCamelCase , output_attentions=lowerCamelCase ) __snake_case : List[Any] = outputs.feature_maps __snake_case : List[Any] = self.decode_head(lowerCamelCase ) __snake_case : List[str] = nn.functional.interpolate(lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : Optional[int] = None if self.auxiliary_head is not None: __snake_case : Dict = self.auxiliary_head(lowerCamelCase ) __snake_case : Dict = nn.functional.interpolate( lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : int = None if labels is not None: if self.config.num_labels == 1: raise ValueError("The number of labels should be greater than one" ) else: # compute weighted loss __snake_case : Any = CrossEntropyLoss(ignore_index=self.config.loss_ignore_index ) __snake_case : Union[str, Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = main_loss + self.config.auxiliary_loss_weight * auxiliary_loss if not return_dict: if output_hidden_states: __snake_case : Any = (logits,) + outputs[1:] else: __snake_case : str = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return SemanticSegmenterOutput( loss=lowerCamelCase , logits=lowerCamelCase , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
81
0
"""simple docstring""" import copy from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging __SCREAMING_SNAKE_CASE =logging.get_logger(__name__) class UpperCamelCase ( _lowerCAmelCase ): lowercase = ["input_features"] def __init__( self ,__UpperCamelCase=80 ,__UpperCamelCase=1_6000 ,__UpperCamelCase=160 ,__UpperCamelCase=30 ,__UpperCamelCase=400 ,__UpperCamelCase=0.0 ,__UpperCamelCase=False ,**__UpperCamelCase ,) -> Dict: '''simple docstring''' super().__init__( feature_size=__UpperCamelCase ,sampling_rate=__UpperCamelCase ,padding_value=__UpperCamelCase ,return_attention_mask=__UpperCamelCase ,**__UpperCamelCase ,) lowercase_ : Union[str, Any] = n_fft lowercase_ : List[Any] = hop_length lowercase_ : Optional[Any] = chunk_length lowercase_ : List[str] = chunk_length * sampling_rate lowercase_ : str = self.n_samples // hop_length lowercase_ : Any = sampling_rate lowercase_ : Dict = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 ,num_mel_filters=__UpperCamelCase ,min_frequency=0.0 ,max_frequency=8000.0 ,sampling_rate=__UpperCamelCase ,norm='slaney' ,mel_scale='slaney' ,) def _UpperCAmelCase ( self ,__UpperCamelCase ) -> np.ndarray: '''simple docstring''' lowercase_ : Union[str, Any] = spectrogram( __UpperCamelCase ,window_function(self.n_fft ,'hann' ) ,frame_length=self.n_fft ,hop_length=self.hop_length ,power=2.0 ,mel_filters=self.mel_filters ,log_mel='log10' ,) lowercase_ : Dict = log_spec[:, :-1] lowercase_ : Any = np.maximum(__UpperCamelCase ,log_spec.max() - 8.0 ) lowercase_ : Union[str, Any] = (log_spec + 4.0) / 4.0 return log_spec @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def _UpperCAmelCase ( __UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase = 0.0 ) -> List[np.ndarray]: '''simple docstring''' if attention_mask is not None: lowercase_ : List[str] = np.array(__UpperCamelCase ,np.intaa ) lowercase_ : int = [] for vector, length in zip(__UpperCamelCase ,attention_mask.sum(-1 ) ): lowercase_ : int = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1e-7 ) if length < normed_slice.shape[0]: lowercase_ : int = padding_value normed_input_values.append(__UpperCamelCase ) else: lowercase_ : List[str] = [(x - x.mean()) / np.sqrt(x.var() + 1e-7 ) for x in input_values] return normed_input_values def __call__( self ,__UpperCamelCase ,__UpperCamelCase = True ,__UpperCamelCase = None ,__UpperCamelCase = None ,__UpperCamelCase = None ,__UpperCamelCase = "max_length" ,__UpperCamelCase = None ,__UpperCamelCase = None ,__UpperCamelCase = None ,**__UpperCamelCase ,) -> BatchFeature: '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a''' f''' sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input''' f''' was sampled with {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) lowercase_ : Any = isinstance(__UpperCamelCase ,np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowercase_ : Union[str, Any] = is_batched_numpy or ( isinstance(__UpperCamelCase ,(list, tuple) ) and (isinstance(raw_speech[0] ,(np.ndarray, tuple, list) )) ) if is_batched: lowercase_ : Any = [np.asarray([speech] ,dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(__UpperCamelCase ,np.ndarray ): lowercase_ : Tuple = np.asarray(__UpperCamelCase ,dtype=np.floataa ) elif isinstance(__UpperCamelCase ,np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowercase_ : Dict = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowercase_ : List[Any] = [np.asarray([raw_speech] ).T] lowercase_ : List[str] = BatchFeature({'input_features': raw_speech} ) # convert into correct format for padding lowercase_ : Union[str, Any] = self.pad( __UpperCamelCase ,padding=__UpperCamelCase ,max_length=max_length if max_length else self.n_samples ,truncation=__UpperCamelCase ,pad_to_multiple_of=__UpperCamelCase ,return_attention_mask=return_attention_mask or do_normalize ,) # zero-mean and unit-variance normalization if do_normalize: lowercase_ : List[str] = self.zero_mean_unit_var_norm( padded_inputs['input_features'] ,attention_mask=padded_inputs['attention_mask'] ,padding_value=self.padding_value ,) lowercase_ : Optional[int] = np.stack(padded_inputs['input_features'] ,axis=0 ) # make sure list is in array format lowercase_ : Optional[int] = padded_inputs.get('input_features' ).transpose(2 ,0 ,1 ) lowercase_ : Tuple = [self._np_extract_fbank_features(__UpperCamelCase ) for waveform in input_features[0]] if isinstance(input_features[0] ,__UpperCamelCase ): lowercase_ : Dict = [np.asarray(__UpperCamelCase ,dtype=np.floataa ) for feature in input_features] else: lowercase_ : Any = input_features if return_attention_mask: # rescale from sample (48000) to feature (3000) lowercase_ : Tuple = padded_inputs["attention_mask"][:, :: self.hop_length] if return_tensors is not None: lowercase_ : Dict = padded_inputs.convert_to_tensors(__UpperCamelCase ) return padded_inputs def _UpperCAmelCase ( self ) -> Dict[str, Any]: '''simple docstring''' lowercase_ : Tuple = copy.deepcopy(self.__dict__ ) lowercase_ : str = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] return output
425
from __future__ import annotations def lowerCAmelCase_ ( __lowerCamelCase ): if not nums: return 0 __snake_case : Any = nums[0] __snake_case : str = 0 for num in nums[1:]: __snake_case , __snake_case : List[str] = ( max_excluding + num, max(__lowerCamelCase , __lowerCamelCase ), ) return max(__lowerCamelCase , __lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
81
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) snake_case_ : Any = { "configuration_roformer": ["ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "RoFormerConfig", "RoFormerOnnxConfig"], "tokenization_roformer": ["RoFormerTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : Dict = ["RoFormerTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : List[str] = [ "ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "RoFormerForCausalLM", "RoFormerForMaskedLM", "RoFormerForMultipleChoice", "RoFormerForQuestionAnswering", "RoFormerForSequenceClassification", "RoFormerForTokenClassification", "RoFormerLayer", "RoFormerModel", "RoFormerPreTrainedModel", "load_tf_weights_in_roformer", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : Optional[Any] = [ "TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "TFRoFormerForCausalLM", "TFRoFormerForMaskedLM", "TFRoFormerForMultipleChoice", "TFRoFormerForQuestionAnswering", "TFRoFormerForSequenceClassification", "TFRoFormerForTokenClassification", "TFRoFormerLayer", "TFRoFormerModel", "TFRoFormerPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : Union[str, Any] = [ "FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "FlaxRoFormerForMaskedLM", "FlaxRoFormerForMultipleChoice", "FlaxRoFormerForQuestionAnswering", "FlaxRoFormerForSequenceClassification", "FlaxRoFormerForTokenClassification", "FlaxRoFormerModel", "FlaxRoFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_roformer import ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, RoFormerConfig, RoFormerOnnxConfig from .tokenization_roformer import RoFormerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roformer_fast import RoFormerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roformer import ( ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, RoFormerForCausalLM, RoFormerForMaskedLM, RoFormerForMultipleChoice, RoFormerForQuestionAnswering, RoFormerForSequenceClassification, RoFormerForTokenClassification, RoFormerLayer, RoFormerModel, RoFormerPreTrainedModel, load_tf_weights_in_roformer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roformer import ( TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForMultipleChoice, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerLayer, TFRoFormerModel, TFRoFormerPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roformer import ( FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxRoFormerForMaskedLM, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerModel, FlaxRoFormerPreTrainedModel, ) else: import sys snake_case_ : List[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
212
from __future__ import annotations from typing import Any def lowerCAmelCase_ ( __lowerCamelCase ): create_state_space_tree(__lowerCamelCase , [] , 0 ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): if index == len(__lowerCamelCase ): print(__lowerCamelCase ) return create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.pop() if __name__ == "__main__": _snake_case : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["A", "B", "C"]) generate_all_subsequences(seq)
81
0
import os from collections.abc import Iterator def __snake_case ( _lowerCAmelCase : List[str] = "." ) -> int: for dir_path, dir_names, filenames in os.walk(__lowerCamelCase ): A_ : Tuple = [d for d in dir_names if d != "scripts" and d[0] not in "._"] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(__lowerCamelCase )[1] in (".py", ".ipynb"): yield os.path.join(__lowerCamelCase , __lowerCamelCase ).lstrip("./" ) def __snake_case ( _lowerCAmelCase : List[str] ) -> Union[str, Any]: return f"{i * ' '}*" if i else "\n##" def __snake_case ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] ) -> List[str]: A_ : List[str] = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(__lowerCamelCase ) or old_parts[i] != new_part) and new_part: print(f"{md_prefix(__lowerCamelCase )} {new_part.replace('_' , ' ' ).title()}" ) return new_path def __snake_case ( _lowerCAmelCase : List[str] = "." ) -> Optional[int]: A_ : Any = "" for filepath in sorted(good_file_paths(__lowerCamelCase ) ): A_ : Any = os.path.split(__lowerCamelCase ) if filepath != old_path: A_ : str = print_path(__lowerCamelCase , __lowerCamelCase ) A_ : Union[str, Any] = (filepath.count(os.sep ) + 1) if filepath else 0 A_ : List[Any] = f"{filepath}/{filename}".replace(" " , "%20" ) A_ : str = os.path.splitext(filename.replace("_" , " " ).title() )[0] print(f"{md_prefix(__lowerCamelCase )} [{filename}]({url})" ) if __name__ == "__main__": print_directory_md('''.''')
454
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case , __snake_case : List[Any] = image.size __snake_case , __snake_case : Tuple = (x - x % 3_2 for x in (w, h)) # resize to integer multiple of 32 __snake_case : str = image.resize((w, h) , resample=PIL_INTERPOLATION["lanczos"] ) __snake_case : int = np.array(__lowerCamelCase ).astype(np.floataa ) / 2_5_5.0 __snake_case : Union[str, Any] = image[None].transpose(0 , 3 , 1 , 2 ) __snake_case : Union[str, Any] = torch.from_numpy(__lowerCamelCase ) return 2.0 * image - 1.0 class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : VQModel , lowerCamelCase : UNetaDModel , lowerCamelCase : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ) -> Union[str, Any]: super().__init__() self.register_modules(vqvae=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase ) @torch.no_grad() def __call__( self : List[str] , lowerCamelCase : Union[torch.Tensor, PIL.Image.Image] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : Optional[int] = 100 , lowerCamelCase : Optional[float] = 0.0 , lowerCamelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , ) -> Union[Tuple, ImagePipelineOutput]: if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : Any = 1 elif isinstance(lowerCamelCase , torch.Tensor ): __snake_case : Any = image.shape[0] else: raise ValueError(F'`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(lowerCamelCase )}' ) if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : List[Any] = preprocess(lowerCamelCase ) __snake_case , __snake_case : int = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image __snake_case : str = (batch_size, self.unet.config.in_channels // 2, height, width) __snake_case : str = next(self.unet.parameters() ).dtype __snake_case : Tuple = randn_tensor(lowerCamelCase , generator=lowerCamelCase , device=self.device , dtype=lowerCamelCase ) __snake_case : List[Any] = image.to(device=self.device , dtype=lowerCamelCase ) # set timesteps and move to the correct device self.scheduler.set_timesteps(lowerCamelCase , device=self.device ) __snake_case : str = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler __snake_case : Dict = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] __snake_case : Union[str, Any] = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __snake_case : int = {} if accepts_eta: __snake_case : List[str] = eta for t in self.progress_bar(lowerCamelCase ): # concat latents and low resolution image in the channel dimension. __snake_case : Union[str, Any] = torch.cat([latents, image] , dim=1 ) __snake_case : Optional[Any] = self.scheduler.scale_model_input(lowerCamelCase , lowerCamelCase ) # predict the noise residual __snake_case : int = self.unet(lowerCamelCase , lowerCamelCase ).sample # compute the previous noisy sample x_t -> x_t-1 __snake_case : Union[str, Any] = self.scheduler.step(lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase ).prev_sample # decode the image latents with the VQVAE __snake_case : List[Any] = self.vqvae.decode(lowerCamelCase ).sample __snake_case : Dict = torch.clamp(lowerCamelCase , -1.0 , 1.0 ) __snake_case : Any = image / 2 + 0.5 __snake_case : Tuple = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __snake_case : Tuple = self.numpy_to_pil(lowerCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCamelCase )
81
0
"""simple docstring""" import doctest from collections import deque import numpy as np class _UpperCamelCase : def __init__(self ): """simple docstring""" A__ = [2, 1, 2, -1] A__ = [1, 2, 3, 4] def A (self ): """simple docstring""" A__ = len(self.first_signal ) A__ = len(self.second_signal ) A__ = max(lowerCamelCase__ , lowerCamelCase__ ) # create a zero matrix of max_length x max_length A__ = [[0] * max_length for i in range(lowerCamelCase__ )] # fills the smaller signal with zeros to make both signals of same length if length_first_signal < length_second_signal: self.first_signal += [0] * (max_length - length_first_signal) elif length_first_signal > length_second_signal: self.second_signal += [0] * (max_length - length_second_signal) for i in range(lowerCamelCase__ ): A__ = deque(self.second_signal ) rotated_signal.rotate(lowerCamelCase__ ) for j, item in enumerate(lowerCamelCase__ ): matrix[i][j] += item # multiply the matrix with the first signal A__ = np.matmul(np.transpose(lowerCamelCase__ ) , np.transpose(self.first_signal ) ) # rounding-off to two decimal places return [round(lowerCamelCase__ , 2 ) for i in final_signal] if __name__ == "__main__": doctest.testmod()
574
import gc import unittest import torch from parameterized import parameterized from diffusers import AutoencoderKL from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = AutoencoderKL __UpperCAmelCase : Optional[Any] = "sample" __UpperCAmelCase : Optional[int] = 1e-2 @property def __snake_case ( self : Dict ) -> Optional[Any]: __snake_case : Optional[Any] = 4 __snake_case : Tuple = 3 __snake_case : List[str] = (32, 32) __snake_case : str = floats_tensor((batch_size, num_channels) + sizes ).to(lowerCamelCase ) return {"sample": image} @property def __snake_case ( self : Union[str, Any] ) -> Tuple: return (3, 32, 32) @property def __snake_case ( self : int ) -> int: return (3, 32, 32) def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Optional[Any] = { "block_out_channels": [32, 64], "in_channels": 3, "out_channels": 3, "down_block_types": ["DownEncoderBlock2D", "DownEncoderBlock2D"], "up_block_types": ["UpDecoderBlock2D", "UpDecoderBlock2D"], "latent_channels": 4, } __snake_case : Any = self.dummy_input return init_dict, inputs_dict def __snake_case ( self : str ) -> Dict: pass def __snake_case ( self : Tuple ) -> List[str]: pass @unittest.skipIf(torch_device == "mps" , "Gradient checkpointing skipped on MPS" ) def __snake_case ( self : Any ) -> Optional[Any]: # enable deterministic behavior for gradient checkpointing __snake_case , __snake_case : int = self.prepare_init_args_and_inputs_for_common() __snake_case : str = self.model_class(**lowerCamelCase ) model.to(lowerCamelCase ) assert not model.is_gradient_checkpointing and model.training __snake_case : str = model(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model.zero_grad() __snake_case : Any = torch.randn_like(lowerCamelCase ) __snake_case : str = (out - labels).mean() loss.backward() # re-instantiate the model now enabling gradient checkpointing __snake_case : Optional[int] = self.model_class(**lowerCamelCase ) # clone model model_a.load_state_dict(model.state_dict() ) model_a.to(lowerCamelCase ) model_a.enable_gradient_checkpointing() assert model_a.is_gradient_checkpointing and model_a.training __snake_case : int = model_a(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model_a.zero_grad() __snake_case : Union[str, Any] = (out_a - labels).mean() loss_a.backward() # compare the output and parameters gradients self.assertTrue((loss - loss_a).abs() < 1E-5 ) __snake_case : Optional[int] = dict(model.named_parameters() ) __snake_case : List[Any] = dict(model_a.named_parameters() ) for name, param in named_params.items(): self.assertTrue(torch_all_close(param.grad.data , named_params_a[name].grad.data , atol=5E-5 ) ) def __snake_case ( self : List[Any] ) -> Optional[int]: __snake_case , __snake_case : Optional[Any] = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" , output_loading_info=lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(lowerCamelCase ) __snake_case : Optional[Any] = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def __snake_case ( self : Optional[Any] ) -> Union[str, Any]: __snake_case : Tuple = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" ) __snake_case : Dict = model.to(lowerCamelCase ) model.eval() if torch_device == "mps": __snake_case : int = torch.manual_seed(0 ) else: __snake_case : str = torch.Generator(device=lowerCamelCase ).manual_seed(0 ) __snake_case : List[str] = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) __snake_case : Union[str, Any] = image.to(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , sample_posterior=lowerCamelCase , generator=lowerCamelCase ).sample __snake_case : List[Any] = output[0, -1, -3:, -3:].flatten().cpu() # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. if torch_device == "mps": __snake_case : Union[str, Any] = torch.tensor( [ -4.0078E-01, -3.8323E-04, -1.2681E-01, -1.1462E-01, 2.0095E-01, 1.0893E-01, -8.8247E-02, -3.0361E-01, -9.8644E-03, ] ) elif torch_device == "cpu": __snake_case : Tuple = torch.tensor( [-0.13_52, 0.08_78, 0.04_19, -0.08_18, -0.10_69, 0.06_88, -0.14_58, -0.44_46, -0.00_26] ) else: __snake_case : List[str] = torch.tensor( [-0.24_21, 0.46_42, 0.25_07, -0.04_38, 0.06_82, 0.31_60, -0.20_18, -0.07_27, 0.24_85] ) self.assertTrue(torch_all_close(lowerCamelCase , lowerCamelCase , rtol=1E-2 ) ) @slow class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : int , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] ) -> List[str]: return F'gaussian_noise_s={seed}_shape={"_".join([str(lowerCamelCase ) for s in shape] )}.npy' def __snake_case ( self : List[Any] ) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __snake_case ( self : Tuple , lowerCamelCase : List[Any]=0 , lowerCamelCase : Tuple=(4, 3, 512, 512) , lowerCamelCase : Optional[int]=False ) -> str: __snake_case : List[Any] = torch.floataa if fpaa else torch.floataa __snake_case : Tuple = torch.from_numpy(load_hf_numpy(self.get_file_format(lowerCamelCase , lowerCamelCase ) ) ).to(lowerCamelCase ).to(lowerCamelCase ) return image def __snake_case ( self : Optional[Any] , lowerCamelCase : int="CompVis/stable-diffusion-v1-4" , lowerCamelCase : int=False ) -> int: __snake_case : str = "fp16" if fpaa else None __snake_case : int = torch.floataa if fpaa else torch.floataa __snake_case : int = AutoencoderKL.from_pretrained( lowerCamelCase , subfolder="vae" , torch_dtype=lowerCamelCase , revision=lowerCamelCase , ) model.to(lowerCamelCase ).eval() return model def __snake_case ( self : str , lowerCamelCase : int=0 ) -> Optional[Any]: if torch_device == "mps": return torch.manual_seed(lowerCamelCase ) return torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) @parameterized.expand( [ # fmt: off [33, [-0.16_03, 0.98_78, -0.04_95, -0.07_90, -0.27_09, 0.83_75, -0.20_60, -0.08_24], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_76, 0.11_68, 0.13_32, -0.48_40, -0.25_08, -0.07_91, -0.04_93, -0.40_89], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[str] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] ) -> List[Any]: __snake_case : Optional[Any] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) __snake_case : Tuple = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : List[Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : int = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [33, [-0.05_13, 0.02_89, 1.37_99, 0.21_66, -0.25_73, -0.08_71, 0.51_03, -0.09_99]], [47, [-0.41_28, -0.13_20, -0.37_04, 0.19_65, -0.41_16, -0.23_32, -0.33_40, 0.22_47]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : Any , lowerCamelCase : List[str] , lowerCamelCase : List[str] ) -> Tuple: __snake_case : Any = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Any = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.16_09, 0.98_66, -0.04_87, -0.07_77, -0.27_16, 0.83_68, -0.20_55, -0.08_14], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_77, 0.11_47, 0.13_33, -0.48_41, -0.25_06, -0.08_05, -0.04_91, -0.40_85], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Any , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) with torch.no_grad(): __snake_case : int = model(lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Union[str, Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : List[str] = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [13, [-0.20_51, -0.18_03, -0.23_11, -0.21_14, -0.32_92, -0.35_74, -0.29_53, -0.33_23]], [37, [-0.26_32, -0.26_25, -0.21_99, -0.27_41, -0.45_39, -0.49_90, -0.37_20, -0.49_25]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : List[str] , lowerCamelCase : Tuple , lowerCamelCase : Any ) -> Optional[Any]: __snake_case : List[str] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : str = sample[-1, -2:, :2, -2:].flatten().cpu() __snake_case : Optional[int] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-3 ) @parameterized.expand( [ # fmt: off [27, [-0.03_69, 0.02_07, -0.07_76, -0.06_82, -0.17_47, -0.19_30, -0.14_65, -0.20_39]], [16, [-0.16_28, -0.21_34, -0.27_47, -0.26_42, -0.37_74, -0.44_04, -0.36_87, -0.42_77]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : str , lowerCamelCase : Optional[int] , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[str] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : Union[str, Any] = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Optional[Any] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=5E-3 ) @parameterized.expand([(13,), (16,), (27,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> Tuple: __snake_case : Dict = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : Any = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Any = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-1 ) @parameterized.expand([(13,), (16,), (37,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : List[Any] , lowerCamelCase : Any ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : Union[str, Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : List[Any] = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Dict = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.30_01, 0.09_18, -2.69_84, -3.97_20, -3.20_99, -5.03_53, 1.73_38, -0.20_65, 3.42_67]], [47, [-1.50_30, -4.38_71, -6.03_55, -9.11_57, -1.66_61, -2.78_53, 2.16_07, -5.08_23, 2.56_33]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Dict ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : int = self.get_sd_image(lowerCamelCase ) __snake_case : int = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model.encode(lowerCamelCase ).latent_dist __snake_case : Dict = dist.sample(generator=lowerCamelCase ) assert list(sample.shape ) == [image.shape[0], 4] + [i // 8 for i in image.shape[2:]] __snake_case : List[str] = sample[0, -1, -3:, -3:].flatten().cpu() __snake_case : Dict = torch.tensor(lowerCamelCase ) __snake_case : Dict = 3E-3 if torch_device != "mps" else 1E-2 assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=lowerCamelCase )
81
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __a: Optional[Any] = logging.get_logger(__name__) __a: Optional[int] = { "xlm-mlm-en-2048": "https://huggingface.co/xlm-mlm-en-2048/resolve/main/config.json", "xlm-mlm-ende-1024": "https://huggingface.co/xlm-mlm-ende-1024/resolve/main/config.json", "xlm-mlm-enfr-1024": "https://huggingface.co/xlm-mlm-enfr-1024/resolve/main/config.json", "xlm-mlm-enro-1024": "https://huggingface.co/xlm-mlm-enro-1024/resolve/main/config.json", "xlm-mlm-tlm-xnli15-1024": "https://huggingface.co/xlm-mlm-tlm-xnli15-1024/resolve/main/config.json", "xlm-mlm-xnli15-1024": "https://huggingface.co/xlm-mlm-xnli15-1024/resolve/main/config.json", "xlm-clm-enfr-1024": "https://huggingface.co/xlm-clm-enfr-1024/resolve/main/config.json", "xlm-clm-ende-1024": "https://huggingface.co/xlm-clm-ende-1024/resolve/main/config.json", "xlm-mlm-17-1280": "https://huggingface.co/xlm-mlm-17-1280/resolve/main/config.json", "xlm-mlm-100-1280": "https://huggingface.co/xlm-mlm-100-1280/resolve/main/config.json", } class UpperCAmelCase ( _lowerCAmelCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = "xlm" SCREAMING_SNAKE_CASE = { "hidden_size": "emb_dim", "num_attention_heads": "n_heads", "num_hidden_layers": "n_layers", "n_words": "vocab_size", # For backward compatibility } def __init__( self , __lowerCAmelCase=30145 , __lowerCAmelCase=2048 , __lowerCAmelCase=12 , __lowerCAmelCase=16 , __lowerCAmelCase=0.1 , __lowerCAmelCase=0.1 , __lowerCAmelCase=True , __lowerCAmelCase=False , __lowerCAmelCase=False , __lowerCAmelCase=False , __lowerCAmelCase=1 , __lowerCAmelCase=True , __lowerCAmelCase=512 , __lowerCAmelCase=2048**-0.5 , __lowerCAmelCase=1E-12 , __lowerCAmelCase=0.0_2 , __lowerCAmelCase=0 , __lowerCAmelCase=1 , __lowerCAmelCase=2 , __lowerCAmelCase=3 , __lowerCAmelCase=5 , __lowerCAmelCase=True , __lowerCAmelCase="first" , __lowerCAmelCase=True , __lowerCAmelCase=None , __lowerCAmelCase=True , __lowerCAmelCase=0.1 , __lowerCAmelCase=5 , __lowerCAmelCase=5 , __lowerCAmelCase=0 , __lowerCAmelCase=0 , __lowerCAmelCase=2 , __lowerCAmelCase=0 , **__lowerCAmelCase , ) -> List[Any]: lowercase__ : Any = vocab_size lowercase__ : List[Any] = emb_dim lowercase__ : List[Any] = n_layers lowercase__ : Dict = n_heads lowercase__ : Any = dropout lowercase__ : List[str] = attention_dropout lowercase__ : Any = gelu_activation lowercase__ : Dict = sinusoidal_embeddings lowercase__ : Optional[int] = causal lowercase__ : Dict = asm lowercase__ : Optional[int] = n_langs lowercase__ : Dict = use_lang_emb lowercase__ : str = layer_norm_eps lowercase__ : Optional[Any] = bos_index lowercase__ : Optional[Any] = eos_index lowercase__ : List[Any] = pad_index lowercase__ : str = unk_index lowercase__ : int = mask_index lowercase__ : Tuple = is_encoder lowercase__ : Optional[int] = max_position_embeddings lowercase__ : Any = embed_init_std lowercase__ : Optional[int] = init_std lowercase__ : Optional[int] = summary_type lowercase__ : Optional[int] = summary_use_proj lowercase__ : Tuple = summary_activation lowercase__ : str = summary_proj_to_labels lowercase__ : Union[str, Any] = summary_first_dropout lowercase__ : Union[str, Any] = start_n_top lowercase__ : List[Any] = end_n_top lowercase__ : Dict = mask_token_id lowercase__ : Tuple = lang_id if "n_words" in kwargs: lowercase__ : str = kwargs["n_words"] super().__init__(pad_token_id=__lowerCAmelCase , bos_token_id=__lowerCAmelCase , **__lowerCAmelCase ) class UpperCAmelCase ( _lowerCAmelCase ): '''simple docstring''' @property def _lowerCAmelCase( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": lowercase__ : Optional[int] = {0: "batch", 1: "choice", 2: "sequence"} else: lowercase__ : List[str] = {0: "batch", 1: "sequence"} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ('''token_type_ids''', dynamic_axis), ] )
152
import warnings from ...utils import logging from .image_processing_flava import FlavaImageProcessor _snake_case : Optional[int] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[Any] , *lowerCamelCase : Any , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class FlavaFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use FlavaImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
class A : def __init__( self , lowercase_ ) -> None: '''simple docstring''' _snake_case : str = set_counts _snake_case : Union[str, Any] = max(lowercase_ ) _snake_case : List[Any] = len(lowercase_ ) _snake_case : Tuple = [1] * num_sets _snake_case : Dict = list(range(lowercase_ ) ) def __a ( self , lowercase_ , lowercase_ ) -> bool: '''simple docstring''' _snake_case : List[Any] = self.get_parent(lowercase_ ) _snake_case : Tuple = self.get_parent(lowercase_ ) if src_parent == dst_parent: return False if self.ranks[dst_parent] >= self.ranks[src_parent]: self.set_counts[dst_parent] += self.set_counts[src_parent] _snake_case : List[str] = 0 _snake_case : List[Any] = dst_parent if self.ranks[dst_parent] == self.ranks[src_parent]: self.ranks[dst_parent] += 1 _snake_case : Dict = self.set_counts[dst_parent] else: self.set_counts[src_parent] += self.set_counts[dst_parent] _snake_case : Union[str, Any] = 0 _snake_case : Optional[int] = src_parent _snake_case : Tuple = self.set_counts[src_parent] _snake_case : str = max(self.max_set , lowercase_ ) return True def __a ( self , lowercase_ ) -> int: '''simple docstring''' if self.parents[disj_set] == disj_set: return disj_set _snake_case : Optional[int] = self.get_parent(self.parents[disj_set] ) return self.parents[disj_set]
326
import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : Tuple ) -> Optional[Any]: __snake_case : Dict = tempfile.mkdtemp() __snake_case : Any = SamImageProcessor() __snake_case : Optional[int] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : Optional[Any] , **lowerCamelCase : Optional[int] ) -> Optional[Any]: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[Any] ) -> Dict: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : int ) -> List[Any]: __snake_case : List[str] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : int = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : List[Any] ) -> Dict: __snake_case : int = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Dict = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Optional[Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : List[str] ) -> Tuple: __snake_case : int = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Optional[int] = self.prepare_image_inputs() __snake_case : List[str] = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : Dict = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_torch def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Tuple = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[str] = [torch.ones((1, 3, 5, 5) )] __snake_case : Tuple = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : int = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , torch.tensor(lowerCamelCase ) , torch.tensor(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : List[str] = [np.ones((1, 3, 5, 5) )] __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : str = [[1, 0], [0, 1]] with self.assertRaises(lowerCamelCase ): __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) @require_vision @require_tf class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[Any] ) -> Union[str, Any]: __snake_case : int = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : str , **lowerCamelCase : Any ) -> Tuple: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> Any: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : str ) -> List[Any]: __snake_case : int = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : Dict = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : int ) -> List[str]: __snake_case : List[Any] = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Optional[Any] = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Tuple = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : Union[str, Any] ) -> List[Any]: __snake_case : str = self.get_image_processor() __snake_case : Union[str, Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : int = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : List[str] = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_tf def __snake_case ( self : Any ) -> Optional[int]: __snake_case : List[str] = self.get_image_processor() __snake_case : Dict = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Union[str, Any] = [tf.ones((1, 3, 5, 5) )] __snake_case : List[Any] = [[1764, 2646]] __snake_case : Dict = [[683, 1024]] __snake_case : List[str] = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , tf.convert_to_tensor(lowerCamelCase ) , tf.convert_to_tensor(lowerCamelCase ) , return_tensors="tf" , ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : Union[str, Any] = [np.ones((1, 3, 5, 5) )] __snake_case : List[str] = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Tuple = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): __snake_case : Dict = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[str] ) -> str: __snake_case : Optional[int] = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : List[str] , **lowerCamelCase : Any ) -> Dict: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> List[Any]: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : str = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : List[Any] = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : str = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[Any] = np.random.randint(0 , 2 , size=(1, 3, 5, 5) ).astype(np.floataa ) __snake_case : Dict = [tf.convert_to_tensor(lowerCamelCase )] __snake_case : List[Any] = [torch.tensor(lowerCamelCase )] __snake_case : Optional[Any] = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : Union[str, Any] = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) __snake_case : Dict = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="pt" ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def __snake_case ( self : List[Any] ) -> List[str]: __snake_case : Any = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : Any = image_processor(lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Optional[Any] = processor(images=lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Tuple = image_processor(lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() __snake_case : List[Any] = processor(images=lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) )
81
0
"""simple docstring""" from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging __magic_name__ = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE_ ( _lowerCAmelCase ): """simple docstring""" __lowercase : List[str] = ["pixel_values"] def __init__( self , lowerCAmelCase__ = True , lowerCAmelCase__ = 1 / 2_5_5 , lowerCAmelCase__ = True , lowerCAmelCase__ = 8 , **lowerCAmelCase__ , ): super().__init__(**lowerCAmelCase__) __SCREAMING_SNAKE_CASE = do_rescale __SCREAMING_SNAKE_CASE = rescale_factor __SCREAMING_SNAKE_CASE = do_pad __SCREAMING_SNAKE_CASE = pad_size def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None , **lowerCAmelCase__): return rescale(lowerCAmelCase__ , scale=lowerCAmelCase__ , data_format=lowerCAmelCase__ , **lowerCAmelCase__) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None): __SCREAMING_SNAKE_CASE = get_image_size(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = (old_height // size + 1) * size - old_height __SCREAMING_SNAKE_CASE = (old_width // size + 1) * size - old_width return pad(lowerCAmelCase__ , ((0, pad_height), (0, pad_width)) , mode="""symmetric""" , data_format=lowerCAmelCase__) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = ChannelDimension.FIRST , **lowerCAmelCase__ , ): __SCREAMING_SNAKE_CASE = do_rescale if do_rescale is not None else self.do_rescale __SCREAMING_SNAKE_CASE = rescale_factor if rescale_factor is not None else self.rescale_factor __SCREAMING_SNAKE_CASE = do_pad if do_pad is not None else self.do_pad __SCREAMING_SNAKE_CASE = pad_size if pad_size is not None else self.pad_size __SCREAMING_SNAKE_CASE = make_list_of_images(lowerCAmelCase__) if not valid_images(lowerCAmelCase__): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""") if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""") # All transformations expect numpy arrays. __SCREAMING_SNAKE_CASE = [to_numpy_array(lowerCAmelCase__) for image in images] if do_rescale: __SCREAMING_SNAKE_CASE = [self.rescale(image=lowerCAmelCase__ , scale=lowerCAmelCase__) for image in images] if do_pad: __SCREAMING_SNAKE_CASE = [self.pad(lowerCAmelCase__ , size=lowerCAmelCase__) for image in images] __SCREAMING_SNAKE_CASE = [to_channel_dimension_format(lowerCAmelCase__ , lowerCAmelCase__) for image in images] __SCREAMING_SNAKE_CASE = {"pixel_values": images} return BatchFeature(data=lowerCAmelCase__ , tensor_type=lowerCAmelCase__)
155
import warnings from ...utils import logging from .image_processing_deit import DeiTImageProcessor _snake_case : List[str] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[str] , *lowerCamelCase : str , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class DeiTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use DeiTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
from importlib import import_module from .logging import get_logger _UpperCAmelCase : Tuple = get_logger(__name__) class lowerCAmelCase_ : def __init__( self : str , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : Dict=None ): lowerCAmelCase__ = attrs or [] if module is not None: for key in module.__dict__: if key in attrs or not key.startswith('''__''' ): setattr(self , SCREAMING_SNAKE_CASE_ , getattr(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) lowerCAmelCase__ = module._original_module if isinstance(SCREAMING_SNAKE_CASE_ , _PatchedModuleObj ) else module class lowerCAmelCase_ : UpperCamelCase_ :str = [] def __init__( self : Dict , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : List[str] , SCREAMING_SNAKE_CASE_ : Any=None ): lowerCAmelCase__ = obj lowerCAmelCase__ = target lowerCAmelCase__ = new lowerCAmelCase__ = target.split('''.''' )[0] lowerCAmelCase__ = {} lowerCAmelCase__ = attrs or [] def __enter__( self : Dict ): lowerCAmelCase__ = self.target.split('''.''' ) # Patch modules: # it's used to patch attributes of submodules like "os.path.join"; # in this case we need to patch "os" and "os.path" for i in range(len(SCREAMING_SNAKE_CASE_ ) ): try: lowerCAmelCase__ = import_module('''.'''.join(submodules[: i + 1] ) ) except ModuleNotFoundError: continue # We iterate over all the globals in self.obj in case we find "os" or "os.path" for attr in self.obj.__dir__(): lowerCAmelCase__ = getattr(self.obj , SCREAMING_SNAKE_CASE_ ) # We don't check for the name of the global, but rather if its value *is* "os" or "os.path". # This allows to patch renamed modules like "from os import path as ospath". if obj_attr is submodule or ( (isinstance(SCREAMING_SNAKE_CASE_ , _PatchedModuleObj ) and obj_attr._original_module is submodule) ): lowerCAmelCase__ = obj_attr # patch at top level setattr(self.obj , SCREAMING_SNAKE_CASE_ , _PatchedModuleObj(SCREAMING_SNAKE_CASE_ , attrs=self.attrs ) ) lowerCAmelCase__ = getattr(self.obj , SCREAMING_SNAKE_CASE_ ) # construct lower levels patches for key in submodules[i + 1 :]: setattr(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , _PatchedModuleObj(getattr(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) , attrs=self.attrs ) ) lowerCAmelCase__ = getattr(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # finally set the target attribute setattr(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , self.new ) # Patch attribute itself: # it's used for builtins like "open", # and also to patch "os.path.join" we may also need to patch "join" # itself if it was imported as "from os.path import join". if submodules: # if it's an attribute of a submodule like "os.path.join" try: lowerCAmelCase__ = getattr(import_module('''.'''.join(SCREAMING_SNAKE_CASE_ ) ) , SCREAMING_SNAKE_CASE_ ) except (AttributeError, ModuleNotFoundError): return # We iterate over all the globals in self.obj in case we find "os.path.join" for attr in self.obj.__dir__(): # We don't check for the name of the global, but rather if its value *is* "os.path.join". # This allows to patch renamed attributes like "from os.path import join as pjoin". if getattr(self.obj , SCREAMING_SNAKE_CASE_ ) is attr_value: lowerCAmelCase__ = getattr(self.obj , SCREAMING_SNAKE_CASE_ ) setattr(self.obj , SCREAMING_SNAKE_CASE_ , self.new ) elif target_attr in globals()["__builtins__"]: # if it'a s builtin like "open" lowerCAmelCase__ = globals()["__builtins__"][target_attr] setattr(self.obj , SCREAMING_SNAKE_CASE_ , self.new ) else: raise RuntimeError(f'Tried to patch attribute {target_attr} instead of a submodule.' ) def __exit__( self : Optional[int] , *SCREAMING_SNAKE_CASE_ : Optional[Any] ): for attr in list(self.original ): setattr(self.obj , SCREAMING_SNAKE_CASE_ , self.original.pop(SCREAMING_SNAKE_CASE_ ) ) def __snake_case ( self : Any ): self.__enter__() self._active_patches.append(self ) def __snake_case ( self : List[str] ): try: self._active_patches.remove(self ) except ValueError: # If the patch hasn't been started this will fail return None return self.__exit__()
668
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case : Union[str, Any] = { "configuration_owlvit": [ "OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "OwlViTConfig", "OwlViTOnnxConfig", "OwlViTTextConfig", "OwlViTVisionConfig", ], "processing_owlvit": ["OwlViTProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[Any] = ["OwlViTFeatureExtractor"] _snake_case : Optional[int] = ["OwlViTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "OwlViTModel", "OwlViTPreTrainedModel", "OwlViTTextModel", "OwlViTVisionModel", "OwlViTForObjectDetection", ] if TYPE_CHECKING: from .configuration_owlvit import ( OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, OwlViTConfig, OwlViTOnnxConfig, OwlViTTextConfig, OwlViTVisionConfig, ) from .processing_owlvit import OwlViTProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_owlvit import OwlViTFeatureExtractor from .image_processing_owlvit import OwlViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_owlvit import ( OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST, OwlViTForObjectDetection, OwlViTModel, OwlViTPreTrainedModel, OwlViTTextModel, OwlViTVisionModel, ) else: import sys _snake_case : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
0
'''simple docstring''' import inspect from typing import Optional, Union import numpy as np import PIL import torch from torch.nn import functional as F from torchvision import transforms from transformers import CLIPFeatureExtractor, CLIPModel, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, DPMSolverMultistepScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion import StableDiffusionPipelineOutput from diffusers.utils import ( PIL_INTERPOLATION, randn_tensor, ) def _a ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Dict: """simple docstring""" if isinstance(__lowerCamelCase , torch.Tensor ): return image elif isinstance(__lowerCamelCase , PIL.Image.Image ): __snake_case : List[Any] = [image] if isinstance(image[0] , PIL.Image.Image ): __snake_case : Any = [np.array(i.resize((w, h) , resample=PIL_INTERPOLATION["""lanczos"""] ) )[None, :] for i in image] __snake_case : Tuple = np.concatenate(__lowerCamelCase , axis=0 ) __snake_case : Tuple = np.array(__lowerCamelCase ).astype(np.floataa ) / 255.0 __snake_case : List[str] = image.transpose(0 , 3 , 1 , 2 ) __snake_case : List[str] = 2.0 * image - 1.0 __snake_case : Union[str, Any] = torch.from_numpy(__lowerCamelCase ) elif isinstance(image[0] , torch.Tensor ): __snake_case : int = torch.cat(__lowerCamelCase , dim=0 ) return image def _a ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase=0.99_95 ) -> Optional[Any]: """simple docstring""" if not isinstance(__lowerCamelCase , np.ndarray ): __snake_case : Any = True __snake_case : Any = va.device __snake_case : str = va.cpu().numpy() __snake_case : Optional[int] = va.cpu().numpy() __snake_case : List[Any] = np.sum(va * va / (np.linalg.norm(__lowerCamelCase ) * np.linalg.norm(__lowerCamelCase )) ) if np.abs(__lowerCamelCase ) > DOT_THRESHOLD: __snake_case : List[Any] = (1 - t) * va + t * va else: __snake_case : Any = np.arccos(__lowerCamelCase ) __snake_case : Optional[Any] = np.sin(__lowerCamelCase ) __snake_case : int = theta_a * t __snake_case : List[Any] = np.sin(__lowerCamelCase ) __snake_case : str = np.sin(theta_a - theta_t ) / sin_theta_a __snake_case : Optional[Any] = sin_theta_t / sin_theta_a __snake_case : Dict = sa * va + sa * va if inputs_are_torch: __snake_case : List[Any] = torch.from_numpy(__lowerCamelCase ).to(__lowerCamelCase ) return va def _a ( _lowerCamelCase , _lowerCamelCase ) -> Tuple: """simple docstring""" __snake_case : Any = F.normalize(__lowerCamelCase , dim=-1 ) __snake_case : Optional[Any] = F.normalize(__lowerCamelCase , dim=-1 ) return (x - y).norm(dim=-1 ).div(2 ).arcsin().pow(2 ).mul(2 ) def _a ( _lowerCamelCase , _lowerCamelCase ) -> Dict: """simple docstring""" for param in model.parameters(): __snake_case : List[str] = value class _A ( _lowerCAmelCase ): def __init__( self : Optional[Any] , __magic_name__ : AutoencoderKL , __magic_name__ : CLIPTextModel , __magic_name__ : CLIPModel , __magic_name__ : CLIPTokenizer , __magic_name__ : UNetaDConditionModel , __magic_name__ : Union[PNDMScheduler, LMSDiscreteScheduler, DDIMScheduler, DPMSolverMultistepScheduler] , __magic_name__ : CLIPFeatureExtractor , __magic_name__ : Any=None , __magic_name__ : Union[str, Any]=None , __magic_name__ : Tuple=None , ) -> Tuple: """simple docstring""" super().__init__() self.register_modules( vae=__magic_name__ , text_encoder=__magic_name__ , clip_model=__magic_name__ , tokenizer=__magic_name__ , unet=__magic_name__ , scheduler=__magic_name__ , feature_extractor=__magic_name__ , coca_model=__magic_name__ , coca_tokenizer=__magic_name__ , coca_transform=__magic_name__ , ) __snake_case : str = ( feature_extractor.size if isinstance(feature_extractor.size , __magic_name__ ) else feature_extractor.size["shortest_edge"] ) __snake_case : Optional[Any] = transforms.Normalize(mean=feature_extractor.image_mean , std=feature_extractor.image_std ) set_requires_grad(self.text_encoder , __magic_name__ ) set_requires_grad(self.clip_model , __magic_name__ ) def lowercase__ ( self : Tuple , __magic_name__ : Optional[Union[str, int]] = "auto" ) -> int: """simple docstring""" if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __snake_case : Optional[Any] = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(__magic_name__ ) def lowercase__ ( self : int ) -> Dict: """simple docstring""" self.enable_attention_slicing(__magic_name__ ) def lowercase__ ( self : Union[str, Any] ) -> List[str]: """simple docstring""" set_requires_grad(self.vae , __magic_name__ ) def lowercase__ ( self : int ) -> List[str]: """simple docstring""" set_requires_grad(self.vae , __magic_name__ ) def lowercase__ ( self : List[Any] ) -> Dict: """simple docstring""" set_requires_grad(self.unet , __magic_name__ ) def lowercase__ ( self : List[str] ) -> List[str]: """simple docstring""" set_requires_grad(self.unet , __magic_name__ ) def lowercase__ ( self : List[Any] , __magic_name__ : Union[str, Any] , __magic_name__ : List[str] , __magic_name__ : List[str] ) -> Optional[Any]: """simple docstring""" __snake_case : str = min(int(num_inference_steps * strength ) , __magic_name__ ) __snake_case : Optional[Any] = max(num_inference_steps - init_timestep , 0 ) __snake_case : Optional[Any] = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def lowercase__ ( self : List[Any] , __magic_name__ : Optional[Any] , __magic_name__ : int , __magic_name__ : List[Any] , __magic_name__ : Union[str, Any] , __magic_name__ : Any , __magic_name__ : List[Any]=None ) -> Optional[Any]: """simple docstring""" if not isinstance(__magic_name__ , torch.Tensor ): raise ValueError(f'''`image` has to be of type `torch.Tensor` but is {type(__magic_name__ )}''' ) __snake_case : Optional[int] = image.to(device=__magic_name__ , dtype=__magic_name__ ) if isinstance(__magic_name__ , __magic_name__ ): __snake_case : Any = [ self.vae.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(__magic_name__ ) ] __snake_case : List[Any] = torch.cat(__magic_name__ , dim=0 ) else: __snake_case : Any = self.vae.encode(__magic_name__ ).latent_dist.sample(__magic_name__ ) # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor __snake_case : Union[str, Any] = 0.18215 * init_latents __snake_case : Optional[Any] = init_latents.repeat_interleave(__magic_name__ , dim=0 ) __snake_case : Union[str, Any] = randn_tensor(init_latents.shape , generator=__magic_name__ , device=__magic_name__ , dtype=__magic_name__ ) # get latents __snake_case : List[Any] = self.scheduler.add_noise(__magic_name__ , __magic_name__ , __magic_name__ ) __snake_case : Tuple = init_latents return latents def lowercase__ ( self : Optional[int] , __magic_name__ : Dict ) -> List[Any]: """simple docstring""" __snake_case : Any = self.coca_transform(__magic_name__ ).unsqueeze(0 ) with torch.no_grad(), torch.cuda.amp.autocast(): __snake_case : Union[str, Any] = self.coca_model.generate(transformed_image.to(device=self.device , dtype=self.coca_model.dtype ) ) __snake_case : List[Any] = self.coca_tokenizer.decode(generated[0].cpu().numpy() ) return generated.split("""<end_of_text>""" )[0].replace("""<start_of_text>""" , """""" ).rstrip(""" .,""" ) def lowercase__ ( self : Optional[Any] , __magic_name__ : Any , __magic_name__ : List[Any] ) -> Union[str, Any]: """simple docstring""" __snake_case : Optional[Any] = self.feature_extractor.preprocess(__magic_name__ ) __snake_case : Optional[int] = torch.from_numpy(clip_image_input["""pixel_values"""][0] ).unsqueeze(0 ).to(self.device ).half() __snake_case : Tuple = self.clip_model.get_image_features(__magic_name__ ) __snake_case : str = image_embeddings_clip / image_embeddings_clip.norm(p=2 , dim=-1 , keepdim=__magic_name__ ) __snake_case : List[str] = image_embeddings_clip.repeat_interleave(__magic_name__ , dim=0 ) return image_embeddings_clip @torch.enable_grad() def lowercase__ ( self : str , __magic_name__ : int , __magic_name__ : Optional[Any] , __magic_name__ : Union[str, Any] , __magic_name__ : str , __magic_name__ : Any , __magic_name__ : Optional[int] , __magic_name__ : int , ) -> List[str]: """simple docstring""" __snake_case : Optional[Any] = latents.detach().requires_grad_() __snake_case : Dict = self.scheduler.scale_model_input(__magic_name__ , __magic_name__ ) # predict the noise residual __snake_case : Optional[Any] = self.unet(__magic_name__ , __magic_name__ , encoder_hidden_states=__magic_name__ ).sample if isinstance(self.scheduler , (PNDMScheduler, DDIMScheduler, DPMSolverMultistepScheduler) ): __snake_case : Tuple = self.scheduler.alphas_cumprod[timestep] __snake_case : Dict = 1 - alpha_prod_t # compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf __snake_case : List[str] = (latents - beta_prod_t ** 0.5 * noise_pred) / alpha_prod_t ** 0.5 __snake_case : Union[str, Any] = torch.sqrt(__magic_name__ ) __snake_case : Union[str, Any] = pred_original_sample * (fac) + latents * (1 - fac) elif isinstance(self.scheduler , __magic_name__ ): __snake_case : Any = self.scheduler.sigmas[index] __snake_case : Any = latents - sigma * noise_pred else: raise ValueError(f'''scheduler type {type(self.scheduler )} not supported''' ) # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor __snake_case : Tuple = 1 / 0.18215 * sample __snake_case : int = self.vae.decode(__magic_name__ ).sample __snake_case : str = (image / 2 + 0.5).clamp(0 , 1 ) __snake_case : str = transforms.Resize(self.feature_extractor_size )(__magic_name__ ) __snake_case : List[str] = self.normalize(__magic_name__ ).to(latents.dtype ) __snake_case : List[str] = self.clip_model.get_image_features(__magic_name__ ) __snake_case : Dict = image_embeddings_clip / image_embeddings_clip.norm(p=2 , dim=-1 , keepdim=__magic_name__ ) __snake_case : str = spherical_dist_loss(__magic_name__ , __magic_name__ ).mean() * clip_guidance_scale __snake_case : Union[str, Any] = -torch.autograd.grad(__magic_name__ , __magic_name__ )[0] if isinstance(self.scheduler , __magic_name__ ): __snake_case : str = latents.detach() + grads * (sigma**2) __snake_case : List[Any] = noise_pred_original else: __snake_case : str = noise_pred_original - torch.sqrt(__magic_name__ ) * grads return noise_pred, latents @torch.no_grad() def __call__( self : List[str] , __magic_name__ : Union[torch.FloatTensor, PIL.Image.Image] , __magic_name__ : Union[torch.FloatTensor, PIL.Image.Image] , __magic_name__ : Optional[str] = None , __magic_name__ : Optional[str] = None , __magic_name__ : Optional[int] = 5_12 , __magic_name__ : Optional[int] = 5_12 , __magic_name__ : float = 0.6 , __magic_name__ : Optional[int] = 50 , __magic_name__ : Optional[float] = 7.5 , __magic_name__ : Optional[int] = 1 , __magic_name__ : float = 0.0 , __magic_name__ : Optional[float] = 1_00 , __magic_name__ : Optional[torch.Generator] = None , __magic_name__ : Optional[str] = "pil" , __magic_name__ : bool = True , __magic_name__ : float = 0.8 , __magic_name__ : float = 0.1 , __magic_name__ : float = 0.1 , ) -> Any: """simple docstring""" if isinstance(__magic_name__ , __magic_name__ ) and len(__magic_name__ ) != batch_size: raise ValueError(f'''You have passed {batch_size} batch_size, but only {len(__magic_name__ )} generators.''' ) if height % 8 != 0 or width % 8 != 0: raise ValueError(f'''`height` and `width` have to be divisible by 8 but are {height} and {width}.''' ) if isinstance(__magic_name__ , torch.Generator ) and batch_size > 1: __snake_case : int = [generator] + [None] * (batch_size - 1) __snake_case : Tuple = [ ("model", self.coca_model is None), ("tokenizer", self.coca_tokenizer is None), ("transform", self.coca_transform is None), ] __snake_case : Dict = [x[0] for x in coca_is_none if x[1]] __snake_case : List[Any] = ", ".join(__magic_name__ ) # generate prompts with coca model if prompt is None if content_prompt is None: if len(__magic_name__ ): raise ValueError( f'''Content prompt is None and CoCa [{coca_is_none_str}] is None.''' f'''Set prompt or pass Coca [{coca_is_none_str}] to DiffusionPipeline.''' ) __snake_case : List[Any] = self.get_image_description(__magic_name__ ) if style_prompt is None: if len(__magic_name__ ): raise ValueError( f'''Style prompt is None and CoCa [{coca_is_none_str}] is None.''' f''' Set prompt or pass Coca [{coca_is_none_str}] to DiffusionPipeline.''' ) __snake_case : Union[str, Any] = self.get_image_description(__magic_name__ ) # get prompt text embeddings for content and style __snake_case : List[Any] = self.tokenizer( __magic_name__ , padding="""max_length""" , max_length=self.tokenizer.model_max_length , truncation=__magic_name__ , return_tensors="""pt""" , ) __snake_case : Union[str, Any] = self.text_encoder(content_text_input.input_ids.to(self.device ) )[0] __snake_case : Tuple = self.tokenizer( __magic_name__ , padding="""max_length""" , max_length=self.tokenizer.model_max_length , truncation=__magic_name__ , return_tensors="""pt""" , ) __snake_case : Any = self.text_encoder(style_text_input.input_ids.to(self.device ) )[0] __snake_case : Optional[Any] = slerp(__magic_name__ , __magic_name__ , __magic_name__ ) # duplicate text embeddings for each generation per prompt __snake_case : str = text_embeddings.repeat_interleave(__magic_name__ , dim=0 ) # set timesteps __snake_case : Tuple = "offset" in set(inspect.signature(self.scheduler.set_timesteps ).parameters.keys() ) __snake_case : str = {} if accepts_offset: __snake_case : Optional[Any] = 1 self.scheduler.set_timesteps(__magic_name__ , **__magic_name__ ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand self.scheduler.timesteps.to(self.device ) __snake_case : str = self.get_timesteps(__magic_name__ , __magic_name__ , self.device ) __snake_case : Dict = timesteps[:1].repeat(__magic_name__ ) # Preprocess image __snake_case : Dict = preprocess(__magic_name__ , __magic_name__ , __magic_name__ ) __snake_case : List[Any] = self.prepare_latents( __magic_name__ , __magic_name__ , __magic_name__ , text_embeddings.dtype , self.device , __magic_name__ ) __snake_case : Dict = preprocess(__magic_name__ , __magic_name__ , __magic_name__ ) __snake_case : Dict = self.prepare_latents( __magic_name__ , __magic_name__ , __magic_name__ , text_embeddings.dtype , self.device , __magic_name__ ) __snake_case : Optional[int] = slerp(__magic_name__ , __magic_name__ , __magic_name__ ) if clip_guidance_scale > 0: __snake_case : List[str] = self.get_clip_image_embeddings(__magic_name__ , __magic_name__ ) __snake_case : Optional[int] = self.get_clip_image_embeddings(__magic_name__ , __magic_name__ ) __snake_case : str = slerp( __magic_name__ , __magic_name__ , __magic_name__ ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. __snake_case : str = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: __snake_case : Any = content_text_input.input_ids.shape[-1] __snake_case : Tuple = self.tokenizer([""""""] , padding="""max_length""" , max_length=__magic_name__ , return_tensors="""pt""" ) __snake_case : List[Any] = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt __snake_case : Dict = uncond_embeddings.repeat_interleave(__magic_name__ , dim=0 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes __snake_case : str = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. __snake_case : Union[str, Any] = (batch_size, self.unet.config.in_channels, height // 8, width // 8) __snake_case : Tuple = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not work reproducibly on mps __snake_case : Union[str, Any] = torch.randn(__magic_name__ , generator=__magic_name__ , device="""cpu""" , dtype=__magic_name__ ).to( self.device ) else: __snake_case : Optional[Any] = torch.randn(__magic_name__ , generator=__magic_name__ , device=self.device , dtype=__magic_name__ ) else: if latents.shape != latents_shape: raise ValueError(f'''Unexpected latents shape, got {latents.shape}, expected {latents_shape}''' ) __snake_case : int = latents.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler __snake_case : List[str] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] __snake_case : Tuple = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __snake_case : Optional[int] = {} if accepts_eta: __snake_case : List[str] = eta # check if the scheduler accepts generator __snake_case : Optional[Any] = "generator" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) if accepts_generator: __snake_case : Any = generator with self.progress_bar(total=__magic_name__ ): for i, t in enumerate(__magic_name__ ): # expand the latents if we are doing classifier free guidance __snake_case : Optional[Any] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents __snake_case : Union[str, Any] = self.scheduler.scale_model_input(__magic_name__ , __magic_name__ ) # predict the noise residual __snake_case : List[Any] = self.unet(__magic_name__ , __magic_name__ , encoder_hidden_states=__magic_name__ ).sample # perform classifier free guidance if do_classifier_free_guidance: __snake_case : Optional[Any] = noise_pred.chunk(2 ) __snake_case : int = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # perform clip guidance if clip_guidance_scale > 0: __snake_case : Dict = ( text_embeddings.chunk(2 )[1] if do_classifier_free_guidance else text_embeddings ) __snake_case : Tuple = self.cond_fn( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , ) # compute the previous noisy sample x_t -> x_t-1 __snake_case : Tuple = self.scheduler.step(__magic_name__ , __magic_name__ , __magic_name__ , **__magic_name__ ).prev_sample # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor __snake_case : int = 1 / 0.18215 * latents __snake_case : Optional[int] = self.vae.decode(__magic_name__ ).sample __snake_case : int = (image / 2 + 0.5).clamp(0 , 1 ) __snake_case : Tuple = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __snake_case : str = self.numpy_to_pil(__magic_name__ ) if not return_dict: return (image, None) return StableDiffusionPipelineOutput(images=__magic_name__ , nsfw_content_detected=__magic_name__ )
26
from ....configuration_utils import PretrainedConfig from ....utils import logging _snake_case : int = logging.get_logger(__name__) _snake_case : str = { "Visual-Attention-Network/van-base": ( "https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json" ), } class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[Any] = "van" def __init__( self : Optional[int] , lowerCamelCase : Any=224 , lowerCamelCase : str=3 , lowerCamelCase : Any=[7, 3, 3, 3] , lowerCamelCase : Dict=[4, 2, 2, 2] , lowerCamelCase : List[Any]=[64, 128, 320, 512] , lowerCamelCase : str=[3, 3, 12, 3] , lowerCamelCase : Dict=[8, 8, 4, 4] , lowerCamelCase : Any="gelu" , lowerCamelCase : Optional[int]=0.02 , lowerCamelCase : Tuple=1E-6 , lowerCamelCase : Optional[int]=1E-2 , lowerCamelCase : int=0.0 , lowerCamelCase : Optional[Any]=0.0 , **lowerCamelCase : Optional[int] , ) -> int: super().__init__(**lowerCamelCase ) __snake_case : Union[str, Any] = image_size __snake_case : Any = num_channels __snake_case : Any = patch_sizes __snake_case : List[Any] = strides __snake_case : str = hidden_sizes __snake_case : Dict = depths __snake_case : Optional[int] = mlp_ratios __snake_case : Dict = hidden_act __snake_case : Union[str, Any] = initializer_range __snake_case : List[str] = layer_norm_eps __snake_case : Optional[int] = layer_scale_init_value __snake_case : List[Any] = drop_path_rate __snake_case : int = dropout_rate
81
0
'''simple docstring''' from __future__ import annotations UpperCAmelCase = { "A": ["B", "C", "E"], "B": ["A", "D", "E"], "C": ["A", "F", "G"], "D": ["B"], "E": ["A", "B", "D"], "F": ["C"], "G": ["C"], } class lowerCAmelCase : def __init__( self : Tuple , __lowercase : dict[str, list[str]] , __lowercase : str ): """simple docstring""" __lowercase =graph # mapping node to its parent in resulting breadth first tree __lowercase ={} __lowercase =source_vertex def snake_case ( self : Tuple ): """simple docstring""" __lowercase ={self.source_vertex} __lowercase =None __lowercase =[self.source_vertex] # first in first out queue while queue: __lowercase =queue.pop(0 ) for adjacent_vertex in self.graph[vertex]: if adjacent_vertex not in visited: visited.add(__lowercase ) __lowercase =vertex queue.append(__lowercase ) def snake_case ( self : Optional[int] , __lowercase : str ): """simple docstring""" if target_vertex == self.source_vertex: return self.source_vertex __lowercase =self.parent.get(__lowercase ) if target_vertex_parent is None: __lowercase =( f'''No path from vertex: {self.source_vertex} to vertex: {target_vertex}''' ) raise ValueError(__lowercase ) return self.shortest_path(__lowercase ) + f'''->{target_vertex}''' if __name__ == "__main__": UpperCAmelCase = Graph(graph, '''G''') g.breath_first_search() print(g.shortest_path('''D''')) print(g.shortest_path('''G''')) print(g.shortest_path('''Foo'''))
119
import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer _snake_case : Union[str, Any] = logging.getLogger(__name__) def lowerCAmelCase_ ( ): __snake_case : int = argparse.ArgumentParser( description="Prepare TFRecord shards from pre-tokenized samples of the wikitext dataset." ) parser.add_argument( "--dataset_name" , type=__lowerCamelCase , default="wikitext" , help="Name of the training. Explore datasets at: hf.co/datasets." , ) parser.add_argument( "--dataset_config" , type=__lowerCamelCase , default="wikitext-103-raw-v1" , help="Configuration name of the dataset." ) parser.add_argument( "--tokenizer_name_or_path" , type=__lowerCamelCase , default="sayakpaul/unigram-tokenizer-wikitext" , help="Tokenizer identifier. Can be a local filepath or a Hub identifier." , ) parser.add_argument( "--shard_size" , type=__lowerCamelCase , default=1_0_0_0 , help="Number of entries to go in a single shard." , ) parser.add_argument("--split" , type=__lowerCamelCase , default="train" , choices=["train", "test", "validation"] ) parser.add_argument( "--limit" , default=__lowerCamelCase , type=__lowerCamelCase , help="Limit the number of shards (used for debugging)." , ) parser.add_argument( "--max_length" , type=__lowerCamelCase , default=5_1_2 , help="Maximum sequence length. For training on TPUs, it helps to have a maximum" " sequence length that is a multiple of 8." , ) parser.add_argument( "--output_dir" , default="tf-tpu" , type=__lowerCamelCase , help="Output directory where the TFRecord shards will be saved. If the" " path is appended with `gs://` ('gs://tf-tpu', for example) then the TFRecord" " shards will be directly saved to a Google Cloud Storage bucket." , ) __snake_case : List[str] = parser.parse_args() return args def lowerCAmelCase_ ( __lowerCamelCase ): def fn(__lowerCamelCase ): return tokenizer(examples["text"] ) return fn def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = [] for i in range(len(tokenized_data["input_ids"] ) ): __snake_case : Tuple = { "input_ids": tf.train.Feature(intaa_list=tf.train.IntaaList(value=tokenized_data["input_ids"][i] ) ), "attention_mask": tf.train.Feature( intaa_list=tf.train.IntaaList(value=tokenized_data["attention_mask"][i] ) ), } __snake_case : List[Any] = tf.train.Features(feature=__lowerCamelCase ) __snake_case : str = tf.train.Example(features=__lowerCamelCase ) __snake_case : List[str] = example.SerializeToString() records.append(__lowerCamelCase ) return records def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Optional[int] = datasets.load_dataset(args.dataset_name , args.dataset_config , split=args.split ) if args.limit is not None: __snake_case : Optional[Any] = min(len(__lowerCamelCase ) , args.limit ) __snake_case : Dict = dataset.select(range(__lowerCamelCase ) ) print(F'Limiting the dataset to {args.limit} entries.' ) __snake_case : Dict = AutoTokenizer.from_pretrained(args.tokenizer_name_or_path ) # Handle output directory creation. # For serializing into a Google Cloud Storage Bucket, one needs to first # create a bucket. if "gs" not in args.output_dir: if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) __snake_case : Dict = os.path.join(args.output_dir , args.split ) if not os.path.exists(__lowerCamelCase ): os.makedirs(__lowerCamelCase ) else: __snake_case : str = os.path.join(args.output_dir , args.split ) # Tokenize the whole dataset at once. __snake_case : Any = tokenize_function(__lowerCamelCase ) __snake_case : Optional[Any] = dataset.map(__lowerCamelCase , batched=__lowerCamelCase , num_proc=4 , remove_columns=["text"] ) # We need to concatenate all our texts together, and then split the result # into chunks of a fixed size, which we will call block_size. To do this, we # will use the map method again, with the option batched=True. When we use batched=True, # the function we pass to map() will be passed multiple inputs at once, allowing us # to group them into more or fewer examples than we had in the input. # This allows us to create our new fixed-length samples. The advantage of this # method is that we don't lose a whole lot of content from the dataset compared to the # case where we simply tokenize with a pre-defined max_length. def group_texts(__lowerCamelCase ): # Concatenate all texts. __snake_case : List[str] = {k: sum(examples[k] , [] ) for k in examples.keys()} __snake_case : List[Any] = len(concatenated_examples[list(examples.keys() )[0]] ) # We drop the small remainder, though you could add padding instead if the model supports it # In this, as in all things, we advise you to follow your heart 🫀 __snake_case : Any = (total_length // args.max_length) * args.max_length # Split by chunks of max_len. __snake_case : int = { k: [t[i : i + args.max_length] for i in range(0 , __lowerCamelCase , args.max_length )] for k, t in concatenated_examples.items() } return result __snake_case : Any = dataset_tokenized.map(__lowerCamelCase , batched=__lowerCamelCase , batch_size=1_0_0_0 , num_proc=4 ) __snake_case : Optional[Any] = 0 __snake_case : Optional[Any] = 0 for shard in range(0 , len(__lowerCamelCase ) , args.shard_size ): __snake_case : List[str] = grouped_dataset[shard : shard + args.shard_size] __snake_case : Any = len(dataset_snapshot["input_ids"] ) __snake_case : List[Any] = os.path.join(__lowerCamelCase , F'dataset-{shard_count}-{records_containing}.tfrecord' ) __snake_case : Optional[Any] = get_serialized_examples(__lowerCamelCase ) with tf.io.TFRecordWriter(__lowerCamelCase ) as out_file: for i in range(len(__lowerCamelCase ) ): __snake_case : Union[str, Any] = serialized_examples[i] out_file.write(__lowerCamelCase ) print("Wrote file {} containing {} records".format(__lowerCamelCase , __lowerCamelCase ) ) shard_count += 1 total_records += records_containing with open(F'split-{args.split}-records-count.txt' , "w" ) as f: print(F'Total {args.split} records: {total_records}' , file=__lowerCamelCase ) if __name__ == "__main__": _snake_case : List[Any] = parse_args() main(args)
81
0
from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from ... import AutoBackbone from ...modeling_outputs import SemanticSegmenterOutput from ...modeling_utils import PreTrainedModel from ...utils import add_start_docstrings, add_start_docstrings_to_model_forward, replace_return_docstrings from ...utils.backbone_utils import BackboneMixin from .configuration_upernet import UperNetConfig SCREAMING_SNAKE_CASE__ : Optional[Any] = [ "openmmlab/upernet-convnext-tiny", # See all UperNet models at https://huggingface.co/models?filter=upernet ] # General docstring SCREAMING_SNAKE_CASE__ : Dict = "UperNetConfig" class lowerCAmelCase__ ( nn.Module ): def __init__( self : Tuple , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Union[int, Tuple[int, int]] , SCREAMING_SNAKE_CASE__ : Union[int, Tuple[int, int], str] = 0 , SCREAMING_SNAKE_CASE__ : bool = False , SCREAMING_SNAKE_CASE__ : Union[int, Tuple[int, int]] = 1 , ) -> None: super().__init__() __lowerCamelCase = nn.Convad( in_channels=SCREAMING_SNAKE_CASE__ , out_channels=SCREAMING_SNAKE_CASE__ , kernel_size=SCREAMING_SNAKE_CASE__ , padding=SCREAMING_SNAKE_CASE__ , bias=SCREAMING_SNAKE_CASE__ , dilation=SCREAMING_SNAKE_CASE__ , ) __lowerCamelCase = nn.BatchNormad(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = nn.ReLU() def __A ( self : List[Any] , SCREAMING_SNAKE_CASE__ : torch.Tensor ) -> torch.Tensor: __lowerCamelCase = self.conv(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = self.batch_norm(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = self.activation(SCREAMING_SNAKE_CASE__ ) return output class lowerCAmelCase__ ( nn.Module ): def __init__( self : str , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int ) -> None: super().__init__() __lowerCamelCase = [ nn.AdaptiveAvgPoolad(SCREAMING_SNAKE_CASE__ ), UperNetConvModule(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , kernel_size=1 ), ] for i, layer in enumerate(self.layers ): self.add_module(str(SCREAMING_SNAKE_CASE__ ) , SCREAMING_SNAKE_CASE__ ) def __A ( self : Dict , SCREAMING_SNAKE_CASE__ : torch.Tensor ) -> torch.Tensor: __lowerCamelCase = input for layer in self.layers: __lowerCamelCase = layer(SCREAMING_SNAKE_CASE__ ) return hidden_state class lowerCAmelCase__ ( nn.Module ): def __init__( self : Any , SCREAMING_SNAKE_CASE__ : Tuple[int, ...] , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : bool ) -> None: super().__init__() __lowerCamelCase = pool_scales __lowerCamelCase = align_corners __lowerCamelCase = in_channels __lowerCamelCase = channels __lowerCamelCase = [] for i, pool_scale in enumerate(SCREAMING_SNAKE_CASE__ ): __lowerCamelCase = UperNetPyramidPoolingBlock(pool_scale=SCREAMING_SNAKE_CASE__ , in_channels=SCREAMING_SNAKE_CASE__ , channels=SCREAMING_SNAKE_CASE__ ) self.blocks.append(SCREAMING_SNAKE_CASE__ ) self.add_module(str(SCREAMING_SNAKE_CASE__ ) , SCREAMING_SNAKE_CASE__ ) def __A ( self : int , SCREAMING_SNAKE_CASE__ : torch.Tensor ) -> List[torch.Tensor]: __lowerCamelCase = [] for ppm in self.blocks: __lowerCamelCase = ppm(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = nn.functional.interpolate( SCREAMING_SNAKE_CASE__ , size=x.size()[2:] , mode='''bilinear''' , align_corners=self.align_corners ) ppm_outs.append(SCREAMING_SNAKE_CASE__ ) return ppm_outs class lowerCAmelCase__ ( nn.Module ): def __init__( self : int , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : str ) -> List[str]: super().__init__() __lowerCamelCase = config __lowerCamelCase = config.pool_scales # e.g. (1, 2, 3, 6) __lowerCamelCase = in_channels __lowerCamelCase = config.hidden_size __lowerCamelCase = False __lowerCamelCase = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) # PSP Module __lowerCamelCase = UperNetPyramidPoolingModule( self.pool_scales , self.in_channels[-1] , self.channels , align_corners=self.align_corners , ) __lowerCamelCase = UperNetConvModule( self.in_channels[-1] + len(self.pool_scales ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) # FPN Module __lowerCamelCase = nn.ModuleList() __lowerCamelCase = nn.ModuleList() for in_channels in self.in_channels[:-1]: # skip the top layer __lowerCamelCase = UperNetConvModule(SCREAMING_SNAKE_CASE__ , self.channels , kernel_size=1 ) __lowerCamelCase = UperNetConvModule(self.channels , self.channels , kernel_size=3 , padding=1 ) self.lateral_convs.append(SCREAMING_SNAKE_CASE__ ) self.fpn_convs.append(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = UperNetConvModule( len(self.in_channels ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) def __A ( self : List[str] ) -> Optional[Any]: self.apply(self._init_weights ) def __A ( self : Tuple , SCREAMING_SNAKE_CASE__ : List[Any] ) -> str: if isinstance(SCREAMING_SNAKE_CASE__ , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __A ( self : List[Any] , SCREAMING_SNAKE_CASE__ : Tuple ) -> Optional[int]: __lowerCamelCase = inputs[-1] __lowerCamelCase = [x] psp_outs.extend(self.psp_modules(SCREAMING_SNAKE_CASE__ ) ) __lowerCamelCase = torch.cat(SCREAMING_SNAKE_CASE__ , dim=1 ) __lowerCamelCase = self.bottleneck(SCREAMING_SNAKE_CASE__ ) return output def __A ( self : int , SCREAMING_SNAKE_CASE__ : torch.Tensor ) -> torch.Tensor: # build laterals __lowerCamelCase = [lateral_conv(encoder_hidden_states[i] ) for i, lateral_conv in enumerate(self.lateral_convs )] laterals.append(self.psp_forward(SCREAMING_SNAKE_CASE__ ) ) # build top-down path __lowerCamelCase = len(SCREAMING_SNAKE_CASE__ ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __lowerCamelCase = laterals[i - 1].shape[2:] __lowerCamelCase = laterals[i - 1] + nn.functional.interpolate( laterals[i] , size=SCREAMING_SNAKE_CASE__ , mode='''bilinear''' , align_corners=self.align_corners ) # build outputs __lowerCamelCase = [self.fpn_convs[i](laterals[i] ) for i in range(used_backbone_levels - 1 )] # append psp feature fpn_outs.append(laterals[-1] ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __lowerCamelCase = nn.functional.interpolate( fpn_outs[i] , size=fpn_outs[0].shape[2:] , mode='''bilinear''' , align_corners=self.align_corners ) __lowerCamelCase = torch.cat(SCREAMING_SNAKE_CASE__ , dim=1 ) __lowerCamelCase = self.fpn_bottleneck(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = self.classifier(SCREAMING_SNAKE_CASE__ ) return output class lowerCAmelCase__ ( nn.Module ): def __init__( self : Any , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : int = 2 , SCREAMING_SNAKE_CASE__ : int = 3 , SCREAMING_SNAKE_CASE__ : Union[int, Tuple[int, int]] = 1 ) -> None: super().__init__() __lowerCamelCase = config __lowerCamelCase = config.auxiliary_in_channels __lowerCamelCase = config.auxiliary_channels __lowerCamelCase = config.auxiliary_num_convs __lowerCamelCase = config.auxiliary_concat_input __lowerCamelCase = in_index __lowerCamelCase = (kernel_size // 2) * dilation __lowerCamelCase = [] convs.append( UperNetConvModule( self.in_channels , self.channels , kernel_size=SCREAMING_SNAKE_CASE__ , padding=SCREAMING_SNAKE_CASE__ , dilation=SCREAMING_SNAKE_CASE__ ) ) for i in range(self.num_convs - 1 ): convs.append( UperNetConvModule( self.channels , self.channels , kernel_size=SCREAMING_SNAKE_CASE__ , padding=SCREAMING_SNAKE_CASE__ , dilation=SCREAMING_SNAKE_CASE__ ) ) if self.num_convs == 0: __lowerCamelCase = nn.Identity() else: __lowerCamelCase = nn.Sequential(*SCREAMING_SNAKE_CASE__ ) if self.concat_input: __lowerCamelCase = UperNetConvModule( self.in_channels + self.channels , self.channels , kernel_size=SCREAMING_SNAKE_CASE__ , padding=kernel_size // 2 ) __lowerCamelCase = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) def __A ( self : Dict ) -> Optional[Any]: self.apply(self._init_weights ) def __A ( self : Tuple , SCREAMING_SNAKE_CASE__ : Tuple ) -> Optional[int]: if isinstance(SCREAMING_SNAKE_CASE__ , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __A ( self : Optional[int] , SCREAMING_SNAKE_CASE__ : torch.Tensor ) -> torch.Tensor: # just take the relevant feature maps __lowerCamelCase = encoder_hidden_states[self.in_index] __lowerCamelCase = self.convs(SCREAMING_SNAKE_CASE__ ) if self.concat_input: __lowerCamelCase = self.conv_cat(torch.cat([hidden_states, output] , dim=1 ) ) __lowerCamelCase = self.classifier(SCREAMING_SNAKE_CASE__ ) return output class lowerCAmelCase__ ( _lowerCAmelCase ): a__ : Optional[Any] = UperNetConfig a__ : int = "pixel_values" a__ : str = True def __A ( self : Union[str, Any] , SCREAMING_SNAKE_CASE__ : List[Any] ) -> List[Any]: if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): module.backbone.init_weights() module.decode_head.init_weights() module.auxiliary_head.init_weights() def __A ( self : Optional[Any] ) -> List[str]: self.backbone.init_weights() self.decode_head.init_weights() self.auxiliary_head.init_weights() def __A ( self : List[Any] , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Optional[Any]=False ) -> Dict: if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): __lowerCamelCase = value SCREAMING_SNAKE_CASE__ : Dict = R"\n Parameters:\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n config ([`UperNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" SCREAMING_SNAKE_CASE__ : Tuple = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using\n [`AutoImageProcessor`]. See [`SegformerImageProcessor.__call__`] for details.\n output_attentions (`bool`, *optional*):\n Whether or not to return the attentions tensors of all attention layers in case the backbone has them. See\n `attentions` under returned tensors for more detail.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers of the backbone. See `hidden_states` under\n returned tensors for more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n" @add_start_docstrings( """UperNet framework leveraging any vision backbone e.g. for ADE20k, CityScapes.""" , _lowerCAmelCase , ) class lowerCAmelCase__ ( _lowerCAmelCase ): def __init__( self : Dict , SCREAMING_SNAKE_CASE__ : int ) -> Optional[int]: super().__init__(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = AutoBackbone.from_config(config.backbone_config ) # Semantic segmentation head(s) __lowerCamelCase = UperNetHead(SCREAMING_SNAKE_CASE__ , in_channels=self.backbone.channels ) __lowerCamelCase = UperNetFCNHead(SCREAMING_SNAKE_CASE__ ) if config.use_auxiliary_head else None # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UPERNET_INPUTS_DOCSTRING.format('''batch_size, sequence_length''' ) ) @replace_return_docstrings(output_type=SCREAMING_SNAKE_CASE__ , config_class=_CONFIG_FOR_DOC ) def __A ( self : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Optional[torch.Tensor] = None , SCREAMING_SNAKE_CASE__ : Optional[bool] = None , SCREAMING_SNAKE_CASE__ : Optional[bool] = None , SCREAMING_SNAKE_CASE__ : Optional[torch.Tensor] = None , SCREAMING_SNAKE_CASE__ : Optional[bool] = None , ) -> Union[tuple, SemanticSegmenterOutput]: __lowerCamelCase = return_dict if return_dict is not None else self.config.use_return_dict __lowerCamelCase = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __lowerCamelCase = output_attentions if output_attentions is not None else self.config.output_attentions __lowerCamelCase = self.backbone.forward_with_filtered_kwargs( SCREAMING_SNAKE_CASE__ , output_hidden_states=SCREAMING_SNAKE_CASE__ , output_attentions=SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = outputs.feature_maps __lowerCamelCase = self.decode_head(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = nn.functional.interpolate(SCREAMING_SNAKE_CASE__ , size=pixel_values.shape[2:] , mode='''bilinear''' , align_corners=SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = None if self.auxiliary_head is not None: __lowerCamelCase = self.auxiliary_head(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = nn.functional.interpolate( SCREAMING_SNAKE_CASE__ , size=pixel_values.shape[2:] , mode='''bilinear''' , align_corners=SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = None if labels is not None: if self.config.num_labels == 1: raise ValueError('''The number of labels should be greater than one''' ) else: # compute weighted loss __lowerCamelCase = CrossEntropyLoss(ignore_index=self.config.loss_ignore_index ) __lowerCamelCase = loss_fct(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = loss_fct(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = main_loss + self.config.auxiliary_loss_weight * auxiliary_loss if not return_dict: if output_hidden_states: __lowerCamelCase = (logits,) + outputs[1:] else: __lowerCamelCase = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return SemanticSegmenterOutput( loss=SCREAMING_SNAKE_CASE__ , logits=SCREAMING_SNAKE_CASE__ , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
298
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) _snake_case : List[Any] = "\\n Text data.\n Second line of data." _snake_case : Tuple = "file" @pytest.fixture(scope="session" ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / (FILE_PATH + ".zstd") __snake_case : Optional[Any] = bytes(__lowerCamelCase , "utf-8" ) with zstd.open(__lowerCamelCase , "wb" ) as f: f.write(__lowerCamelCase ) return path @pytest.fixture def lowerCAmelCase_ ( __lowerCamelCase ): with open(os.path.join(tmpfs.local_root_dir , __lowerCamelCase ) , "w" ) as f: f.write(__lowerCamelCase ) return FILE_PATH @pytest.mark.parametrize("compression_format" , ["gzip", "xz", "zstd"] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Optional[int] = {"gzip": gz_file, "xz": xz_file, "zstd": zstd_path} __snake_case : str = input_paths[compression_format] __snake_case : Optional[Any] = tmp_path / "cache" __snake_case : Optional[int] = DownloadConfig(cache_dir=__lowerCamelCase , extract_compressed_file=__lowerCamelCase ) __snake_case : Union[str, Any] = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) with open(__lowerCamelCase ) as f: __snake_case : Dict = f.read() with open(__lowerCamelCase ) as f: __snake_case : Tuple = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("default_extracted" , [True, False] ) @pytest.mark.parametrize("default_cache_dir" , [True, False] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Tuple = "custom_cache" __snake_case : List[str] = "custom_extracted_dir" __snake_case : Any = tmp_path / "custom_extracted_path" if default_extracted: __snake_case : List[Any] = ("downloads" if default_cache_dir else custom_cache_dir, "extracted") else: monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_DIR" , __lowerCamelCase ) monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_PATH" , str(__lowerCamelCase ) ) __snake_case : Optional[Any] = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) __snake_case : Optional[int] = xz_file __snake_case : Optional[int] = ( DownloadConfig(extract_compressed_file=__lowerCamelCase ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=__lowerCamelCase ) ) __snake_case : str = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) assert Path(__lowerCamelCase ).parent.parts[-2:] == expected def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : Optional[Any] = str(Path(__lowerCamelCase ).resolve() ) assert cached_path(__lowerCamelCase ) == text_file # relative path __snake_case : Any = str(Path(__lowerCamelCase ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(__lowerCamelCase ) == text_file def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : List[Any] = str(tmp_path.resolve() / "__missing_file__.txt" ) with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) # relative path __snake_case : Optional[int] = "./__missing_file__.txt" with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : str = get_from_cache(F'tmp://{tmpfs_file}' ) with open(__lowerCamelCase ) as f: __snake_case : Union[str, Any] = f.read() assert output_file_content == FILE_CONTENT @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( ): with pytest.raises(__lowerCamelCase ): cached_path("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[Any] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): http_get("https://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): http_head("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[str] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): ftp_get("ftp://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): ftp_head("ftp://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): fsspec_get("s3://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): fsspec_head("s3://huggingface.co" )
81
0
"""simple docstring""" from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE =logging.get_logger(__name__) __SCREAMING_SNAKE_CASE ={ "huggingface/informer-tourism-monthly": ( "https://huggingface.co/huggingface/informer-tourism-monthly/resolve/main/config.json" ), # See all Informer models at https://huggingface.co/models?filter=informer } class UpperCamelCase ( _lowerCAmelCase ): lowercase = "informer" lowercase = { "hidden_size": "d_model", "num_attention_heads": "encoder_attention_heads", "num_hidden_layers": "encoder_layers", } def __init__( self ,__UpperCamelCase = None ,__UpperCamelCase = None ,__UpperCamelCase = "student_t" ,__UpperCamelCase = "nll" ,__UpperCamelCase = 1 ,__UpperCamelCase = None ,__UpperCamelCase = "mean" ,__UpperCamelCase = 0 ,__UpperCamelCase = 0 ,__UpperCamelCase = 0 ,__UpperCamelCase = 0 ,__UpperCamelCase = None ,__UpperCamelCase = None ,__UpperCamelCase = 64 ,__UpperCamelCase = 32 ,__UpperCamelCase = 32 ,__UpperCamelCase = 2 ,__UpperCamelCase = 2 ,__UpperCamelCase = 2 ,__UpperCamelCase = 2 ,__UpperCamelCase = True ,__UpperCamelCase = "gelu" ,__UpperCamelCase = 0.05 ,__UpperCamelCase = 0.1 ,__UpperCamelCase = 0.1 ,__UpperCamelCase = 0.1 ,__UpperCamelCase = 0.1 ,__UpperCamelCase = 100 ,__UpperCamelCase = 0.02 ,__UpperCamelCase=True ,__UpperCamelCase = "prob" ,__UpperCamelCase = 5 ,__UpperCamelCase = True ,**__UpperCamelCase ,) -> str: '''simple docstring''' lowercase_ : Any = prediction_length lowercase_ : str = context_length or prediction_length lowercase_ : Any = distribution_output lowercase_ : Optional[Any] = loss lowercase_ : List[Any] = input_size lowercase_ : List[Any] = num_time_features lowercase_ : Union[str, Any] = lags_sequence if lags_sequence is not None else [1, 2, 3, 4, 5, 6, 7] lowercase_ : Tuple = scaling lowercase_ : List[Any] = num_dynamic_real_features lowercase_ : Tuple = num_static_real_features lowercase_ : Union[str, Any] = num_static_categorical_features # set cardinality if cardinality and num_static_categorical_features > 0: if len(__UpperCamelCase ) != num_static_categorical_features: raise ValueError( 'The cardinality should be a list of the same length as `num_static_categorical_features`' ) lowercase_ : Tuple = cardinality else: lowercase_ : Tuple = [0] # set embedding_dimension if embedding_dimension and num_static_categorical_features > 0: if len(__UpperCamelCase ) != num_static_categorical_features: raise ValueError( 'The embedding dimension should be a list of the same length as `num_static_categorical_features`' ) lowercase_ : Tuple = embedding_dimension else: lowercase_ : Union[str, Any] = [min(50 ,(cat + 1) // 2 ) for cat in self.cardinality] lowercase_ : List[str] = num_parallel_samples # Transformer architecture configuration lowercase_ : Optional[int] = input_size * len(self.lags_sequence ) + self._number_of_features lowercase_ : str = d_model lowercase_ : int = encoder_attention_heads lowercase_ : Any = decoder_attention_heads lowercase_ : Any = encoder_ffn_dim lowercase_ : List[Any] = decoder_ffn_dim lowercase_ : Dict = encoder_layers lowercase_ : Dict = decoder_layers lowercase_ : Union[str, Any] = dropout lowercase_ : Tuple = attention_dropout lowercase_ : int = activation_dropout lowercase_ : List[str] = encoder_layerdrop lowercase_ : Optional[int] = decoder_layerdrop lowercase_ : List[str] = activation_function lowercase_ : Dict = init_std lowercase_ : Union[str, Any] = use_cache # Informer lowercase_ : Optional[Any] = attention_type lowercase_ : List[Any] = sampling_factor lowercase_ : Optional[int] = distil super().__init__(is_encoder_decoder=__UpperCamelCase ,**__UpperCamelCase ) @property def _UpperCAmelCase ( self ) -> int: '''simple docstring''' return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
425
_snake_case : Optional[int] = {"a": ["c", "b"], "b": ["d", "e"], "c": [], "d": [], "e": []} _snake_case : Dict = ["a", "b", "c", "d", "e"] def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : List[str] = start # add current to visited visited.append(__lowerCamelCase ) __snake_case : List[Any] = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: __snake_case : Tuple = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # if all neighbors visited add current to sort sort.append(__lowerCamelCase ) # if all vertices haven't been visited select a new one to visit if len(__lowerCamelCase ) != len(__lowerCamelCase ): for vertice in vertices: if vertice not in visited: __snake_case : int = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # return sort return sort if __name__ == "__main__": _snake_case : List[Any] = topological_sort("a", [], []) print(sort)
81
0
'''simple docstring''' from __future__ import annotations def __snake_case ( _UpperCAmelCase : int): if len(__lowerCamelCase) == 0: return array UpperCamelCase = min(__lowerCamelCase), max(__lowerCamelCase) # Compute the variables UpperCamelCase = _max - _min + 1 UpperCamelCase = [0] * holes_range, [0] * holes_range # Make the sorting. for i in array: UpperCamelCase = i - _min UpperCamelCase = i holes_repeat[index] += 1 # Makes the array back by replacing the numbers. UpperCamelCase = 0 for i in range(__lowerCamelCase): while holes_repeat[i] > 0: UpperCamelCase = holes[i] index += 1 holes_repeat[i] -= 1 # Returns the sorted array. return array if __name__ == "__main__": import doctest doctest.testmod() snake_case_ : Optional[Any] = input('Enter numbers separated by comma:\n') snake_case_ : List[str] = [int(x) for x in user_input.split(',')] print(pigeon_sort(unsorted))
212
def lowerCAmelCase_ ( ): return [ a * b * (1_0_0_0 - a - b) for a in range(1 , 9_9_9 ) for b in range(__lowerCamelCase , 9_9_9 ) if (a * a + b * b == (1_0_0_0 - a - b) ** 2) ][0] if __name__ == "__main__": print(f'''{solution() = }''')
81
0
from __future__ import annotations from collections.abc import Callable from typing import Any, Generic, TypeVar _lowerCAmelCase : Union[str, Any] = TypeVar('''T''') class __magic_name__ ( Generic[T] ): """simple docstring""" def __init__( self :Tuple , snake_case :list[T] , snake_case :Callable[[T, T], T] ): '''simple docstring''' A_ : Any | T = None A_ : int = len(snake_case ) A_ : list[T] = [any_type for _ in range(self.N )] + arr A_ : Tuple = fnc self.build() def SCREAMING_SNAKE_CASE ( self :Dict ): '''simple docstring''' for p in range(self.N - 1 , 0 , -1 ): A_ : List[str] = self.fn(self.st[p * 2] , self.st[p * 2 + 1] ) def SCREAMING_SNAKE_CASE ( self :Optional[Any] , snake_case :int , snake_case :T ): '''simple docstring''' p += self.N A_ : str = v while p > 1: A_ : Dict = p // 2 A_ : Optional[Any] = self.fn(self.st[p * 2] , self.st[p * 2 + 1] ) def SCREAMING_SNAKE_CASE ( self :List[str] , snake_case :int , snake_case :int ): # noqa: E741 '''simple docstring''' A_ : Optional[int] = l + self.N, r + self.N A_ : T | None = None while l <= r: if l % 2 == 1: A_ : Optional[Any] = self.st[l] if res is None else self.fn(snake_case , self.st[l] ) if r % 2 == 0: A_ : int = self.st[r] if res is None else self.fn(snake_case , self.st[r] ) A_ : List[str] = (l + 1) // 2, (r - 1) // 2 return res if __name__ == "__main__": from functools import reduce _lowerCAmelCase : Union[str, Any] = [1, 10, -2, 9, -3, 8, 4, -7, 5, 6, 11, -12] _lowerCAmelCase : Dict = { 0: 7, 1: 2, 2: 6, 3: -14, 4: 5, 5: 4, 6: 7, 7: -10, 8: 9, 9: 10, 10: 12, 11: 1, } _lowerCAmelCase : Optional[Any] = SegmentTree(test_array, min) _lowerCAmelCase : Optional[Any] = SegmentTree(test_array, max) _lowerCAmelCase : Optional[Any] = SegmentTree(test_array, lambda a, b: a + b) def __snake_case ( ) -> List[Any]: for i in range(len(__lowerCamelCase ) ): for j in range(__lowerCamelCase , len(__lowerCamelCase ) ): A_ : Optional[int] = reduce(__lowerCamelCase , test_array[i : j + 1] ) A_ : List[Any] = reduce(__lowerCamelCase , test_array[i : j + 1] ) A_ : Union[str, Any] = reduce(lambda _lowerCAmelCase , _lowerCAmelCase : a + b , test_array[i : j + 1] ) assert min_range == min_segment_tree.query(__lowerCamelCase , __lowerCamelCase ) assert max_range == max_segment_tree.query(__lowerCamelCase , __lowerCamelCase ) assert sum_range == sum_segment_tree.query(__lowerCamelCase , __lowerCamelCase ) test_all_segments() for index, value in test_updates.items(): _lowerCAmelCase : List[Any] = value min_segment_tree.update(index, value) max_segment_tree.update(index, value) sum_segment_tree.update(index, value) test_all_segments()
454
from math import sqrt def lowerCAmelCase_ ( __lowerCamelCase = 1_0_0_0_0_0_0 ): __snake_case : int = 0 __snake_case : int = 0 __snake_case : int while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(__lowerCamelCase , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(f'''{solution() = }''')
81
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowerCamelCase__ = { "configuration_bridgetower": [ "BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP", "BridgeTowerConfig", "BridgeTowerTextConfig", "BridgeTowerVisionConfig", ], "processing_bridgetower": ["BridgeTowerProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ["BridgeTowerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ "BRIDGETOWER_PRETRAINED_MODEL_ARCHIVE_LIST", "BridgeTowerForContrastiveLearning", "BridgeTowerForImageAndTextRetrieval", "BridgeTowerForMaskedLM", "BridgeTowerModel", "BridgeTowerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_bridgetower import ( BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP, BridgeTowerConfig, BridgeTowerTextConfig, BridgeTowerVisionConfig, ) from .processing_bridgetower import BridgeTowerProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_bridgetower import BridgeTowerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bridgetower import ( BRIDGETOWER_PRETRAINED_MODEL_ARCHIVE_LIST, BridgeTowerForContrastiveLearning, BridgeTowerForImageAndTextRetrieval, BridgeTowerForMaskedLM, BridgeTowerModel, BridgeTowerPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["__file__"], _import_structure)
574
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class a (_lowerCAmelCase ): """simple docstring""" def __snake_case ( self : str ) -> str: __snake_case : Tuple = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCamelCase , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "neck_hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "num_attention_heads" ) ) class a : """simple docstring""" def __init__( self : Optional[int] , lowerCamelCase : List[str] , lowerCamelCase : Tuple=13 , lowerCamelCase : str=32 , lowerCamelCase : Dict=2 , lowerCamelCase : List[str]=3 , lowerCamelCase : Any=640 , lowerCamelCase : Optional[Any]=4 , lowerCamelCase : Tuple="silu" , lowerCamelCase : int=3 , lowerCamelCase : Dict=32 , lowerCamelCase : str=0.1 , lowerCamelCase : Optional[int]=0.1 , lowerCamelCase : Optional[Any]=0.1 , lowerCamelCase : Dict=0.02 , lowerCamelCase : Union[str, Any]=True , lowerCamelCase : Optional[int]=True , lowerCamelCase : Union[str, Any]=10 , lowerCamelCase : int=None , ) -> str: __snake_case : Optional[Any] = parent __snake_case : Optional[Any] = batch_size __snake_case : Any = image_size __snake_case : List[Any] = patch_size __snake_case : Any = num_channels __snake_case : Union[str, Any] = last_hidden_size __snake_case : Any = num_attention_heads __snake_case : Any = hidden_act __snake_case : Tuple = conv_kernel_size __snake_case : Any = output_stride __snake_case : Any = hidden_dropout_prob __snake_case : List[Any] = attention_probs_dropout_prob __snake_case : Optional[Any] = classifier_dropout_prob __snake_case : Union[str, Any] = use_labels __snake_case : Optional[int] = is_training __snake_case : Dict = num_labels __snake_case : Any = initializer_range __snake_case : Optional[int] = scope def __snake_case ( self : str ) -> Union[str, Any]: __snake_case : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : List[Any] = None __snake_case : Optional[int] = None if self.use_labels: __snake_case : Optional[int] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : Any = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __snake_case : Optional[Any] = self.get_config() return config, pixel_values, labels, pixel_labels def __snake_case ( self : Any ) -> Union[str, Any]: return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __snake_case ( self : Any , lowerCamelCase : Tuple , lowerCamelCase : int , lowerCamelCase : Tuple , lowerCamelCase : Optional[int] ) -> Dict: __snake_case : List[Any] = MobileViTModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[str] = model(lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[Any] , lowerCamelCase : List[str] , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] , lowerCamelCase : Tuple ) -> List[str]: __snake_case : str = self.num_labels __snake_case : List[Any] = MobileViTForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : Dict , lowerCamelCase : Any , lowerCamelCase : Any , lowerCamelCase : Optional[Any] , lowerCamelCase : Dict ) -> Dict: __snake_case : Union[str, Any] = self.num_labels __snake_case : Optional[int] = MobileViTForSemanticSegmentation(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Tuple = model(lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[int] ) -> List[Any]: __snake_case : Optional[Any] = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case , __snake_case : Union[str, Any] = config_and_inputs __snake_case : Dict = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) __UpperCAmelCase : Optional[Any] = ( { "feature-extraction": MobileViTModel, "image-classification": MobileViTForImageClassification, "image-segmentation": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) __UpperCAmelCase : List[str] = False __UpperCAmelCase : int = False __UpperCAmelCase : Optional[int] = False __UpperCAmelCase : Optional[int] = False def __snake_case ( self : Optional[int] ) -> Dict: __snake_case : Tuple = MobileViTModelTester(self ) __snake_case : Any = MobileViTConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase ) def __snake_case ( self : Optional[int] ) -> Dict: self.config_tester.run_common_tests() @unittest.skip(reason="MobileViT does not use inputs_embeds" ) def __snake_case ( self : Dict ) -> Any: pass @unittest.skip(reason="MobileViT does not support input and output embeddings" ) def __snake_case ( self : Dict ) -> List[Any]: pass @unittest.skip(reason="MobileViT does not output attentions" ) def __snake_case ( self : int ) -> Dict: pass def __snake_case ( self : int ) -> Union[str, Any]: __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Optional[Any] = model_class(lowerCamelCase ) __snake_case : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : int = [*signature.parameters.keys()] __snake_case : List[str] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __snake_case ( self : int ) -> Tuple: pass def __snake_case ( self : Any ) -> Tuple: __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def __snake_case ( self : Any ) -> str: def check_hidden_states_output(lowerCamelCase : Union[str, Any] , lowerCamelCase : Dict , lowerCamelCase : Any ): __snake_case : int = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : int = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : Union[str, Any] = outputs.hidden_states __snake_case : int = 5 self.assertEqual(len(lowerCamelCase ) , lowerCamelCase ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. __snake_case : List[Any] = 2 for i in range(len(lowerCamelCase ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) __snake_case , __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : str = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : List[Any] = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Any ) -> Any: __snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) def __snake_case ( self : List[str] ) -> List[str]: __snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowerCamelCase ) @slow def __snake_case ( self : List[str] ) -> Any: for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : List[str] = MobileViTModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def lowerCAmelCase_ ( ): __snake_case : Optional[int] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a (unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self : str ) -> Dict: return MobileViTImageProcessor.from_pretrained("apple/mobilevit-xx-small" ) if is_vision_available() else None @slow def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : Tuple = MobileViTForImageClassification.from_pretrained("apple/mobilevit-xx-small" ).to(lowerCamelCase ) __snake_case : Optional[Any] = self.default_image_processor __snake_case : Union[str, Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Dict = model(**lowerCamelCase ) # verify the logits __snake_case : Union[str, Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([-1.93_64, -1.23_27, -0.46_53] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : str ) -> Optional[int]: __snake_case : Optional[int] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : str = model.to(lowerCamelCase ) __snake_case : int = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Optional[int] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : List[str] = model(**lowerCamelCase ) __snake_case : Union[str, Any] = outputs.logits # verify the logits __snake_case : Tuple = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , lowerCamelCase ) __snake_case : Union[str, Any] = torch.tensor( [ [[6.97_13, 6.97_86, 7.24_22], [7.28_93, 7.28_25, 7.44_46], [7.65_80, 7.87_97, 7.94_20]], [[-10.68_69, -10.32_50, -10.34_71], [-10.42_28, -9.98_68, -9.71_32], [-11.04_05, -11.02_21, -10.73_18]], [[-3.30_89, -2.85_39, -2.67_40], [-3.27_06, -2.56_21, -2.51_08], [-3.25_34, -2.66_15, -2.66_51]], ] , device=lowerCamelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : Union[str, Any] ) -> Optional[int]: __snake_case : Optional[Any] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Tuple = model.to(lowerCamelCase ) __snake_case : Dict = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : List[Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Any = model(**lowerCamelCase ) __snake_case : Dict = outputs.logits.detach().cpu() __snake_case : Any = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase , target_sizes=[(50, 60)] ) __snake_case : int = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , lowerCamelCase ) __snake_case : List[str] = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase ) __snake_case : Optional[int] = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , lowerCamelCase )
81
0
'''simple docstring''' import logging import os from typing import Dict, List, Optional, Union import torch import torch.nn as nn from accelerate.utils.imports import ( is_abit_bnb_available, is_abit_bnb_available, is_bnb_available, ) from ..big_modeling import dispatch_model, init_empty_weights from .dataclasses import BnbQuantizationConfig from .modeling import ( find_tied_parameters, get_balanced_memory, infer_auto_device_map, load_checkpoint_in_model, offload_weight, set_module_tensor_to_device, ) if is_bnb_available(): import bitsandbytes as bnb from copy import deepcopy __a: str = logging.getLogger(__name__) def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = False , ): lowercase__ : Union[str, Any] = bnb_quantization_config.load_in_abit lowercase__ : List[str] = bnb_quantization_config.load_in_abit if load_in_abit and not is_abit_bnb_available(): raise ImportError( '''You have a version of `bitsandbytes` that is not compatible with 8bit quantization,''' ''' make sure you have the latest version of `bitsandbytes` installed.''' ) if load_in_abit and not is_abit_bnb_available(): raise ValueError( '''You have a version of `bitsandbytes` that is not compatible with 4bit quantization,''' '''make sure you have the latest version of `bitsandbytes` installed.''' ) lowercase__ : Tuple = [] # custom device map if isinstance(__lowerCamelCase , __lowerCamelCase ) and len(device_map.keys() ) > 1: lowercase__ : int = [key for key, value in device_map.items() if value in ["disk", "cpu"]] # We keep some modules such as the lm_head in their original dtype for numerical stability reasons if bnb_quantization_config.skip_modules is None: lowercase__ : str = get_keys_to_not_convert(__lowerCamelCase ) # add cpu modules to skip modules only for 4-bit modules if load_in_abit: bnb_quantization_config.skip_modules.extend(__lowerCamelCase ) lowercase__ : int = bnb_quantization_config.skip_modules # We add the modules we want to keep in full precision if bnb_quantization_config.keep_in_fpaa_modules is None: lowercase__ : Union[str, Any] = [] lowercase__ : Any = bnb_quantization_config.keep_in_fpaa_modules modules_to_not_convert.extend(__lowerCamelCase ) # compatibility with peft lowercase__ : int = load_in_abit lowercase__ : Union[str, Any] = load_in_abit lowercase__ : Union[str, Any] = get_parameter_device(__lowerCamelCase ) if model_device.type != "meta": # quantization of an already loaded model logger.warning( '''It is not recommended to quantize a loaded model. ''' '''The model should be instantiated under the `init_empty_weights` context manager.''' ) lowercase__ : Optional[int] = replace_with_bnb_layers(__lowerCamelCase , __lowerCamelCase , modules_to_not_convert=__lowerCamelCase ) # convert param to the right dtype lowercase__ : Any = bnb_quantization_config.torch_dtype for name, param in model.state_dict().items(): if any(module_to_keep_in_fpaa in name for module_to_keep_in_fpaa in keep_in_fpaa_modules ): param.to(torch.floataa ) if param.dtype != torch.floataa: lowercase__ : Optional[int] = name.replace('''.weight''' , '''''' ).replace('''.bias''' , '''''' ) lowercase__ : Dict = getattr(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) if param is not None: param.to(torch.floataa ) elif torch.is_floating_point(__lowerCamelCase ): param.to(__lowerCamelCase ) if model_device.type == "cuda": # move everything to cpu in the first place because we can't do quantization if the weights are already on cuda model.cuda(torch.cuda.current_device() ) torch.cuda.empty_cache() elif torch.cuda.is_available(): model.to(torch.cuda.current_device() ) else: raise RuntimeError('''No GPU found. A GPU is needed for quantization.''' ) logger.info( F"""The model device type is {model_device.type}. However, cuda is needed for quantization.""" '''We move the model to cuda.''' ) return model elif weights_location is None: raise RuntimeError( F"""`weights_location` needs to be the folder path containing the weights of the model, but we found {weights_location} """ ) else: with init_empty_weights(): lowercase__ : List[Any] = replace_with_bnb_layers( __lowerCamelCase , __lowerCamelCase , modules_to_not_convert=__lowerCamelCase ) lowercase__ : Any = get_quantized_model_device_map( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , max_memory=__lowerCamelCase , no_split_module_classes=__lowerCamelCase , ) if offload_state_dict is None and device_map is not None and "disk" in device_map.values(): lowercase__ : str = True lowercase__ : int = any(x in list(device_map.values() ) for x in ['''cpu''', '''disk'''] ) load_checkpoint_in_model( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , dtype=bnb_quantization_config.torch_dtype , offload_folder=__lowerCamelCase , offload_state_dict=__lowerCamelCase , keep_in_fpaa_modules=bnb_quantization_config.keep_in_fpaa_modules , offload_abit_bnb=load_in_abit and offload , ) return dispatch_model(__lowerCamelCase , device_map=__lowerCamelCase , offload_dir=__lowerCamelCase ) def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase=None ): if device_map is None: if torch.cuda.is_available(): lowercase__ : Dict = {"": torch.cuda.current_device()} else: raise RuntimeError('''No GPU found. A GPU is needed for quantization.''' ) logger.info('''The device_map was not initialized.''' '''Setting device_map to `{\'\':torch.cuda.current_device()}`.''' ) if isinstance(__lowerCamelCase , __lowerCamelCase ): if device_map not in ["auto", "balanced", "balanced_low_0", "sequential"]: raise ValueError( '''If passing a string for `device_map`, please choose \'auto\', \'balanced\', \'balanced_low_0\' or ''' '''\'sequential\'.''' ) lowercase__ : Union[str, Any] = {} special_dtypes.update( { name: bnb_quantization_config.torch_dtype for name, _ in model.named_parameters() if any(m in name for m in bnb_quantization_config.skip_modules ) } ) special_dtypes.update( { name: torch.floataa for name, _ in model.named_parameters() if any(m in name for m in bnb_quantization_config.keep_in_fpaa_modules ) } ) lowercase__ : Any = {} lowercase__ : Tuple = special_dtypes lowercase__ : Dict = no_split_module_classes lowercase__ : Union[str, Any] = bnb_quantization_config.target_dtype # get max_memory for each device. if device_map != "sequential": lowercase__ : Optional[int] = get_balanced_memory( __lowerCamelCase , low_zero=(device_map == '''balanced_low_0''') , max_memory=__lowerCamelCase , **__lowerCamelCase , ) lowercase__ : Optional[Any] = max_memory lowercase__ : Tuple = infer_auto_device_map(__lowerCamelCase , **__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ): # check if don't have any quantized module on the cpu lowercase__ : str = bnb_quantization_config.skip_modules + bnb_quantization_config.keep_in_fpaa_modules lowercase__ : int = { key: device_map[key] for key in device_map.keys() if key not in modules_not_to_convert } for device in ["cpu", "disk"]: if device in device_map_without_some_modules.values(): if bnb_quantization_config.load_in_abit: raise ValueError( '''\n Some modules are dispatched on the CPU or the disk. Make sure you have enough GPU RAM to fit\n the quantized model. If you want to dispatch the model on the CPU or the disk while keeping\n these modules in `torch_dtype`, you need to pass a custom `device_map` to\n `load_and_quantize_model`. Check\n https://huggingface.co/docs/accelerate/main/en/usage_guides/quantization#offload-modules-to-cpu-and-disk\n for more details.\n ''' ) else: logger.info( '''Some modules are are offloaded to the CPU or the disk. Note that these modules will be converted to 8-bit''' ) del device_map_without_some_modules return device_map def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=None , UpperCAmelCase=None ): if modules_to_not_convert is None: lowercase__ : Any = [] lowercase__ : Union[str, Any] = _replace_with_bnb_layers( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) if not has_been_replaced: logger.warning( '''You are loading your model in 8bit or 4bit but no linear modules were found in your model.''' ''' this can happen for some architectures such as gpt2 that uses Conv1D instead of Linear layers.''' ''' Please double check your model architecture, or submit an issue on github if you think this is''' ''' a bug.''' ) return model def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=None , UpperCAmelCase=None , ): lowercase__ : List[str] = False for name, module in model.named_children(): if current_key_name is None: lowercase__ : List[Any] = [] current_key_name.append(__lowerCamelCase ) if isinstance(__lowerCamelCase , nn.Linear ) and name not in modules_to_not_convert: # Check if the current key is not in the `modules_to_not_convert` lowercase__ : Optional[int] = ".".join(__lowerCamelCase ) lowercase__ : List[str] = True for key in modules_to_not_convert: if ( (key in current_key_name_str) and (key + "." in current_key_name_str) ) or key == current_key_name_str: lowercase__ : Union[str, Any] = False break if proceed: # Load bnb module with empty weight and replace ``nn.Linear` module if bnb_quantization_config.load_in_abit: lowercase__ : Dict = bnb.nn.LinearabitLt( module.in_features , module.out_features , module.bias is not None , has_fpaa_weights=__lowerCamelCase , threshold=bnb_quantization_config.llm_inta_threshold , ) elif bnb_quantization_config.load_in_abit: lowercase__ : Optional[int] = bnb.nn.Linearabit( module.in_features , module.out_features , module.bias is not None , bnb_quantization_config.bnb_abit_compute_dtype , compress_statistics=bnb_quantization_config.bnb_abit_use_double_quant , quant_type=bnb_quantization_config.bnb_abit_quant_type , ) else: raise ValueError('''load_in_8bit and load_in_4bit can\'t be both False''' ) lowercase__ : Dict = module.weight.data if module.bias is not None: lowercase__ : Any = module.bias.data bnb_module.requires_grad_(__lowerCamelCase ) setattr(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) lowercase__ : Optional[Any] = True if len(list(module.children() ) ) > 0: lowercase__ : Optional[Any] = _replace_with_bnb_layers( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) lowercase__ : Optional[Any] = has_been_replaced | _has_been_replaced # Remove the last key for recursion current_key_name.pop(-1 ) return model, has_been_replaced def __UpperCamelCase ( UpperCAmelCase ): # Create a copy of the model with init_empty_weights(): lowercase__ : List[str] = deepcopy(__lowerCamelCase ) # this has 0 cost since it is done inside `init_empty_weights` context manager` lowercase__ : Tuple = find_tied_parameters(__lowerCamelCase ) # For compatibility with Accelerate < 0.18 if isinstance(__lowerCamelCase , __lowerCamelCase ): lowercase__ : Tuple = sum(list(tied_params.values() ) , [] ) + list(tied_params.keys() ) else: lowercase__ : Optional[int] = sum(__lowerCamelCase , [] ) lowercase__ : Union[str, Any] = len(__lowerCamelCase ) > 0 # Check if it is a base model lowercase__ : int = False if hasattr(__lowerCamelCase , '''base_model_prefix''' ): lowercase__ : Dict = not hasattr(__lowerCamelCase , model.base_model_prefix ) # Ignore this for base models (BertModel, GPT2Model, etc.) if (not has_tied_params) and is_base_model: return [] # otherwise they have an attached head lowercase__ : Dict = list(model.named_children() ) lowercase__ : List[Any] = [list_modules[-1][0]] # add last module together with tied weights lowercase__ : Dict = set(__lowerCamelCase ) - set(__lowerCamelCase ) lowercase__ : Union[str, Any] = list(set(__lowerCamelCase ) ) + list(__lowerCamelCase ) # remove ".weight" from the keys lowercase__ : Optional[int] = [".weight", ".bias"] lowercase__ : Dict = [] for name in list_untouched: for name_to_remove in names_to_remove: if name_to_remove in name: lowercase__ : int = name.replace(__lowerCamelCase , '''''' ) filtered_module_names.append(__lowerCamelCase ) return filtered_module_names def __UpperCamelCase ( UpperCAmelCase ): for m in model.modules(): if isinstance(__lowerCamelCase , bnb.nn.Linearabit ): return True return False def __UpperCamelCase ( UpperCAmelCase ): return next(parameter.parameters() ).device def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): # if it is not quantized, we quantize and offload the quantized weights and the SCB stats if fpaa_statistics is None: set_module_tensor_to_device(__lowerCamelCase , __lowerCamelCase , 0 , dtype=__lowerCamelCase , value=__lowerCamelCase ) lowercase__ : Tuple = param_name lowercase__ : Union[str, Any] = model if "." in tensor_name: lowercase__ : Any = tensor_name.split('''.''' ) for split in splits[:-1]: lowercase__ : int = getattr(__lowerCamelCase , __lowerCamelCase ) if new_module is None: raise ValueError(F"""{module} has no attribute {split}.""" ) lowercase__ : str = new_module lowercase__ : List[str] = splits[-1] # offload weights lowercase__ : int = False offload_weight(module._parameters[tensor_name] , __lowerCamelCase , __lowerCamelCase , index=__lowerCamelCase ) if hasattr(module._parameters[tensor_name] , '''SCB''' ): offload_weight( module._parameters[tensor_name].SCB , param_name.replace('''weight''' , '''SCB''' ) , __lowerCamelCase , index=__lowerCamelCase , ) else: offload_weight(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , index=__lowerCamelCase ) offload_weight(__lowerCamelCase , param_name.replace('''weight''' , '''SCB''' ) , __lowerCamelCase , index=__lowerCamelCase ) set_module_tensor_to_device(__lowerCamelCase , __lowerCamelCase , '''meta''' , dtype=__lowerCamelCase , value=torch.empty(*param.size() ) )
152
import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor _snake_case : int = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[int] , *lowerCamelCase : List[Any] , **lowerCamelCase : int ) -> None: warnings.warn( "The class ImageGPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use ImageGPTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision import transforms from transformers import BitImageProcessor, FocalNetConfig, FocalNetForImageClassification from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def A_ ( lowercase_ ) -> Any: _snake_case : List[Any] = [2, 2, 6, 2] if "tiny" in model_name else [2, 2, 18, 2] _snake_case : Optional[int] = True if "large" in model_name or "huge" in model_name else False _snake_case : List[Any] = True if "large" in model_name or "huge" in model_name else False _snake_case : List[Any] = True if "large" in model_name or "huge" in model_name else False if "large" in model_name or "xlarge" in model_name or "huge" in model_name: if "fl3" in model_name: _snake_case : Optional[Any] = [3, 3, 3, 3] _snake_case : Optional[Any] = [5, 5, 5, 5] elif "fl4" in model_name: _snake_case : List[Any] = [4, 4, 4, 4] _snake_case : Any = [3, 3, 3, 3] if "tiny" in model_name or "small" in model_name or "base" in model_name: _snake_case : List[str] = [3, 3, 3, 3] if "lrf" in model_name: _snake_case : Any = [3, 3, 3, 3] else: _snake_case : str = [2, 2, 2, 2] if "tiny" in model_name: _snake_case : int = 96 elif "small" in model_name: _snake_case : Dict = 96 elif "base" in model_name: _snake_case : Any = 128 elif "large" in model_name: _snake_case : List[str] = 192 elif "xlarge" in model_name: _snake_case : List[str] = 256 elif "huge" in model_name: _snake_case : List[str] = 352 # set label information _snake_case : List[Any] = "huggingface/label-files" if "large" in model_name or "huge" in model_name: _snake_case : Any = "imagenet-22k-id2label.json" else: _snake_case : str = "imagenet-1k-id2label.json" _snake_case : Optional[Any] = json.load(open(hf_hub_download(__lowerCamelCase , __lowerCamelCase , repo_type='''dataset''' ) , '''r''' ) ) _snake_case : Any = {int(__lowerCamelCase ): v for k, v in idalabel.items()} _snake_case : Optional[Any] = {v: k for k, v in idalabel.items()} _snake_case : Dict = FocalNetConfig( embed_dim=__lowerCamelCase , depths=__lowerCamelCase , focal_levels=__lowerCamelCase , focal_windows=__lowerCamelCase , use_conv_embed=__lowerCamelCase , idalabel=__lowerCamelCase , labelaid=__lowerCamelCase , use_post_layernorm=__lowerCamelCase , use_layerscale=__lowerCamelCase , ) return config def A_ ( lowercase_ ) -> str: if "patch_embed.proj" in name: _snake_case : Union[str, Any] = name.replace('''patch_embed.proj''' , '''embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: _snake_case : Optional[Any] = name.replace('''patch_embed.norm''' , '''embeddings.norm''' ) if "layers" in name: _snake_case : Optional[int] = "encoder." + name if "encoder.layers" in name: _snake_case : List[str] = name.replace('''encoder.layers''' , '''encoder.stages''' ) if "downsample.proj" in name: _snake_case : Optional[Any] = name.replace('''downsample.proj''' , '''downsample.projection''' ) if "blocks" in name: _snake_case : Dict = name.replace('''blocks''' , '''layers''' ) if "modulation.f.weight" in name or "modulation.f.bias" in name: _snake_case : Tuple = name.replace('''modulation.f''' , '''modulation.projection_in''' ) if "modulation.h.weight" in name or "modulation.h.bias" in name: _snake_case : Union[str, Any] = name.replace('''modulation.h''' , '''modulation.projection_context''' ) if "modulation.proj.weight" in name or "modulation.proj.bias" in name: _snake_case : Any = name.replace('''modulation.proj''' , '''modulation.projection_out''' ) if name == "norm.weight": _snake_case : Any = "layernorm.weight" if name == "norm.bias": _snake_case : Optional[int] = "layernorm.bias" if "head" in name: _snake_case : Union[str, Any] = name.replace('''head''' , '''classifier''' ) else: _snake_case : Union[str, Any] = "focalnet." + name return name def A_ ( lowercase_ , lowercase_ , lowercase_=False ) -> List[Any]: # fmt: off _snake_case : str = { "focalnet-tiny": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_srf.pth", "focalnet-tiny-lrf": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_lrf.pth", "focalnet-small": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_srf.pth", "focalnet-small-lrf": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_lrf.pth", "focalnet-base": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_srf.pth", "focalnet-base-lrf": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_lrf.pth", "focalnet-large-lrf-fl3": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384.pth", "focalnet-large-lrf-fl4": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384_fl4.pth", "focalnet-xlarge-lrf-fl3": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384.pth", "focalnet-xlarge-lrf-fl4": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384_fl4.pth", } # fmt: on _snake_case : Union[str, Any] = model_name_to_url[model_name] print('''Checkpoint URL: ''' , __lowerCamelCase ) _snake_case : Tuple = torch.hub.load_state_dict_from_url(__lowerCamelCase , map_location='''cpu''' )["model"] # rename keys for key in state_dict.copy().keys(): _snake_case : Tuple = state_dict.pop(__lowerCamelCase ) _snake_case : List[str] = val _snake_case : Any = get_focalnet_config(__lowerCamelCase ) _snake_case : Optional[Any] = FocalNetForImageClassification(__lowerCamelCase ) model.eval() # load state dict model.load_state_dict(__lowerCamelCase ) # verify conversion _snake_case : Union[str, Any] = "http://images.cocodataset.org/val2017/000000039769.jpg" _snake_case : List[str] = BitImageProcessor( do_resize=__lowerCamelCase , size={'''shortest_edge''': 256} , resample=PILImageResampling.BILINEAR , do_center_crop=__lowerCamelCase , crop_size=224 , do_normalize=__lowerCamelCase , image_mean=__lowerCamelCase , image_std=__lowerCamelCase , ) _snake_case : Optional[int] = Image.open(requests.get(__lowerCamelCase , stream=__lowerCamelCase ).raw ) _snake_case : Tuple = processor(images=__lowerCamelCase , return_tensors='''pt''' ) _snake_case : Tuple = transforms.Compose( [ transforms.Resize(256 ), transforms.CenterCrop(224 ), transforms.ToTensor(), transforms.Normalize(mean=[0.4_85, 0.4_56, 0.4_06] , std=[0.2_29, 0.2_24, 0.2_25] ), ] ) _snake_case : Tuple = image_transforms(__lowerCamelCase ).unsqueeze(0 ) # verify pixel_values assert torch.allclose(inputs.pixel_values , __lowerCamelCase , atol=1E-4 ) _snake_case : Union[str, Any] = model(**__lowerCamelCase ) _snake_case : List[str] = outputs.logits.argmax(-1 ).item() print('''Predicted class:''' , model.config.idalabel[predicted_class_idx] ) print('''First values of logits:''' , outputs.logits[0, :3] ) if model_name == "focalnet-tiny": _snake_case : List[Any] = torch.tensor([0.21_66, -0.43_68, 0.21_91] ) elif model_name == "focalnet-tiny-lrf": _snake_case : Union[str, Any] = torch.tensor([1.16_69, 0.01_25, -0.16_95] ) elif model_name == "focalnet-small": _snake_case : List[Any] = torch.tensor([0.49_17, -0.04_30, 0.13_41] ) elif model_name == "focalnet-small-lrf": _snake_case : List[str] = torch.tensor([-0.25_88, -0.53_42, -0.23_31] ) elif model_name == "focalnet-base": _snake_case : List[str] = torch.tensor([-0.16_55, -0.40_90, -0.17_30] ) elif model_name == "focalnet-base-lrf": _snake_case : Tuple = torch.tensor([0.53_06, -0.04_83, -0.39_28] ) assert torch.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1E-4 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: print(f'''Saving model and processor of {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(__lowerCamelCase ) processor.save_pretrained(__lowerCamelCase ) if push_to_hub: print(f'''Pushing model and processor of {model_name} to the hub...''' ) model.push_to_hub(f'''{model_name}''' ) processor.push_to_hub(f'''{model_name}''' ) if __name__ == "__main__": lowerCAmelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="focalnet-tiny", type=str, help="Name of the FocalNet model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to push the model and processor to the hub.", ) lowerCAmelCase_ = parser.parse_args() convert_focalnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
326
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _snake_case : Optional[Any] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[str] = ["pixel_values"] def __init__( self : List[Any] , lowerCamelCase : bool = True , lowerCamelCase : Union[int, float] = 1 / 255 , lowerCamelCase : bool = True , lowerCamelCase : int = 8 , **lowerCamelCase : Tuple , ) -> None: super().__init__(**lowerCamelCase ) __snake_case : Dict = do_rescale __snake_case : Dict = rescale_factor __snake_case : Optional[Any] = do_pad __snake_case : Tuple = pad_size def __snake_case ( self : Dict , lowerCamelCase : np.ndarray , lowerCamelCase : float , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase : Optional[int] ) -> np.ndarray: return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def __snake_case ( self : Optional[Any] , lowerCamelCase : np.ndarray , lowerCamelCase : int , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None ) -> Tuple: __snake_case , __snake_case : List[str] = get_image_size(lowerCamelCase ) __snake_case : Optional[Any] = (old_height // size + 1) * size - old_height __snake_case : List[Any] = (old_width // size + 1) * size - old_width return pad(lowerCamelCase , ((0, pad_height), (0, pad_width)) , mode="symmetric" , data_format=lowerCamelCase ) def __snake_case ( self : Tuple , lowerCamelCase : ImageInput , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[float] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[int] = None , lowerCamelCase : Optional[Union[str, TensorType]] = None , lowerCamelCase : Union[str, ChannelDimension] = ChannelDimension.FIRST , **lowerCamelCase : Union[str, Any] , ) -> List[str]: __snake_case : int = do_rescale if do_rescale is not None else self.do_rescale __snake_case : Dict = rescale_factor if rescale_factor is not None else self.rescale_factor __snake_case : str = do_pad if do_pad is not None else self.do_pad __snake_case : Any = pad_size if pad_size is not None else self.pad_size __snake_case : int = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) # All transformations expect numpy arrays. __snake_case : str = [to_numpy_array(lowerCamelCase ) for image in images] if do_rescale: __snake_case : Optional[int] = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_pad: __snake_case : Optional[Any] = [self.pad(lowerCamelCase , size=lowerCamelCase ) for image in images] __snake_case : int = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __snake_case : Union[str, Any] = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
81
0
"""simple docstring""" def _lowerCAmelCase ( ): __SCREAMING_SNAKE_CASE = [] __SCREAMING_SNAKE_CASE = 1 while len(__lowerCamelCase ) < 1e6: constant.append(str(__lowerCamelCase ) ) i += 1 __SCREAMING_SNAKE_CASE = "".join(__lowerCamelCase ) return ( int(constant[0] ) * int(constant[9] ) * int(constant[99] ) * int(constant[999] ) * int(constant[9999] ) * int(constant[9_9999] ) * int(constant[99_9999] ) ) if __name__ == "__main__": print(solution())
155
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case : Any = { "configuration_informer": [ "INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "InformerConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "InformerForPrediction", "InformerModel", "InformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys _snake_case : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
0
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetrImageProcessor class lowerCAmelCase_ ( unittest.TestCase ): def __init__( self : Tuple , SCREAMING_SNAKE_CASE_ : Optional[Any] , SCREAMING_SNAKE_CASE_ : List[str]=7 , SCREAMING_SNAKE_CASE_ : Union[str, Any]=3 , SCREAMING_SNAKE_CASE_ : Dict=30 , SCREAMING_SNAKE_CASE_ : Any=400 , SCREAMING_SNAKE_CASE_ : List[str]=True , SCREAMING_SNAKE_CASE_ : Union[str, Any]=None , SCREAMING_SNAKE_CASE_ : Tuple=True , SCREAMING_SNAKE_CASE_ : int=1 / 255 , SCREAMING_SNAKE_CASE_ : Union[str, Any]=True , SCREAMING_SNAKE_CASE_ : Optional[int]=[0.5, 0.5, 0.5] , SCREAMING_SNAKE_CASE_ : int=[0.5, 0.5, 0.5] , SCREAMING_SNAKE_CASE_ : Union[str, Any]=True , ): # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p lowerCAmelCase__ = size if size is not None else {"shortest_edge": 18, "longest_edge": 1_333} lowerCAmelCase__ = parent lowerCAmelCase__ = batch_size lowerCAmelCase__ = num_channels lowerCAmelCase__ = min_resolution lowerCAmelCase__ = max_resolution lowerCAmelCase__ = do_resize lowerCAmelCase__ = size lowerCAmelCase__ = do_rescale lowerCAmelCase__ = rescale_factor lowerCAmelCase__ = do_normalize lowerCAmelCase__ = image_mean lowerCAmelCase__ = image_std lowerCAmelCase__ = do_pad def __snake_case ( self : Any ): return { "do_resize": self.do_resize, "size": self.size, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_pad": self.do_pad, } def __snake_case ( self : List[str] , SCREAMING_SNAKE_CASE_ : Optional[Any] , SCREAMING_SNAKE_CASE_ : List[str]=False ): if not batched: lowerCAmelCase__ = image_inputs[0] if isinstance(SCREAMING_SNAKE_CASE_ , Image.Image ): lowerCAmelCase__ = image.size else: lowerCAmelCase__ = image.shape[1], image.shape[2] if w < h: lowerCAmelCase__ = int(self.size['''shortest_edge'''] * h / w ) lowerCAmelCase__ = self.size["shortest_edge"] elif w > h: lowerCAmelCase__ = self.size["shortest_edge"] lowerCAmelCase__ = int(self.size['''shortest_edge'''] * w / h ) else: lowerCAmelCase__ = self.size["shortest_edge"] lowerCAmelCase__ = self.size["shortest_edge"] else: lowerCAmelCase__ = [] for image in image_inputs: lowerCAmelCase__ = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) lowerCAmelCase__ = max(SCREAMING_SNAKE_CASE_ , key=lambda SCREAMING_SNAKE_CASE_ : item[0] )[0] lowerCAmelCase__ = max(SCREAMING_SNAKE_CASE_ , key=lambda SCREAMING_SNAKE_CASE_ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class lowerCAmelCase_ ( _lowerCAmelCase , unittest.TestCase ): UpperCamelCase_ :Union[str, Any] = DetrImageProcessor if is_vision_available() else None def __snake_case ( self : Optional[int] ): lowerCAmelCase__ = DetrImageProcessingTester(self ) @property def __snake_case ( self : Union[str, Any] ): return self.image_processor_tester.prepare_image_processor_dict() def __snake_case ( self : Optional[Any] ): lowerCAmelCase__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , '''image_mean''' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , '''image_std''' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , '''do_normalize''' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , '''do_rescale''' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , '''rescale_factor''' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , '''do_resize''' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , '''size''' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , '''do_pad''' ) ) def __snake_case ( self : List[str] ): lowerCAmelCase__ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 18, '''longest_edge''': 1_333} ) self.assertEqual(image_processor.do_pad , SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=SCREAMING_SNAKE_CASE_ ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42, '''longest_edge''': 84} ) self.assertEqual(image_processor.do_pad , SCREAMING_SNAKE_CASE_ ) def __snake_case ( self : Any ): pass def __snake_case ( self : str ): # Initialize image_processing lowerCAmelCase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCAmelCase__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=SCREAMING_SNAKE_CASE_ ) for image in image_inputs: self.assertIsInstance(SCREAMING_SNAKE_CASE_ , Image.Image ) # Test not batched input lowerCAmelCase__ = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values lowerCAmelCase__ = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCAmelCase__ = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ = image_processing(SCREAMING_SNAKE_CASE_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : List[Any] ): # Initialize image_processing lowerCAmelCase__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCAmelCase__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=SCREAMING_SNAKE_CASE_ , numpify=SCREAMING_SNAKE_CASE_ ) for image in image_inputs: self.assertIsInstance(SCREAMING_SNAKE_CASE_ , np.ndarray ) # Test not batched input lowerCAmelCase__ = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values lowerCAmelCase__ = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCAmelCase__ = image_processing(SCREAMING_SNAKE_CASE_ , return_tensors='''pt''' ).pixel_values lowerCAmelCase__ = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : Tuple ): # Initialize image_processing lowerCAmelCase__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCAmelCase__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=SCREAMING_SNAKE_CASE_ , torchify=SCREAMING_SNAKE_CASE_ ) for image in image_inputs: self.assertIsInstance(SCREAMING_SNAKE_CASE_ , torch.Tensor ) # Test not batched input lowerCAmelCase__ = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values lowerCAmelCase__ = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCAmelCase__ = image_processing(SCREAMING_SNAKE_CASE_ , return_tensors='''pt''' ).pixel_values lowerCAmelCase__ = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def __snake_case ( self : Dict ): # prepare image and target lowerCAmelCase__ = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_annotations.txt''' , '''r''' ) as f: lowerCAmelCase__ = json.loads(f.read() ) lowerCAmelCase__ = {"image_id": 39_769, "annotations": target} # encode them lowerCAmelCase__ = DetrImageProcessor.from_pretrained('''facebook/detr-resnet-50''' ) lowerCAmelCase__ = image_processing(images=SCREAMING_SNAKE_CASE_ , annotations=SCREAMING_SNAKE_CASE_ , return_tensors='''pt''' ) # verify pixel values lowerCAmelCase__ = torch.Size([1, 3, 800, 1_066] ) self.assertEqual(encoding['''pixel_values'''].shape , SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ = torch.tensor([0.2_796, 0.3_138, 0.3_481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , SCREAMING_SNAKE_CASE_ , atol=1e-4 ) ) # verify area lowerCAmelCase__ = torch.tensor([5_887.9_600, 11_250.2_061, 489_353.8_438, 837_122.7_500, 147_967.5_156, 165_732.3_438] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , SCREAMING_SNAKE_CASE_ ) ) # verify boxes lowerCAmelCase__ = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ = torch.tensor([0.5_503, 0.2_765, 0.0_604, 0.2_215] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , SCREAMING_SNAKE_CASE_ , atol=1e-3 ) ) # verify image_id lowerCAmelCase__ = torch.tensor([39_769] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , SCREAMING_SNAKE_CASE_ ) ) # verify is_crowd lowerCAmelCase__ = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , SCREAMING_SNAKE_CASE_ ) ) # verify class_labels lowerCAmelCase__ = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , SCREAMING_SNAKE_CASE_ ) ) # verify orig_size lowerCAmelCase__ = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , SCREAMING_SNAKE_CASE_ ) ) # verify size lowerCAmelCase__ = torch.tensor([800, 1_066] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , SCREAMING_SNAKE_CASE_ ) ) @slow def __snake_case ( self : Tuple ): # prepare image, target and masks_path lowerCAmelCase__ = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt''' , '''r''' ) as f: lowerCAmelCase__ = json.loads(f.read() ) lowerCAmelCase__ = {"file_name": "000000039769.png", "image_id": 39_769, "segments_info": target} lowerCAmelCase__ = pathlib.Path('''./tests/fixtures/tests_samples/COCO/coco_panoptic''' ) # encode them lowerCAmelCase__ = DetrImageProcessor.from_pretrained('''facebook/detr-resnet-50-panoptic''' ) lowerCAmelCase__ = image_processing(images=SCREAMING_SNAKE_CASE_ , annotations=SCREAMING_SNAKE_CASE_ , masks_path=SCREAMING_SNAKE_CASE_ , return_tensors='''pt''' ) # verify pixel values lowerCAmelCase__ = torch.Size([1, 3, 800, 1_066] ) self.assertEqual(encoding['''pixel_values'''].shape , SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ = torch.tensor([0.2_796, 0.3_138, 0.3_481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , SCREAMING_SNAKE_CASE_ , atol=1e-4 ) ) # verify area lowerCAmelCase__ = torch.tensor([147_979.6_875, 165_527.0_469, 484_638.5_938, 11_292.9_375, 5_879.6_562, 7_634.1_147] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , SCREAMING_SNAKE_CASE_ ) ) # verify boxes lowerCAmelCase__ = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ = torch.tensor([0.2_625, 0.5_437, 0.4_688, 0.8_625] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , SCREAMING_SNAKE_CASE_ , atol=1e-3 ) ) # verify image_id lowerCAmelCase__ = torch.tensor([39_769] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , SCREAMING_SNAKE_CASE_ ) ) # verify is_crowd lowerCAmelCase__ = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , SCREAMING_SNAKE_CASE_ ) ) # verify class_labels lowerCAmelCase__ = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , SCREAMING_SNAKE_CASE_ ) ) # verify masks lowerCAmelCase__ = 822_873 self.assertEqual(encoding['''labels'''][0]['''masks'''].sum().item() , SCREAMING_SNAKE_CASE_ ) # verify orig_size lowerCAmelCase__ = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , SCREAMING_SNAKE_CASE_ ) ) # verify size lowerCAmelCase__ = torch.tensor([800, 1_066] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , SCREAMING_SNAKE_CASE_ ) )
668
from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging _snake_case : Any = logging.get_logger(__name__) # pylint: disable=invalid-name class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[Any] , lowerCamelCase : CLIPSegForImageSegmentation , lowerCamelCase : CLIPSegProcessor , lowerCamelCase : AutoencoderKL , lowerCamelCase : CLIPTextModel , lowerCamelCase : CLIPTokenizer , lowerCamelCase : UNetaDConditionModel , lowerCamelCase : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , lowerCamelCase : StableDiffusionSafetyChecker , lowerCamelCase : CLIPImageProcessor , ) -> Tuple: super().__init__() if hasattr(scheduler.config , "steps_offset" ) and scheduler.config.steps_offset != 1: __snake_case : Tuple = ( F'The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`' F' should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure ' "to update the config accordingly as leaving `steps_offset` might led to incorrect results" " in future versions. If you have downloaded this checkpoint from the Hugging Face Hub," " it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`" " file" ) deprecate("steps_offset!=1" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : Any = dict(scheduler.config ) __snake_case : List[Any] = 1 __snake_case : Tuple = FrozenDict(lowerCamelCase ) if hasattr(scheduler.config , "skip_prk_steps" ) and scheduler.config.skip_prk_steps is False: __snake_case : List[str] = ( F'The configuration file of this scheduler: {scheduler} has not set the configuration' " `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make" " sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to" " incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face" " Hub, it would be very nice if you could open a Pull request for the" " `scheduler/scheduler_config.json` file" ) deprecate("skip_prk_steps not set" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : List[str] = dict(scheduler.config ) __snake_case : List[str] = True __snake_case : Any = FrozenDict(lowerCamelCase ) if safety_checker is None: logger.warning( F'You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure' " that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered" " results in services or applications open to the public. Both the diffusers team and Hugging Face" " strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling" " it only for use-cases that involve analyzing network behavior or auditing its results. For more" " information, please have a look at https://github.com/huggingface/diffusers/pull/254 ." ) self.register_modules( segmentation_model=lowerCamelCase , segmentation_processor=lowerCamelCase , vae=lowerCamelCase , text_encoder=lowerCamelCase , tokenizer=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase , safety_checker=lowerCamelCase , feature_extractor=lowerCamelCase , ) def __snake_case ( self : Dict , lowerCamelCase : Optional[Union[str, int]] = "auto" ) -> Tuple: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __snake_case : Any = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowerCamelCase ) def __snake_case ( self : List[Any] ) -> Any: self.enable_attention_slicing(lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> str: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) __snake_case : Optional[int] = torch.device("cuda" ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(lowerCamelCase , lowerCamelCase ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __snake_case ( self : int ) -> Any: if self.device != torch.device("meta" ) or not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowerCamelCase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self : List[Any] , lowerCamelCase : Union[str, List[str]] , lowerCamelCase : Union[torch.FloatTensor, PIL.Image.Image] , lowerCamelCase : str , lowerCamelCase : int = 512 , lowerCamelCase : int = 512 , lowerCamelCase : int = 50 , lowerCamelCase : float = 7.5 , lowerCamelCase : Optional[Union[str, List[str]]] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : float = 0.0 , lowerCamelCase : Optional[torch.Generator] = None , lowerCamelCase : Optional[torch.FloatTensor] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , lowerCamelCase : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , lowerCamelCase : int = 1 , **lowerCamelCase : Dict , ) -> List[str]: __snake_case : Tuple = self.segmentation_processor( text=[text] , images=[image] , padding="max_length" , return_tensors="pt" ).to(self.device ) __snake_case : str = self.segmentation_model(**lowerCamelCase ) __snake_case : Dict = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() __snake_case : List[Any] = self.numpy_to_pil(lowerCamelCase )[0].resize(image.size ) # Run inpainting pipeline with the generated mask __snake_case : Tuple = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=lowerCamelCase , image=lowerCamelCase , mask_image=lowerCamelCase , height=lowerCamelCase , width=lowerCamelCase , num_inference_steps=lowerCamelCase , guidance_scale=lowerCamelCase , negative_prompt=lowerCamelCase , num_images_per_prompt=lowerCamelCase , eta=lowerCamelCase , generator=lowerCamelCase , latents=lowerCamelCase , output_type=lowerCamelCase , return_dict=lowerCamelCase , callback=lowerCamelCase , callback_steps=lowerCamelCase , )
81
0
'''simple docstring''' import argparse import json from pathlib import Path import torch import torchaudio from datasets import load_dataset from huggingface_hub import hf_hub_download from transformers import ASTConfig, ASTFeatureExtractor, ASTForAudioClassification from transformers.utils import logging logging.set_verbosity_info() __UpperCamelCase = logging.get_logger(__name__) def _a ( _lowerCamelCase ) -> str: """simple docstring""" __snake_case : Optional[Any] = ASTConfig() if "10-10" in model_name: pass elif "speech-commands" in model_name: __snake_case : Union[str, Any] = 128 elif "12-12" in model_name: __snake_case : str = 12 __snake_case : Optional[int] = 12 elif "14-14" in model_name: __snake_case : List[Any] = 14 __snake_case : Dict = 14 elif "16-16" in model_name: __snake_case : Optional[int] = 16 __snake_case : Optional[int] = 16 else: raise ValueError("""Model not supported""" ) __snake_case : Optional[int] = "huggingface/label-files" if "speech-commands" in model_name: __snake_case : Tuple = 35 __snake_case : Optional[int] = "speech-commands-v2-id2label.json" else: __snake_case : List[Any] = 527 __snake_case : Tuple = "audioset-id2label.json" __snake_case : Optional[int] = json.load(open(hf_hub_download(__lowerCamelCase , __lowerCamelCase , repo_type="""dataset""" ) , """r""" ) ) __snake_case : Optional[int] = {int(__lowerCamelCase ): v for k, v in idalabel.items()} __snake_case : str = idalabel __snake_case : Tuple = {v: k for k, v in idalabel.items()} return config def _a ( _lowerCamelCase ) -> Optional[int]: """simple docstring""" if "module.v" in name: __snake_case : List[Any] = name.replace("""module.v""" , """audio_spectrogram_transformer""" ) if "cls_token" in name: __snake_case : Any = name.replace("""cls_token""" , """embeddings.cls_token""" ) if "dist_token" in name: __snake_case : List[Any] = name.replace("""dist_token""" , """embeddings.distillation_token""" ) if "pos_embed" in name: __snake_case : Tuple = name.replace("""pos_embed""" , """embeddings.position_embeddings""" ) if "patch_embed.proj" in name: __snake_case : Dict = name.replace("""patch_embed.proj""" , """embeddings.patch_embeddings.projection""" ) # transformer blocks if "blocks" in name: __snake_case : Union[str, Any] = name.replace("""blocks""" , """encoder.layer""" ) if "attn.proj" in name: __snake_case : List[Any] = name.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in name: __snake_case : Optional[int] = name.replace("""attn""" , """attention.self""" ) if "norm1" in name: __snake_case : Any = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: __snake_case : int = name.replace("""norm2""" , """layernorm_after""" ) if "mlp.fc1" in name: __snake_case : List[str] = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: __snake_case : Union[str, Any] = name.replace("""mlp.fc2""" , """output.dense""" ) # final layernorm if "audio_spectrogram_transformer.norm" in name: __snake_case : Union[str, Any] = name.replace("""audio_spectrogram_transformer.norm""" , """audio_spectrogram_transformer.layernorm""" ) # classifier head if "module.mlp_head.0" in name: __snake_case : List[str] = name.replace("""module.mlp_head.0""" , """classifier.layernorm""" ) if "module.mlp_head.1" in name: __snake_case : str = name.replace("""module.mlp_head.1""" , """classifier.dense""" ) return name def _a ( _lowerCamelCase , _lowerCamelCase ) -> List[str]: """simple docstring""" for key in orig_state_dict.copy().keys(): __snake_case : List[str] = orig_state_dict.pop(__lowerCamelCase ) if "qkv" in key: __snake_case : Optional[int] = key.split(""".""" ) __snake_case : int = int(key_split[3] ) __snake_case : str = config.hidden_size if "weight" in key: __snake_case : int = val[:dim, :] __snake_case : Optional[Any] = val[dim : dim * 2, :] __snake_case : int = val[-dim:, :] else: __snake_case : Union[str, Any] = val[:dim] __snake_case : int = val[dim : dim * 2] __snake_case : Any = val[-dim:] else: __snake_case : Optional[int] = val return orig_state_dict def _a ( _lowerCamelCase ) -> Any: """simple docstring""" __snake_case : List[Any] = [ "module.v.head.weight", "module.v.head.bias", "module.v.head_dist.weight", "module.v.head_dist.bias", ] for k in ignore_keys: state_dict.pop(__lowerCamelCase , __lowerCamelCase ) @torch.no_grad() def _a ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase=False ) -> Union[str, Any]: """simple docstring""" __snake_case : Optional[Any] = get_audio_spectrogram_transformer_config(__lowerCamelCase ) __snake_case : List[str] = { "ast-finetuned-audioset-10-10-0.4593": ( "https://www.dropbox.com/s/ca0b1v2nlxzyeb4/audioset_10_10_0.4593.pth?dl=1" ), "ast-finetuned-audioset-10-10-0.450": ( "https://www.dropbox.com/s/1tv0hovue1bxupk/audioset_10_10_0.4495.pth?dl=1" ), "ast-finetuned-audioset-10-10-0.448": ( "https://www.dropbox.com/s/6u5sikl4b9wo4u5/audioset_10_10_0.4483.pth?dl=1" ), "ast-finetuned-audioset-10-10-0.448-v2": ( "https://www.dropbox.com/s/kt6i0v9fvfm1mbq/audioset_10_10_0.4475.pth?dl=1" ), "ast-finetuned-audioset-12-12-0.447": ( "https://www.dropbox.com/s/snfhx3tizr4nuc8/audioset_12_12_0.4467.pth?dl=1" ), "ast-finetuned-audioset-14-14-0.443": ( "https://www.dropbox.com/s/z18s6pemtnxm4k7/audioset_14_14_0.4431.pth?dl=1" ), "ast-finetuned-audioset-16-16-0.442": ( "https://www.dropbox.com/s/mdsa4t1xmcimia6/audioset_16_16_0.4422.pth?dl=1" ), "ast-finetuned-speech-commands-v2": ( "https://www.dropbox.com/s/q0tbqpwv44pquwy/speechcommands_10_10_0.9812.pth?dl=1" ), } # load original state_dict __snake_case : List[Any] = model_name_to_url[model_name] __snake_case : List[str] = torch.hub.load_state_dict_from_url(__lowerCamelCase , map_location="""cpu""" ) # remove some keys remove_keys(__lowerCamelCase ) # rename some keys __snake_case : Tuple = convert_state_dict(__lowerCamelCase , __lowerCamelCase ) # load 🤗 model __snake_case : Union[str, Any] = ASTForAudioClassification(__lowerCamelCase ) model.eval() model.load_state_dict(__lowerCamelCase ) # verify outputs on dummy input # source: https://github.com/YuanGongND/ast/blob/79e873b8a54d0a3b330dd522584ff2b9926cd581/src/run.py#L62 __snake_case : Union[str, Any] = -4.2_67_73_93 if "speech-commands" not in model_name else -6.84_59_78 __snake_case : Tuple = 4.5_68_99_74 if "speech-commands" not in model_name else 5.5_65_45_26 __snake_case : str = 1024 if "speech-commands" not in model_name else 128 __snake_case : Tuple = ASTFeatureExtractor(mean=__lowerCamelCase , std=__lowerCamelCase , max_length=__lowerCamelCase ) if "speech-commands" in model_name: __snake_case : Dict = load_dataset("""speech_commands""" , """v0.02""" , split="""validation""" ) __snake_case : Optional[int] = dataset[0]["audio"]["array"] else: __snake_case : int = hf_hub_download( repo_id="""nielsr/audio-spectogram-transformer-checkpoint""" , filename="""sample_audio.flac""" , repo_type="""dataset""" , ) __snake_case : Optional[Any] = torchaudio.load(__lowerCamelCase ) __snake_case : Optional[Any] = waveform.squeeze().numpy() __snake_case : List[str] = feature_extractor(__lowerCamelCase , sampling_rate=1_6000 , return_tensors="""pt""" ) # forward pass __snake_case : List[Any] = model(**__lowerCamelCase ) __snake_case : Tuple = outputs.logits if model_name == "ast-finetuned-audioset-10-10-0.4593": __snake_case : List[str] = torch.tensor([-0.87_60, -7.00_42, -8.66_02] ) elif model_name == "ast-finetuned-audioset-10-10-0.450": __snake_case : List[str] = torch.tensor([-1.19_86, -7.09_03, -8.27_18] ) elif model_name == "ast-finetuned-audioset-10-10-0.448": __snake_case : Optional[int] = torch.tensor([-2.61_28, -8.00_80, -9.43_44] ) elif model_name == "ast-finetuned-audioset-10-10-0.448-v2": __snake_case : Optional[Any] = torch.tensor([-1.50_80, -7.45_34, -8.89_17] ) elif model_name == "ast-finetuned-audioset-12-12-0.447": __snake_case : Optional[Any] = torch.tensor([-0.50_50, -6.58_33, -8.08_43] ) elif model_name == "ast-finetuned-audioset-14-14-0.443": __snake_case : Optional[int] = torch.tensor([-0.38_26, -7.03_36, -8.24_13] ) elif model_name == "ast-finetuned-audioset-16-16-0.442": __snake_case : int = torch.tensor([-1.21_13, -6.91_01, -8.34_70] ) elif model_name == "ast-finetuned-speech-commands-v2": __snake_case : List[str] = torch.tensor([6.15_89, -8.05_66, -8.79_84] ) else: raise ValueError("""Unknown model name""" ) if not torch.allclose(logits[0, :3] , __lowerCamelCase , atol=1E-4 ): raise ValueError("""Logits don't match""" ) print("""Looks ok!""" ) if pytorch_dump_folder_path is not None: Path(__lowerCamelCase ).mkdir(exist_ok=__lowerCamelCase ) print(F'''Saving model {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(__lowerCamelCase ) print(F'''Saving feature extractor to {pytorch_dump_folder_path}''' ) feature_extractor.save_pretrained(__lowerCamelCase ) if push_to_hub: print("""Pushing model and feature extractor to the hub...""" ) model.push_to_hub(F'''MIT/{model_name}''' ) feature_extractor.push_to_hub(F'''MIT/{model_name}''' ) if __name__ == "__main__": __UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="ast-finetuned-audioset-10-10-0.4593", type=str, help="Name of the Audio Spectrogram Transformer model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) __UpperCamelCase = parser.parse_args() convert_audio_spectrogram_transformer_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
26
class a : """simple docstring""" def __init__( self : Tuple , lowerCamelCase : list ) -> None: __snake_case : str = set_counts __snake_case : Union[str, Any] = max(lowerCamelCase ) __snake_case : List[Any] = len(lowerCamelCase ) __snake_case : Tuple = [1] * num_sets __snake_case : Dict = list(range(lowerCamelCase ) ) def __snake_case ( self : str , lowerCamelCase : int , lowerCamelCase : int ) -> bool: __snake_case : List[Any] = self.get_parent(lowerCamelCase ) __snake_case : Tuple = self.get_parent(lowerCamelCase ) if src_parent == dst_parent: return False if self.ranks[dst_parent] >= self.ranks[src_parent]: self.set_counts[dst_parent] += self.set_counts[src_parent] __snake_case : List[str] = 0 __snake_case : List[Any] = dst_parent if self.ranks[dst_parent] == self.ranks[src_parent]: self.ranks[dst_parent] += 1 __snake_case : Dict = self.set_counts[dst_parent] else: self.set_counts[src_parent] += self.set_counts[dst_parent] __snake_case : Union[str, Any] = 0 __snake_case : Optional[int] = src_parent __snake_case : Tuple = self.set_counts[src_parent] __snake_case : str = max(self.max_set , lowerCamelCase ) return True def __snake_case ( self : int , lowerCamelCase : int ) -> int: if self.parents[disj_set] == disj_set: return disj_set __snake_case : Optional[int] = self.get_parent(self.parents[disj_set] ) return self.parents[disj_set]
81
0
'''simple docstring''' from typing import List, Optional, Union import numpy as np import tensorflow as tf from .utils import logging UpperCAmelCase = logging.get_logger(__name__) def __UpperCamelCase ( lowercase__ : Dict ): '''simple docstring''' if isinstance(__lowerCamelCase, np.ndarray ): return list(tensor.shape ) __lowercase =tf.shape(__lowerCamelCase ) if tensor.shape == tf.TensorShape(__lowerCamelCase ): return dynamic __lowercase =tensor.shape.as_list() return [dynamic[i] if s is None else s for i, s in enumerate(__lowerCamelCase )] def __UpperCamelCase ( lowercase__ : Any, lowercase__ : str = None, lowercase__ : List[Any] = None ): '''simple docstring''' return tf.nn.softmax(logits=logits + 1E-9, axis=__lowerCamelCase, name=__lowerCamelCase ) def __UpperCamelCase ( lowercase__ : Dict, lowercase__ : Optional[int], lowercase__ : List[str], lowercase__ : List[str]=1E-5, lowercase__ : Optional[Any]=-1 ): '''simple docstring''' if weight.shape.rank != 1 or bias.shape.rank != 1 or not isinstance(__lowerCamelCase, __lowerCamelCase ): raise NotImplementedError('Only 1D weight and bias tensors are supported for now, with only a single axis.' ) # Get mean and variance on the axis to be normalized __lowercase =tf.nn.moments(__lowerCamelCase, axes=[axis], keepdims=__lowerCamelCase ) if axis != -1: # Reshape scale and weight to have the same rank as inputs, but with 1 dimensions # on every dimension except axis __lowercase =[1] * inputs.shape.rank __lowercase =shape_list(__lowerCamelCase )[axis] __lowercase =tf.reshape(__lowerCamelCase, __lowerCamelCase ) __lowercase =tf.reshape(__lowerCamelCase, __lowerCamelCase ) # Compute layer normalization using the batch_normalization # function. __lowercase =tf.nn.batch_normalization( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, offset=__lowerCamelCase, scale=__lowerCamelCase, variance_epsilon=__lowerCamelCase, ) return outputs def __UpperCamelCase ( lowercase__ : Dict, lowercase__ : Union[str, Any]=0, lowercase__ : int=-1 ): '''simple docstring''' if end_dim < 0: end_dim += input.shape.rank if start_dim < 0: start_dim += input.shape.rank if start_dim == end_dim: return input __lowercase =tf.shape(__lowerCamelCase ) __lowercase =tf.math.reduce_prod(in_shape[start_dim : end_dim + 1] ) __lowercase =tf.concat([in_shape[:start_dim], [flattened_dim], in_shape[end_dim + 1 :]], axis=0 ) return tf.reshape(__lowerCamelCase, __lowerCamelCase ) def __UpperCamelCase ( lowercase__ : Tuple ): '''simple docstring''' if not isinstance(__lowerCamelCase, tf.Tensor ): __lowercase =tf.convert_to_tensor(__lowerCamelCase ) # Catches stray NumPy inputs if encoder_attention_mask.shape.rank == 3: __lowercase =encoder_attention_mask[:, None, :, :] if encoder_attention_mask.shape.rank == 2: __lowercase =encoder_attention_mask[:, None, None, :] # T5 has a mask that can compare sequence ids, we can simulate this here with this transposition # Cf. https://github.com/tensorflow/mesh/blob/8d2465e9bc93129b913b5ccc6a59aa97abd96ec6/mesh_tensorflow # /transformer/transformer_layers.py#L270 # encoder_extended_attention_mask = (encoder_extended_attention_mask == # encoder_extended_attention_mask.transpose(-1, -2)) __lowercase =( tf.cast(1, encoder_attention_mask.dtype ) - encoder_extended_attention_mask ) * encoder_extended_attention_mask.dtype.min return encoder_extended_attention_mask def __UpperCamelCase ( lowercase__ : Optional[int], lowercase__ : Tuple, lowercase__ : Tuple = "input_ids" ): '''simple docstring''' tf.debugging.assert_less( __lowerCamelCase, tf.cast(__lowerCamelCase, dtype=tensor.dtype ), message=( F'''The maximum value of {tensor_name} ({tf.math.reduce_max(__lowerCamelCase )}) must be smaller than the embedding ''' F'''layer\'s input dimension ({embed_dim}). The likely cause is some problem at tokenization time.''' ), ) def __UpperCamelCase ( lowercase__ : List[str], lowercase__ : List[str], lowercase__ : Union[str, Any] ): '''simple docstring''' __lowercase =6_45_12 # Check that no item in `data` is larger than `HDF5_OBJECT_HEADER_LIMIT` # because in that case even chunking the array would not make the saving # possible. __lowercase =[x for x in data if len(__lowerCamelCase ) > HDF5_OBJECT_HEADER_LIMIT] # Expecting this to never be true. if bad_attributes: raise RuntimeError( 'The following attributes cannot be saved to HDF5 file because ' F'''they are larger than {HDF5_OBJECT_HEADER_LIMIT} ''' F'''bytes: {bad_attributes}''' ) __lowercase =np.asarray(__lowerCamelCase ) __lowercase =1 __lowercase =np.array_split(__lowerCamelCase, __lowerCamelCase ) # This will never loop forever thanks to the test above. while any(x.nbytes > HDF5_OBJECT_HEADER_LIMIT for x in chunked_data ): num_chunks += 1 __lowercase =np.array_split(__lowerCamelCase, __lowerCamelCase ) if num_chunks > 1: for chunk_id, chunk_data in enumerate(__lowerCamelCase ): __lowercase =chunk_data else: __lowercase =data def __UpperCamelCase ( lowercase__ : int, lowercase__ : Dict ): '''simple docstring''' if name in group.attrs: __lowercase =[n.decode('utf8' ) if hasattr(__lowerCamelCase, 'decode' ) else n for n in group.attrs[name]] else: __lowercase =[] __lowercase =0 while "%s%d" % (name, chunk_id) in group.attrs: data.extend( [n.decode('utf8' ) if hasattr(__lowerCamelCase, 'decode' ) else n for n in group.attrs['%s%d' % (name, chunk_id)]] ) chunk_id += 1 return data def __UpperCamelCase ( lowercase__ : Optional[Any] ): '''simple docstring''' def _expand_single_ad_tensor(lowercase__ : Any ): if isinstance(__lowerCamelCase, tf.Tensor ) and t.shape.rank == 1: return tf.expand_dims(__lowerCamelCase, axis=-1 ) return t return tf.nest.map_structure(_expand_single_ad_tensor, __lowerCamelCase )
119
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class a (unittest.TestCase ): """simple docstring""" def __init__( self : int , lowerCamelCase : Any , lowerCamelCase : int=7 , lowerCamelCase : str=3 , lowerCamelCase : Optional[int]=30 , lowerCamelCase : Dict=400 , lowerCamelCase : str=True , lowerCamelCase : str=None , lowerCamelCase : Any=True , lowerCamelCase : Union[str, Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[str]=True , lowerCamelCase : Optional[int]=1 / 255 , lowerCamelCase : Any=True , ) -> str: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __snake_case : Optional[Any] = size if size is not None else {"shortest_edge": 18, "longest_edge": 1333} __snake_case : Optional[Any] = parent __snake_case : List[Any] = batch_size __snake_case : Optional[int] = num_channels __snake_case : str = min_resolution __snake_case : int = max_resolution __snake_case : int = do_resize __snake_case : Tuple = size __snake_case : Any = do_normalize __snake_case : int = image_mean __snake_case : Tuple = image_std __snake_case : Dict = do_rescale __snake_case : Optional[Any] = rescale_factor __snake_case : str = do_pad def __snake_case ( self : Any ) -> int: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def __snake_case ( self : Dict , lowerCamelCase : List[Any] , lowerCamelCase : Optional[Any]=False ) -> List[str]: if not batched: __snake_case : Dict = image_inputs[0] if isinstance(lowerCamelCase , Image.Image ): __snake_case , __snake_case : Dict = image.size else: __snake_case , __snake_case : List[str] = image.shape[1], image.shape[2] if w < h: __snake_case : Optional[int] = int(self.size["shortest_edge"] * h / w ) __snake_case : int = self.size["shortest_edge"] elif w > h: __snake_case : List[str] = self.size["shortest_edge"] __snake_case : Optional[Any] = int(self.size["shortest_edge"] * w / h ) else: __snake_case : List[Any] = self.size["shortest_edge"] __snake_case : Any = self.size["shortest_edge"] else: __snake_case : int = [] for image in image_inputs: __snake_case , __snake_case : List[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[0] )[0] __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class a (_lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ConditionalDetrImageProcessor if is_vision_available() else None def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : Optional[Any] = ConditionalDetrImageProcessingTester(self ) @property def __snake_case ( self : Any ) -> str: return self.image_processor_tester.prepare_image_processor_dict() def __snake_case ( self : Optional[Any] ) -> Optional[int]: __snake_case : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase , "image_mean" ) ) self.assertTrue(hasattr(lowerCamelCase , "image_std" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_normalize" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase , "size" ) ) def __snake_case ( self : Any ) -> Dict: __snake_case : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 1333} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) __snake_case : str = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=lowerCamelCase ) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> Dict: pass def __snake_case ( self : Tuple ) -> str: # Initialize image_processing __snake_case : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , Image.Image ) # Test not batched input __snake_case : int = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) __snake_case : Dict = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> str: # Initialize image_processing __snake_case : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __snake_case : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , numpify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , np.ndarray ) # Test not batched input __snake_case : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : List[Any] = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> List[str]: # Initialize image_processing __snake_case : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , torchify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , torch.Tensor ) # Test not batched input __snake_case : Optional[Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : List[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : int = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def __snake_case ( self : Any ) -> Optional[int]: # prepare image and target __snake_case : List[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : List[Any] = {"image_id": 39769, "annotations": target} # encode them __snake_case : List[str] = ConditionalDetrImageProcessor.from_pretrained("microsoft/conditional-detr-resnet-50" ) __snake_case : List[str] = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : Tuple = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Optional[int] = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : List[Any] = torch.tensor([58_87.96_00, 1_12_50.20_61, 48_93_53.84_38, 83_71_22.75_00, 14_79_67.51_56, 16_57_32.34_38] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : Optional[int] = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([0.55_03, 0.27_65, 0.06_04, 0.22_15] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Optional[Any] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : List[Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : Dict = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify orig_size __snake_case : int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Tuple = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) ) @slow def __snake_case ( self : str ) -> Tuple: # prepare image, target and masks_path __snake_case : Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : str = {"file_name": "000000039769.png", "image_id": 39769, "segments_info": target} __snake_case : Optional[int] = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them __snake_case : int = ConditionalDetrImageProcessor(format="coco_panoptic" ) __snake_case : str = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , masks_path=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : List[str] = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Dict = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : Any = torch.tensor([14_79_79.68_75, 16_55_27.04_69, 48_46_38.59_38, 1_12_92.93_75, 58_79.65_62, 76_34.11_47] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : str = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : Optional[Any] = torch.tensor([0.26_25, 0.54_37, 0.46_88, 0.86_25] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Tuple = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : Dict = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : int = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify masks __snake_case : List[Any] = 822873 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , lowerCamelCase ) # verify orig_size __snake_case : List[str] = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Any = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) )
81
0
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device SCREAMING_SNAKE_CASE__ : Tuple = False class lowerCAmelCase__ ( unittest.TestCase ): pass @nightly @require_torch_gpu class lowerCAmelCase__ ( unittest.TestCase ): def __A ( self : List[str] ) -> str: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __A ( self : Any ) -> Dict: __lowerCamelCase = VersatileDiffusionTextToImagePipeline.from_pretrained('''shi-labs/versatile-diffusion''' ) # remove text_unet pipe.remove_unused_weights() pipe.to(SCREAMING_SNAKE_CASE__ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = "A painting of a squirrel eating a burger " __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe( prompt=SCREAMING_SNAKE_CASE__ , generator=SCREAMING_SNAKE_CASE__ , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = VersatileDiffusionTextToImagePipeline.from_pretrained(SCREAMING_SNAKE_CASE__ ) pipe.to(SCREAMING_SNAKE_CASE__ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = generator.manual_seed(0 ) __lowerCamelCase = pipe( prompt=SCREAMING_SNAKE_CASE__ , generator=SCREAMING_SNAKE_CASE__ , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' ).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def __A ( self : Dict ) -> Dict: __lowerCamelCase = VersatileDiffusionTextToImagePipeline.from_pretrained( '''shi-labs/versatile-diffusion''' , torch_dtype=torch.floataa ) pipe.to(SCREAMING_SNAKE_CASE__ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = "A painting of a squirrel eating a burger " __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe( prompt=SCREAMING_SNAKE_CASE__ , generator=SCREAMING_SNAKE_CASE__ , guidance_scale=7.5 , num_inference_steps=50 , output_type='''numpy''' ).images __lowerCamelCase = image[0, 2_53:2_56, 2_53:2_56, -1] assert image.shape == (1, 5_12, 5_12, 3) __lowerCamelCase = np.array([0.3367, 0.3169, 0.2656, 0.3870, 0.4790, 0.3796, 0.4009, 0.4878, 0.4778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
298
from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from ... import AutoBackbone from ...modeling_outputs import SemanticSegmenterOutput from ...modeling_utils import PreTrainedModel from ...utils import add_start_docstrings, add_start_docstrings_to_model_forward, replace_return_docstrings from ...utils.backbone_utils import BackboneMixin from .configuration_upernet import UperNetConfig _snake_case : Optional[Any] = [ "openmmlab/upernet-convnext-tiny", # See all UperNet models at https://huggingface.co/models?filter=upernet ] # General docstring _snake_case : Dict = "UperNetConfig" class a (nn.Module ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : Union[int, Tuple[int, int]] , lowerCamelCase : Union[int, Tuple[int, int], str] = 0 , lowerCamelCase : bool = False , lowerCamelCase : Union[int, Tuple[int, int]] = 1 , ) -> None: super().__init__() __snake_case : Union[str, Any] = nn.Convad( in_channels=lowerCamelCase , out_channels=lowerCamelCase , kernel_size=lowerCamelCase , padding=lowerCamelCase , bias=lowerCamelCase , dilation=lowerCamelCase , ) __snake_case : Dict = nn.BatchNormad(lowerCamelCase ) __snake_case : List[Any] = nn.ReLU() def __snake_case ( self : List[Any] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : Dict = self.conv(lowerCamelCase ) __snake_case : int = self.batch_norm(lowerCamelCase ) __snake_case : Optional[Any] = self.activation(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : str , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : int ) -> None: super().__init__() __snake_case : Tuple = [ nn.AdaptiveAvgPoolad(lowerCamelCase ), UperNetConvModule(lowerCamelCase , lowerCamelCase , kernel_size=1 ), ] for i, layer in enumerate(self.layers ): self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : Dict , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : List[str] = input for layer in self.layers: __snake_case : Tuple = layer(lowerCamelCase ) return hidden_state class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Tuple[int, ...] , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : bool ) -> None: super().__init__() __snake_case : Dict = pool_scales __snake_case : List[str] = align_corners __snake_case : List[Any] = in_channels __snake_case : str = channels __snake_case : Optional[Any] = [] for i, pool_scale in enumerate(lowerCamelCase ): __snake_case : Union[str, Any] = UperNetPyramidPoolingBlock(pool_scale=lowerCamelCase , in_channels=lowerCamelCase , channels=lowerCamelCase ) self.blocks.append(lowerCamelCase ) self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> List[torch.Tensor]: __snake_case : Tuple = [] for ppm in self.blocks: __snake_case : Any = ppm(lowerCamelCase ) __snake_case : List[Any] = nn.functional.interpolate( lowerCamelCase , size=x.size()[2:] , mode="bilinear" , align_corners=self.align_corners ) ppm_outs.append(lowerCamelCase ) return ppm_outs class a (nn.Module ): """simple docstring""" def __init__( self : int , lowerCamelCase : str , lowerCamelCase : str ) -> List[str]: super().__init__() __snake_case : Dict = config __snake_case : List[str] = config.pool_scales # e.g. (1, 2, 3, 6) __snake_case : Tuple = in_channels __snake_case : str = config.hidden_size __snake_case : List[str] = False __snake_case : Any = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) # PSP Module __snake_case : Tuple = UperNetPyramidPoolingModule( self.pool_scales , self.in_channels[-1] , self.channels , align_corners=self.align_corners , ) __snake_case : List[str] = UperNetConvModule( self.in_channels[-1] + len(self.pool_scales ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) # FPN Module __snake_case : List[Any] = nn.ModuleList() __snake_case : Dict = nn.ModuleList() for in_channels in self.in_channels[:-1]: # skip the top layer __snake_case : Union[str, Any] = UperNetConvModule(lowerCamelCase , self.channels , kernel_size=1 ) __snake_case : Optional[int] = UperNetConvModule(self.channels , self.channels , kernel_size=3 , padding=1 ) self.lateral_convs.append(lowerCamelCase ) self.fpn_convs.append(lowerCamelCase ) __snake_case : int = UperNetConvModule( len(self.in_channels ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) def __snake_case ( self : List[str] ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> str: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : List[Any] , lowerCamelCase : Tuple ) -> Optional[int]: __snake_case : str = inputs[-1] __snake_case : int = [x] psp_outs.extend(self.psp_modules(lowerCamelCase ) ) __snake_case : Tuple = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Union[str, Any] = self.bottleneck(lowerCamelCase ) return output def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # build laterals __snake_case : Any = [lateral_conv(encoder_hidden_states[i] ) for i, lateral_conv in enumerate(self.lateral_convs )] laterals.append(self.psp_forward(lowerCamelCase ) ) # build top-down path __snake_case : Dict = len(lowerCamelCase ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Union[str, Any] = laterals[i - 1].shape[2:] __snake_case : Optional[int] = laterals[i - 1] + nn.functional.interpolate( laterals[i] , size=lowerCamelCase , mode="bilinear" , align_corners=self.align_corners ) # build outputs __snake_case : str = [self.fpn_convs[i](laterals[i] ) for i in range(used_backbone_levels - 1 )] # append psp feature fpn_outs.append(laterals[-1] ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Tuple = nn.functional.interpolate( fpn_outs[i] , size=fpn_outs[0].shape[2:] , mode="bilinear" , align_corners=self.align_corners ) __snake_case : str = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Optional[Any] = self.fpn_bottleneck(lowerCamelCase ) __snake_case : Tuple = self.classifier(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Optional[int] , lowerCamelCase : int = 2 , lowerCamelCase : int = 3 , lowerCamelCase : Union[int, Tuple[int, int]] = 1 ) -> None: super().__init__() __snake_case : List[Any] = config __snake_case : List[str] = config.auxiliary_in_channels __snake_case : List[Any] = config.auxiliary_channels __snake_case : Tuple = config.auxiliary_num_convs __snake_case : int = config.auxiliary_concat_input __snake_case : Optional[int] = in_index __snake_case : Tuple = (kernel_size // 2) * dilation __snake_case : Optional[int] = [] convs.append( UperNetConvModule( self.in_channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) for i in range(self.num_convs - 1 ): convs.append( UperNetConvModule( self.channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) if self.num_convs == 0: __snake_case : Union[str, Any] = nn.Identity() else: __snake_case : Any = nn.Sequential(*lowerCamelCase ) if self.concat_input: __snake_case : int = UperNetConvModule( self.in_channels + self.channels , self.channels , kernel_size=lowerCamelCase , padding=kernel_size // 2 ) __snake_case : Optional[Any] = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) def __snake_case ( self : Dict ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : Tuple ) -> Optional[int]: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : Optional[int] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # just take the relevant feature maps __snake_case : List[str] = encoder_hidden_states[self.in_index] __snake_case : Optional[Any] = self.convs(lowerCamelCase ) if self.concat_input: __snake_case : Optional[Any] = self.conv_cat(torch.cat([hidden_states, output] , dim=1 ) ) __snake_case : Union[str, Any] = self.classifier(lowerCamelCase ) return output class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = UperNetConfig __UpperCAmelCase : int = "pixel_values" __UpperCAmelCase : str = True def __snake_case ( self : Union[str, Any] , lowerCamelCase : List[Any] ) -> List[Any]: if isinstance(lowerCamelCase , lowerCamelCase ): module.backbone.init_weights() module.decode_head.init_weights() module.auxiliary_head.init_weights() def __snake_case ( self : Optional[Any] ) -> List[str]: self.backbone.init_weights() self.decode_head.init_weights() self.auxiliary_head.init_weights() def __snake_case ( self : List[Any] , lowerCamelCase : Any , lowerCamelCase : Optional[Any]=False ) -> Dict: if isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : Union[str, Any] = value _snake_case : Dict = R"\n Parameters:\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n config ([`UperNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" _snake_case : Tuple = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using\n [`AutoImageProcessor`]. See [`SegformerImageProcessor.__call__`] for details.\n output_attentions (`bool`, *optional*):\n Whether or not to return the attentions tensors of all attention layers in case the backbone has them. See\n `attentions` under returned tensors for more detail.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers of the backbone. See `hidden_states` under\n returned tensors for more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n" @add_start_docstrings( "UperNet framework leveraging any vision backbone e.g. for ADE20k, CityScapes." , _lowerCAmelCase , ) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Dict , lowerCamelCase : int ) -> Optional[int]: super().__init__(lowerCamelCase ) __snake_case : Any = AutoBackbone.from_config(config.backbone_config ) # Semantic segmentation head(s) __snake_case : Union[str, Any] = UperNetHead(lowerCamelCase , in_channels=self.backbone.channels ) __snake_case : Any = UperNetFCNHead(lowerCamelCase ) if config.use_auxiliary_head else None # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UPERNET_INPUTS_DOCSTRING.format("batch_size, sequence_length" ) ) @replace_return_docstrings(output_type=lowerCamelCase , config_class=_CONFIG_FOR_DOC ) def __snake_case ( self : Union[str, Any] , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , ) -> Union[tuple, SemanticSegmenterOutput]: __snake_case : Any = return_dict if return_dict is not None else self.config.use_return_dict __snake_case : int = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __snake_case : str = output_attentions if output_attentions is not None else self.config.output_attentions __snake_case : Tuple = self.backbone.forward_with_filtered_kwargs( lowerCamelCase , output_hidden_states=lowerCamelCase , output_attentions=lowerCamelCase ) __snake_case : List[Any] = outputs.feature_maps __snake_case : List[Any] = self.decode_head(lowerCamelCase ) __snake_case : List[str] = nn.functional.interpolate(lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : Optional[int] = None if self.auxiliary_head is not None: __snake_case : Dict = self.auxiliary_head(lowerCamelCase ) __snake_case : Dict = nn.functional.interpolate( lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : int = None if labels is not None: if self.config.num_labels == 1: raise ValueError("The number of labels should be greater than one" ) else: # compute weighted loss __snake_case : Any = CrossEntropyLoss(ignore_index=self.config.loss_ignore_index ) __snake_case : Union[str, Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = main_loss + self.config.auxiliary_loss_weight * auxiliary_loss if not return_dict: if output_hidden_states: __snake_case : Any = (logits,) + outputs[1:] else: __snake_case : str = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return SemanticSegmenterOutput( loss=lowerCamelCase , logits=lowerCamelCase , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
81
0
"""simple docstring""" def lowercase__( __SCREAMING_SNAKE_CASE : Dict = 10 ): if not isinstance(__lowerCamelCase , __lowerCamelCase ) or n < 0: raise ValueError('Invalid input' ) lowercase_ : List[str] = 10**n lowercase_ : Optional[int] = 2_84_33 * (pow(2 , 7_83_04_57 , __lowerCamelCase )) + 1 return str(number % modulus ) if __name__ == "__main__": from doctest import testmod testmod() print(F"{solution(10) = }")
425
from __future__ import annotations def lowerCAmelCase_ ( __lowerCamelCase ): if not nums: return 0 __snake_case : Any = nums[0] __snake_case : str = 0 for num in nums[1:]: __snake_case , __snake_case : List[str] = ( max_excluding + num, max(__lowerCamelCase , __lowerCamelCase ), ) return max(__lowerCamelCase , __lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
81
0
'''simple docstring''' import copy import inspect import unittest from transformers import PretrainedConfig, SwiftFormerConfig from transformers.testing_utils import ( require_torch, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwiftFormerForImageClassification, SwiftFormerModel from transformers.models.swiftformer.modeling_swiftformer import SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class lowercase__ : '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=3 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=2_2_4 , lowerCamelCase__=1_0_0_0 , lowerCamelCase__=[3, 3, 6, 4] , lowerCamelCase__=[4_8, 5_6, 1_1_2, 2_2_0] , ): '''simple docstring''' UpperCamelCase = parent UpperCamelCase = batch_size UpperCamelCase = num_channels UpperCamelCase = is_training UpperCamelCase = use_labels UpperCamelCase = hidden_dropout_prob UpperCamelCase = attention_probs_dropout_prob UpperCamelCase = num_labels UpperCamelCase = image_size UpperCamelCase = layer_depths UpperCamelCase = embed_dims def UpperCAmelCase ( self ): '''simple docstring''' UpperCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase = None if self.use_labels: UpperCamelCase = ids_tensor([self.batch_size] , self.num_labels ) UpperCamelCase = self.get_config() return config, pixel_values, labels def UpperCAmelCase ( self ): '''simple docstring''' return SwiftFormerConfig( depths=self.layer_depths , embed_dims=self.embed_dims , mlp_ratio=4 , downsamples=[True, True, True, True] , hidden_act='''gelu''' , num_labels=self.num_labels , down_patch_size=3 , down_stride=2 , down_pad=1 , drop_rate=0.0 , drop_path_rate=0.0 , use_layer_scale=lowerCamelCase__ , layer_scale_init_value=1e-5 , ) def UpperCAmelCase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' UpperCamelCase = SwiftFormerModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() UpperCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dims[-1], 7, 7) ) def UpperCAmelCase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' UpperCamelCase = self.num_labels UpperCamelCase = SwiftFormerForImageClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() UpperCamelCase = model(lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) UpperCamelCase = SwiftFormerForImageClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() UpperCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase ( self ): '''simple docstring''' (UpperCamelCase) = self.prepare_config_and_inputs() UpperCamelCase = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class lowercase__ ( _lowerCAmelCase, _lowerCAmelCase, unittest.TestCase ): '''simple docstring''' _snake_case = (SwiftFormerModel, SwiftFormerForImageClassification) if is_torch_available() else () _snake_case = ( {"feature-extraction": SwiftFormerModel, "image-classification": SwiftFormerForImageClassification} if is_torch_available() else {} ) _snake_case = False _snake_case = False _snake_case = False _snake_case = False _snake_case = False def UpperCAmelCase ( self ): '''simple docstring''' UpperCamelCase = SwiftFormerModelTester(self ) UpperCamelCase = ConfigTester( self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ , hidden_size=3_7 , num_attention_heads=1_2 , num_hidden_layers=1_2 , ) def UpperCAmelCase ( self ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='''SwiftFormer does not use inputs_embeds''' ) def UpperCAmelCase ( self ): '''simple docstring''' pass def UpperCAmelCase ( self ): '''simple docstring''' UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase = model_class(lowerCamelCase__ ) UpperCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase__ , nn.Linear ) ) def UpperCAmelCase ( self ): '''simple docstring''' UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase = model_class(lowerCamelCase__ ) UpperCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase = [*signature.parameters.keys()] UpperCamelCase = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def UpperCAmelCase ( self ): '''simple docstring''' UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def UpperCAmelCase ( self ): '''simple docstring''' UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase__ ) @slow def UpperCAmelCase ( self ): '''simple docstring''' for model_name in SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase = SwiftFormerModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) @unittest.skip(reason='''SwiftFormer does not output attentions''' ) def UpperCAmelCase ( self ): '''simple docstring''' pass def UpperCAmelCase ( self ): '''simple docstring''' def check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): UpperCamelCase = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): UpperCamelCase = model(**self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) UpperCamelCase = outputs.hidden_states UpperCamelCase = 8 self.assertEqual(len(lowerCamelCase__ ) , lowerCamelCase__ ) # TODO # SwiftFormer's feature maps are of shape (batch_size, embed_dims, height, width) # with the width and height being successively divided by 2, after every 2 blocks for i in range(len(lowerCamelCase__ ) ): self.assertEqual( hidden_states[i].shape , torch.Size( [ self.model_tester.batch_size, self.model_tester.embed_dims[i // 2], (self.model_tester.image_size // 4) // 2 ** (i // 2), (self.model_tester.image_size // 4) // 2 ** (i // 2), ] ) , ) UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase = True check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCamelCase = True check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def UpperCAmelCase ( self ): '''simple docstring''' def _config_zero_init(lowerCamelCase__ ): UpperCamelCase = copy.deepcopy(lowerCamelCase__ ) for key in configs_no_init.__dict__.keys(): if "_range" in key or "_std" in key or "initializer_factor" in key or "layer_scale" in key: setattr(lowerCamelCase__ , lowerCamelCase__ , 1e-1_0 ) if isinstance(getattr(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) , lowerCamelCase__ ): UpperCamelCase = _config_zero_init(getattr(lowerCamelCase__ , lowerCamelCase__ ) ) setattr(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) return configs_no_init UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase = _config_zero_init(lowerCamelCase__ ) for model_class in self.all_model_classes: UpperCamelCase = model_class(config=lowerCamelCase__ ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9) / 1e9).round().item() , [0.0, 1.0] , msg=f'Parameter {name} of model {model_class} seems not properly initialized' , ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def UpperCAmelCase ( self ): '''simple docstring''' pass def __snake_case ( ): UpperCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''') return image @require_torch @require_vision class lowercase__ ( unittest.TestCase ): '''simple docstring''' @cached_property def UpperCAmelCase ( self ): '''simple docstring''' return ViTImageProcessor.from_pretrained('''MBZUAI/swiftformer-xs''' ) if is_vision_available() else None @slow def UpperCAmelCase ( self ): '''simple docstring''' UpperCamelCase = SwiftFormerForImageClassification.from_pretrained('''MBZUAI/swiftformer-xs''' ).to(lowerCamelCase__ ) UpperCamelCase = self.default_image_processor UpperCamelCase = prepare_img() UpperCamelCase = image_processor(images=lowerCamelCase__ , return_tensors='''pt''' ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): UpperCamelCase = model(**lowerCamelCase__ ) # verify the logits UpperCamelCase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) UpperCamelCase = torch.tensor([[-2.1_7_0_3e0_0, 2.1_1_0_7e0_0, -2.0_8_1_1e0_0]] ).to(lowerCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase__ , atol=1e-4 ) )
212
from __future__ import annotations from typing import Any def lowerCAmelCase_ ( __lowerCamelCase ): create_state_space_tree(__lowerCamelCase , [] , 0 ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): if index == len(__lowerCamelCase ): print(__lowerCamelCase ) return create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.pop() if __name__ == "__main__": _snake_case : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["A", "B", "C"]) generate_all_subsequences(seq)
81
0
import warnings from transformers import AutoTokenizer from transformers.utils import is_torch_available from transformers.utils.generic import ExplicitEnum from ...processing_utils import ProcessorMixin if is_torch_available(): import torch class __magic_name__ ( _lowerCAmelCase ): """simple docstring""" __UpperCamelCase = "char" __UpperCamelCase = "bpe" __UpperCamelCase = "wp" _lowerCAmelCase : List[Any] = (DecodeType.CHARACTER, DecodeType.BPE, DecodeType.WORDPIECE) class __magic_name__ ( _lowerCAmelCase ): """simple docstring""" __UpperCamelCase = ["image_processor", "char_tokenizer"] __UpperCamelCase = "ViTImageProcessor" __UpperCamelCase = "MgpstrTokenizer" def __init__( self :Any , snake_case :Tuple=None , snake_case :Optional[int]=None , **snake_case :List[str] ): '''simple docstring''' A_ : str = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , snake_case , ) A_ : int = kwargs.pop("feature_extractor" ) A_ : int = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) A_ : List[str] = tokenizer A_ : Optional[Any] = AutoTokenizer.from_pretrained("gpt2" ) A_ : Union[str, Any] = AutoTokenizer.from_pretrained("bert-base-uncased" ) super().__init__(snake_case , snake_case ) def __call__( self :List[Any] , snake_case :List[str]=None , snake_case :Any=None , snake_case :Optional[int]=None , **snake_case :Any ): '''simple docstring''' if images is None and text is None: raise ValueError("You need to specify either an `images` or `text` input to process." ) if images is not None: A_ : str = self.image_processor(snake_case , return_tensors=snake_case , **snake_case ) if text is not None: A_ : Tuple = self.char_tokenizer(snake_case , return_tensors=snake_case , **snake_case ) if text is None: return inputs elif images is None: return encodings else: A_ : Dict = encodings["input_ids"] return inputs def SCREAMING_SNAKE_CASE ( self :List[str] , snake_case :str ): '''simple docstring''' A_ : Tuple = sequences A_ : List[str] = char_preds.size(0 ) A_ : Optional[Any] = self._decode_helper(snake_case , "char" ) A_ : Tuple = self._decode_helper(snake_case , "bpe" ) A_ : Optional[int] = self._decode_helper(snake_case , "wp" ) A_ : Tuple = [] A_ : Union[str, Any] = [] for i in range(snake_case ): A_ : List[str] = [char_scores[i], bpe_scores[i], wp_scores[i]] A_ : Optional[Any] = [char_strs[i], bpe_strs[i], wp_strs[i]] A_ : List[str] = scores.index(max(snake_case ) ) final_strs.append(strs[max_score_index] ) final_scores.append(scores[max_score_index] ) A_ : Union[str, Any] = {} A_ : List[Any] = final_strs A_ : Union[str, Any] = final_scores A_ : Any = char_strs A_ : List[str] = bpe_strs A_ : Union[str, Any] = wp_strs return out def SCREAMING_SNAKE_CASE ( self :Any , snake_case :List[Any] , snake_case :Dict ): '''simple docstring''' if format == DecodeType.CHARACTER: A_ : Optional[Any] = self.char_decode A_ : Union[str, Any] = 1 A_ : Tuple = "[s]" elif format == DecodeType.BPE: A_ : int = self.bpe_decode A_ : int = 2 A_ : List[str] = "#" elif format == DecodeType.WORDPIECE: A_ : int = self.wp_decode A_ : Tuple = 102 A_ : str = "[SEP]" else: raise ValueError(f"Format {format} is not supported." ) A_ : Dict = [], [] A_ : Dict = pred_logits.size(0 ) A_ : str = pred_logits.size(1 ) A_ : Any = pred_logits.topk(1 , dim=-1 , largest=snake_case , sorted=snake_case ) A_ : Any = preds_index.view(-1 , snake_case )[:, 1:] A_ : str = decoder(snake_case ) A_ : List[Any] = torch.nn.functional.softmax(snake_case , dim=2 ).max(dim=2 ) A_ : Tuple = preds_max_prob[:, 1:] for index in range(snake_case ): A_ : int = preds_str[index].find(snake_case ) A_ : str = preds_str[index][:pred_eos] A_ : List[Any] = preds_index[index].cpu().tolist() A_ : Union[str, Any] = pred_index.index(snake_case ) if eos_token in pred_index else -1 A_ : Any = preds_max_prob[index][: pred_eos_index + 1] A_ : Dict = pred_max_prob.cumprod(dim=0 )[-1] if pred_max_prob.nelement() != 0 else 0.0 dec_strs.append(snake_case ) conf_scores.append(snake_case ) return dec_strs, conf_scores def SCREAMING_SNAKE_CASE ( self :str , snake_case :Union[str, Any] ): '''simple docstring''' A_ : Optional[Any] = [seq.replace(" " , "" ) for seq in self.char_tokenizer.batch_decode(snake_case )] return decode_strs def SCREAMING_SNAKE_CASE ( self :List[str] , snake_case :Tuple ): '''simple docstring''' return self.bpe_tokenizer.batch_decode(snake_case ) def SCREAMING_SNAKE_CASE ( self :Optional[Any] , snake_case :List[str] ): '''simple docstring''' A_ : List[str] = [seq.replace(" " , "" ) for seq in self.wp_tokenizer.batch_decode(snake_case )] return decode_strs
454
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case , __snake_case : List[Any] = image.size __snake_case , __snake_case : Tuple = (x - x % 3_2 for x in (w, h)) # resize to integer multiple of 32 __snake_case : str = image.resize((w, h) , resample=PIL_INTERPOLATION["lanczos"] ) __snake_case : int = np.array(__lowerCamelCase ).astype(np.floataa ) / 2_5_5.0 __snake_case : Union[str, Any] = image[None].transpose(0 , 3 , 1 , 2 ) __snake_case : Union[str, Any] = torch.from_numpy(__lowerCamelCase ) return 2.0 * image - 1.0 class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : VQModel , lowerCamelCase : UNetaDModel , lowerCamelCase : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ) -> Union[str, Any]: super().__init__() self.register_modules(vqvae=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase ) @torch.no_grad() def __call__( self : List[str] , lowerCamelCase : Union[torch.Tensor, PIL.Image.Image] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : Optional[int] = 100 , lowerCamelCase : Optional[float] = 0.0 , lowerCamelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , ) -> Union[Tuple, ImagePipelineOutput]: if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : Any = 1 elif isinstance(lowerCamelCase , torch.Tensor ): __snake_case : Any = image.shape[0] else: raise ValueError(F'`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(lowerCamelCase )}' ) if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : List[Any] = preprocess(lowerCamelCase ) __snake_case , __snake_case : int = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image __snake_case : str = (batch_size, self.unet.config.in_channels // 2, height, width) __snake_case : str = next(self.unet.parameters() ).dtype __snake_case : Tuple = randn_tensor(lowerCamelCase , generator=lowerCamelCase , device=self.device , dtype=lowerCamelCase ) __snake_case : List[Any] = image.to(device=self.device , dtype=lowerCamelCase ) # set timesteps and move to the correct device self.scheduler.set_timesteps(lowerCamelCase , device=self.device ) __snake_case : str = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler __snake_case : Dict = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] __snake_case : Union[str, Any] = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __snake_case : int = {} if accepts_eta: __snake_case : List[str] = eta for t in self.progress_bar(lowerCamelCase ): # concat latents and low resolution image in the channel dimension. __snake_case : Union[str, Any] = torch.cat([latents, image] , dim=1 ) __snake_case : Optional[Any] = self.scheduler.scale_model_input(lowerCamelCase , lowerCamelCase ) # predict the noise residual __snake_case : int = self.unet(lowerCamelCase , lowerCamelCase ).sample # compute the previous noisy sample x_t -> x_t-1 __snake_case : Union[str, Any] = self.scheduler.step(lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase ).prev_sample # decode the image latents with the VQVAE __snake_case : List[Any] = self.vqvae.decode(lowerCamelCase ).sample __snake_case : Dict = torch.clamp(lowerCamelCase , -1.0 , 1.0 ) __snake_case : Any = image / 2 + 0.5 __snake_case : Tuple = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __snake_case : Tuple = self.numpy_to_pil(lowerCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCamelCase )
81
0
"""simple docstring""" import argparse from pathlib import Path import torch from packaging import version from torch.onnx import export from diffusers import AutoencoderKL lowerCamelCase__ = version.parse(version.parse(torch.__version__).base_version) < version.parse("1.11") def _SCREAMING_SNAKE_CASE ( UpperCamelCase : List[str] , UpperCamelCase : int , UpperCamelCase : List[Any] , UpperCamelCase : Union[str, Any] , UpperCamelCase : Union[str, Any] , UpperCamelCase : Union[str, Any] , UpperCamelCase : Dict , UpperCamelCase : List[str]=False , ): output_path.parent.mkdir(parents=__lowerCamelCase , exist_ok=__lowerCamelCase ) # PyTorch deprecated the `enable_onnx_checker` and `use_external_data_format` arguments in v1.11, # so we check the torch version for backwards compatibility if is_torch_less_than_1_11: export( __lowerCamelCase , __lowerCamelCase , f=output_path.as_posix() , input_names=__lowerCamelCase , output_names=__lowerCamelCase , dynamic_axes=__lowerCamelCase , do_constant_folding=__lowerCamelCase , use_external_data_format=__lowerCamelCase , enable_onnx_checker=__lowerCamelCase , opset_version=__lowerCamelCase , ) else: export( __lowerCamelCase , __lowerCamelCase , f=output_path.as_posix() , input_names=__lowerCamelCase , output_names=__lowerCamelCase , dynamic_axes=__lowerCamelCase , do_constant_folding=__lowerCamelCase , opset_version=__lowerCamelCase , ) @torch.no_grad() def _SCREAMING_SNAKE_CASE ( UpperCamelCase : List[Any] , UpperCamelCase : Optional[Any] , UpperCamelCase : Any , UpperCamelCase : Any = False ): A__ = torch.floataa if fpaa else torch.floataa if fpaa and torch.cuda.is_available(): A__ = "cuda" elif fpaa and not torch.cuda.is_available(): raise ValueError("""`float16` model export is only supported on GPUs with CUDA""" ) else: A__ = "cpu" A__ = Path(__lowerCamelCase ) # VAE DECODER A__ = AutoencoderKL.from_pretrained(model_path + """/vae""" ) A__ = vae_decoder.config.latent_channels # forward only through the decoder part A__ = vae_decoder.decode onnx_export( __lowerCamelCase , model_args=( torch.randn(1 , __lowerCamelCase , 25 , 25 ).to(device=__lowerCamelCase , dtype=__lowerCamelCase ), False, ) , output_path=output_path / """vae_decoder""" / """model.onnx""" , ordered_input_names=["""latent_sample""", """return_dict"""] , output_names=["""sample"""] , dynamic_axes={ """latent_sample""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, } , opset=__lowerCamelCase , ) del vae_decoder if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() parser.add_argument( "--model_path", type=str, required=True, help="Path to the `diffusers` checkpoint to convert (either a local directory or on the Hub).", ) parser.add_argument("--output_path", type=str, required=True, help="Path to the output model.") parser.add_argument( "--opset", default=14, type=int, help="The version of the ONNX operator set to use.", ) parser.add_argument("--fp16", action="store_true", default=False, help="Export the models in `float16` mode") lowerCamelCase__ = parser.parse_args() print(args.output_path) convert_models(args.model_path, args.output_path, args.opset, args.fpaa) print("SD: Done: ONNX")
574
import gc import unittest import torch from parameterized import parameterized from diffusers import AutoencoderKL from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = AutoencoderKL __UpperCAmelCase : Optional[Any] = "sample" __UpperCAmelCase : Optional[int] = 1e-2 @property def __snake_case ( self : Dict ) -> Optional[Any]: __snake_case : Optional[Any] = 4 __snake_case : Tuple = 3 __snake_case : List[str] = (32, 32) __snake_case : str = floats_tensor((batch_size, num_channels) + sizes ).to(lowerCamelCase ) return {"sample": image} @property def __snake_case ( self : Union[str, Any] ) -> Tuple: return (3, 32, 32) @property def __snake_case ( self : int ) -> int: return (3, 32, 32) def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Optional[Any] = { "block_out_channels": [32, 64], "in_channels": 3, "out_channels": 3, "down_block_types": ["DownEncoderBlock2D", "DownEncoderBlock2D"], "up_block_types": ["UpDecoderBlock2D", "UpDecoderBlock2D"], "latent_channels": 4, } __snake_case : Any = self.dummy_input return init_dict, inputs_dict def __snake_case ( self : str ) -> Dict: pass def __snake_case ( self : Tuple ) -> List[str]: pass @unittest.skipIf(torch_device == "mps" , "Gradient checkpointing skipped on MPS" ) def __snake_case ( self : Any ) -> Optional[Any]: # enable deterministic behavior for gradient checkpointing __snake_case , __snake_case : int = self.prepare_init_args_and_inputs_for_common() __snake_case : str = self.model_class(**lowerCamelCase ) model.to(lowerCamelCase ) assert not model.is_gradient_checkpointing and model.training __snake_case : str = model(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model.zero_grad() __snake_case : Any = torch.randn_like(lowerCamelCase ) __snake_case : str = (out - labels).mean() loss.backward() # re-instantiate the model now enabling gradient checkpointing __snake_case : Optional[int] = self.model_class(**lowerCamelCase ) # clone model model_a.load_state_dict(model.state_dict() ) model_a.to(lowerCamelCase ) model_a.enable_gradient_checkpointing() assert model_a.is_gradient_checkpointing and model_a.training __snake_case : int = model_a(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model_a.zero_grad() __snake_case : Union[str, Any] = (out_a - labels).mean() loss_a.backward() # compare the output and parameters gradients self.assertTrue((loss - loss_a).abs() < 1E-5 ) __snake_case : Optional[int] = dict(model.named_parameters() ) __snake_case : List[Any] = dict(model_a.named_parameters() ) for name, param in named_params.items(): self.assertTrue(torch_all_close(param.grad.data , named_params_a[name].grad.data , atol=5E-5 ) ) def __snake_case ( self : List[Any] ) -> Optional[int]: __snake_case , __snake_case : Optional[Any] = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" , output_loading_info=lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(lowerCamelCase ) __snake_case : Optional[Any] = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def __snake_case ( self : Optional[Any] ) -> Union[str, Any]: __snake_case : Tuple = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" ) __snake_case : Dict = model.to(lowerCamelCase ) model.eval() if torch_device == "mps": __snake_case : int = torch.manual_seed(0 ) else: __snake_case : str = torch.Generator(device=lowerCamelCase ).manual_seed(0 ) __snake_case : List[str] = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) __snake_case : Union[str, Any] = image.to(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , sample_posterior=lowerCamelCase , generator=lowerCamelCase ).sample __snake_case : List[Any] = output[0, -1, -3:, -3:].flatten().cpu() # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. if torch_device == "mps": __snake_case : Union[str, Any] = torch.tensor( [ -4.0078E-01, -3.8323E-04, -1.2681E-01, -1.1462E-01, 2.0095E-01, 1.0893E-01, -8.8247E-02, -3.0361E-01, -9.8644E-03, ] ) elif torch_device == "cpu": __snake_case : Tuple = torch.tensor( [-0.13_52, 0.08_78, 0.04_19, -0.08_18, -0.10_69, 0.06_88, -0.14_58, -0.44_46, -0.00_26] ) else: __snake_case : List[str] = torch.tensor( [-0.24_21, 0.46_42, 0.25_07, -0.04_38, 0.06_82, 0.31_60, -0.20_18, -0.07_27, 0.24_85] ) self.assertTrue(torch_all_close(lowerCamelCase , lowerCamelCase , rtol=1E-2 ) ) @slow class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : int , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] ) -> List[str]: return F'gaussian_noise_s={seed}_shape={"_".join([str(lowerCamelCase ) for s in shape] )}.npy' def __snake_case ( self : List[Any] ) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __snake_case ( self : Tuple , lowerCamelCase : List[Any]=0 , lowerCamelCase : Tuple=(4, 3, 512, 512) , lowerCamelCase : Optional[int]=False ) -> str: __snake_case : List[Any] = torch.floataa if fpaa else torch.floataa __snake_case : Tuple = torch.from_numpy(load_hf_numpy(self.get_file_format(lowerCamelCase , lowerCamelCase ) ) ).to(lowerCamelCase ).to(lowerCamelCase ) return image def __snake_case ( self : Optional[Any] , lowerCamelCase : int="CompVis/stable-diffusion-v1-4" , lowerCamelCase : int=False ) -> int: __snake_case : str = "fp16" if fpaa else None __snake_case : int = torch.floataa if fpaa else torch.floataa __snake_case : int = AutoencoderKL.from_pretrained( lowerCamelCase , subfolder="vae" , torch_dtype=lowerCamelCase , revision=lowerCamelCase , ) model.to(lowerCamelCase ).eval() return model def __snake_case ( self : str , lowerCamelCase : int=0 ) -> Optional[Any]: if torch_device == "mps": return torch.manual_seed(lowerCamelCase ) return torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) @parameterized.expand( [ # fmt: off [33, [-0.16_03, 0.98_78, -0.04_95, -0.07_90, -0.27_09, 0.83_75, -0.20_60, -0.08_24], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_76, 0.11_68, 0.13_32, -0.48_40, -0.25_08, -0.07_91, -0.04_93, -0.40_89], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[str] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] ) -> List[Any]: __snake_case : Optional[Any] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) __snake_case : Tuple = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : List[Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : int = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [33, [-0.05_13, 0.02_89, 1.37_99, 0.21_66, -0.25_73, -0.08_71, 0.51_03, -0.09_99]], [47, [-0.41_28, -0.13_20, -0.37_04, 0.19_65, -0.41_16, -0.23_32, -0.33_40, 0.22_47]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : Any , lowerCamelCase : List[str] , lowerCamelCase : List[str] ) -> Tuple: __snake_case : Any = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Any = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.16_09, 0.98_66, -0.04_87, -0.07_77, -0.27_16, 0.83_68, -0.20_55, -0.08_14], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_77, 0.11_47, 0.13_33, -0.48_41, -0.25_06, -0.08_05, -0.04_91, -0.40_85], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Any , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) with torch.no_grad(): __snake_case : int = model(lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Union[str, Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : List[str] = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [13, [-0.20_51, -0.18_03, -0.23_11, -0.21_14, -0.32_92, -0.35_74, -0.29_53, -0.33_23]], [37, [-0.26_32, -0.26_25, -0.21_99, -0.27_41, -0.45_39, -0.49_90, -0.37_20, -0.49_25]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : List[str] , lowerCamelCase : Tuple , lowerCamelCase : Any ) -> Optional[Any]: __snake_case : List[str] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : str = sample[-1, -2:, :2, -2:].flatten().cpu() __snake_case : Optional[int] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-3 ) @parameterized.expand( [ # fmt: off [27, [-0.03_69, 0.02_07, -0.07_76, -0.06_82, -0.17_47, -0.19_30, -0.14_65, -0.20_39]], [16, [-0.16_28, -0.21_34, -0.27_47, -0.26_42, -0.37_74, -0.44_04, -0.36_87, -0.42_77]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : str , lowerCamelCase : Optional[int] , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[str] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : Union[str, Any] = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Optional[Any] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=5E-3 ) @parameterized.expand([(13,), (16,), (27,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> Tuple: __snake_case : Dict = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : Any = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Any = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-1 ) @parameterized.expand([(13,), (16,), (37,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : List[Any] , lowerCamelCase : Any ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : Union[str, Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : List[Any] = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Dict = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.30_01, 0.09_18, -2.69_84, -3.97_20, -3.20_99, -5.03_53, 1.73_38, -0.20_65, 3.42_67]], [47, [-1.50_30, -4.38_71, -6.03_55, -9.11_57, -1.66_61, -2.78_53, 2.16_07, -5.08_23, 2.56_33]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Dict ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : int = self.get_sd_image(lowerCamelCase ) __snake_case : int = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model.encode(lowerCamelCase ).latent_dist __snake_case : Dict = dist.sample(generator=lowerCamelCase ) assert list(sample.shape ) == [image.shape[0], 4] + [i // 8 for i in image.shape[2:]] __snake_case : List[str] = sample[0, -1, -3:, -3:].flatten().cpu() __snake_case : Dict = torch.tensor(lowerCamelCase ) __snake_case : Dict = 3E-3 if torch_device != "mps" else 1E-2 assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=lowerCamelCase )
81
0
'''simple docstring''' import argparse import pytorch_lightning as pl import torch from torch import nn from transformers import LongformerForQuestionAnswering, LongformerModel class UpperCAmelCase ( pl.LightningModule ): '''simple docstring''' def __init__( self , __lowerCAmelCase ) -> Any: super().__init__() lowercase__ : Any = model lowercase__ : List[str] = 2 lowercase__ : List[Any] = nn.Linear(self.model.config.hidden_size , self.num_labels ) def _lowerCAmelCase( self ) -> List[Any]: pass def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): # load longformer model from model identifier lowercase__ : List[str] = LongformerModel.from_pretrained(__lowerCamelCase ) lowercase__ : Union[str, Any] = LightningModel(__lowerCamelCase ) lowercase__ : Dict = torch.load(__lowerCamelCase , map_location=torch.device('''cpu''' ) ) lightning_model.load_state_dict(ckpt['''state_dict'''] ) # init longformer question answering model lowercase__ : List[str] = LongformerForQuestionAnswering.from_pretrained(__lowerCamelCase ) # transfer weights longformer_for_qa.longformer.load_state_dict(lightning_model.model.state_dict() ) longformer_for_qa.qa_outputs.load_state_dict(lightning_model.qa_outputs.state_dict() ) longformer_for_qa.eval() # save model longformer_for_qa.save_pretrained(__lowerCamelCase ) print(F"""Conversion successful. Model saved under {pytorch_dump_folder_path}""" ) if __name__ == "__main__": __a: int = argparse.ArgumentParser() # Required parameters parser.add_argument( """--longformer_model""", default=None, type=str, required=True, help="""model identifier of longformer. Should be either `longformer-base-4096` or `longformer-large-4096`.""", ) parser.add_argument( """--longformer_question_answering_ckpt_path""", default=None, type=str, required=True, help="""Path the official PyTorch Lightning Checkpoint.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) __a: str = parser.parse_args() convert_longformer_qa_checkpoint_to_pytorch( args.longformer_model, args.longformer_question_answering_ckpt_path, args.pytorch_dump_folder_path )
152
import warnings from ...utils import logging from .image_processing_flava import FlavaImageProcessor _snake_case : Optional[int] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[Any] , *lowerCamelCase : Any , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class FlavaFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use FlavaImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roberta import RobertaTokenizer lowerCAmelCase_ = logging.get_logger(__name__) lowerCAmelCase_ = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} lowerCAmelCase_ = { "vocab_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/vocab.json", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/vocab.json", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/vocab.json", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/vocab.json", "roberta-base-openai-detector": "https://huggingface.co/roberta-base-openai-detector/resolve/main/vocab.json", "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/vocab.json" ), }, "merges_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/merges.txt", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/merges.txt", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/merges.txt", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/merges.txt", "roberta-base-openai-detector": "https://huggingface.co/roberta-base-openai-detector/resolve/main/merges.txt", "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/merges.txt" ), }, "tokenizer_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/tokenizer.json", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/tokenizer.json", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/tokenizer.json", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/tokenizer.json", "roberta-base-openai-detector": ( "https://huggingface.co/roberta-base-openai-detector/resolve/main/tokenizer.json" ), "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/tokenizer.json" ), }, } lowerCAmelCase_ = { "roberta-base": 5_1_2, "roberta-large": 5_1_2, "roberta-large-mnli": 5_1_2, "distilroberta-base": 5_1_2, "roberta-base-openai-detector": 5_1_2, "roberta-large-openai-detector": 5_1_2, } class A (_lowerCAmelCase ): _SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES _SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP _SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _SCREAMING_SNAKE_CASE = ["input_ids", "attention_mask"] _SCREAMING_SNAKE_CASE = RobertaTokenizer def __init__( self , lowercase_=None , lowercase_=None , lowercase_=None , lowercase_="replace" , lowercase_="<s>" , lowercase_="</s>" , lowercase_="</s>" , lowercase_="<s>" , lowercase_="<unk>" , lowercase_="<pad>" , lowercase_="<mask>" , lowercase_=False , lowercase_=True , **lowercase_ , ) -> Union[str, Any]: '''simple docstring''' super().__init__( lowercase_ , lowercase_ , tokenizer_file=lowercase_ , errors=lowercase_ , bos_token=lowercase_ , eos_token=lowercase_ , sep_token=lowercase_ , cls_token=lowercase_ , unk_token=lowercase_ , pad_token=lowercase_ , mask_token=lowercase_ , add_prefix_space=lowercase_ , trim_offsets=lowercase_ , **lowercase_ , ) _snake_case : Optional[Any] = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('''add_prefix_space''' , lowercase_ ) != add_prefix_space: _snake_case : Any = getattr(lowercase_ , pre_tok_state.pop('''type''' ) ) _snake_case : List[Any] = add_prefix_space _snake_case : Optional[Any] = pre_tok_class(**lowercase_ ) _snake_case : Any = add_prefix_space _snake_case : str = "post_processor" _snake_case : Optional[Any] = getattr(self.backend_tokenizer , lowercase_ , lowercase_ ) if tokenizer_component_instance: _snake_case : int = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: _snake_case : Tuple = tuple(state['''sep'''] ) if "cls" in state: _snake_case : List[Any] = tuple(state['''cls'''] ) _snake_case : Union[str, Any] = False if state.get('''add_prefix_space''' , lowercase_ ) != add_prefix_space: _snake_case : Dict = add_prefix_space _snake_case : Optional[int] = True if state.get('''trim_offsets''' , lowercase_ ) != trim_offsets: _snake_case : Optional[Any] = trim_offsets _snake_case : Optional[Any] = True if changes_to_apply: _snake_case : List[str] = getattr(lowercase_ , state.pop('''type''' ) ) _snake_case : List[Any] = component_class(**lowercase_ ) setattr(self.backend_tokenizer , lowercase_ , lowercase_ ) @property def __a ( self ) -> str: '''simple docstring''' if self._mask_token is None: if self.verbose: logger.error('''Using mask_token, but it is not set yet.''' ) return None return str(self._mask_token ) @mask_token.setter def __a ( self , lowercase_ ) -> List[Any]: '''simple docstring''' _snake_case : Union[str, Any] = AddedToken(lowercase_ , lstrip=lowercase_ , rstrip=lowercase_ ) if isinstance(lowercase_ , lowercase_ ) else value _snake_case : Optional[int] = value def __a ( self , *lowercase_ , **lowercase_ ) -> BatchEncoding: '''simple docstring''' _snake_case : Union[str, Any] = kwargs.get('''is_split_into_words''' , lowercase_ ) assert self.add_prefix_space or not is_split_into_words, ( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*lowercase_ , **lowercase_ ) def __a ( self , *lowercase_ , **lowercase_ ) -> BatchEncoding: '''simple docstring''' _snake_case : Dict = kwargs.get('''is_split_into_words''' , lowercase_ ) assert self.add_prefix_space or not is_split_into_words, ( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._encode_plus(*lowercase_ , **lowercase_ ) def __a ( self , lowercase_ , lowercase_ = None ) -> Tuple[str]: '''simple docstring''' _snake_case : Optional[int] = self._tokenizer.model.save(lowercase_ , name=lowercase_ ) return tuple(lowercase_ ) def __a ( self , lowercase_ , lowercase_=None ) -> int: '''simple docstring''' _snake_case : str = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def __a ( self , lowercase_ , lowercase_ = None ) -> List[int]: '''simple docstring''' _snake_case : Union[str, Any] = [self.sep_token_id] _snake_case : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
326
import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : Tuple ) -> Optional[Any]: __snake_case : Dict = tempfile.mkdtemp() __snake_case : Any = SamImageProcessor() __snake_case : Optional[int] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : Optional[Any] , **lowerCamelCase : Optional[int] ) -> Optional[Any]: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[Any] ) -> Dict: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : int ) -> List[Any]: __snake_case : List[str] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : int = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : List[Any] ) -> Dict: __snake_case : int = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Dict = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Optional[Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : List[str] ) -> Tuple: __snake_case : int = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Optional[int] = self.prepare_image_inputs() __snake_case : List[str] = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : Dict = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_torch def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Tuple = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[str] = [torch.ones((1, 3, 5, 5) )] __snake_case : Tuple = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : int = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , torch.tensor(lowerCamelCase ) , torch.tensor(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : List[str] = [np.ones((1, 3, 5, 5) )] __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : str = [[1, 0], [0, 1]] with self.assertRaises(lowerCamelCase ): __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) @require_vision @require_tf class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[Any] ) -> Union[str, Any]: __snake_case : int = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : str , **lowerCamelCase : Any ) -> Tuple: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> Any: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : str ) -> List[Any]: __snake_case : int = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : Dict = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : int ) -> List[str]: __snake_case : List[Any] = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Optional[Any] = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Tuple = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : Union[str, Any] ) -> List[Any]: __snake_case : str = self.get_image_processor() __snake_case : Union[str, Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : int = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : List[str] = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_tf def __snake_case ( self : Any ) -> Optional[int]: __snake_case : List[str] = self.get_image_processor() __snake_case : Dict = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Union[str, Any] = [tf.ones((1, 3, 5, 5) )] __snake_case : List[Any] = [[1764, 2646]] __snake_case : Dict = [[683, 1024]] __snake_case : List[str] = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , tf.convert_to_tensor(lowerCamelCase ) , tf.convert_to_tensor(lowerCamelCase ) , return_tensors="tf" , ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : Union[str, Any] = [np.ones((1, 3, 5, 5) )] __snake_case : List[str] = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Tuple = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): __snake_case : Dict = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[str] ) -> str: __snake_case : Optional[int] = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : List[str] , **lowerCamelCase : Any ) -> Dict: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> List[Any]: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : str = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : List[Any] = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : str = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[Any] = np.random.randint(0 , 2 , size=(1, 3, 5, 5) ).astype(np.floataa ) __snake_case : Dict = [tf.convert_to_tensor(lowerCamelCase )] __snake_case : List[Any] = [torch.tensor(lowerCamelCase )] __snake_case : Optional[Any] = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : Union[str, Any] = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) __snake_case : Dict = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="pt" ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def __snake_case ( self : List[Any] ) -> List[str]: __snake_case : Any = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : Any = image_processor(lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Optional[Any] = processor(images=lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Tuple = image_processor(lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() __snake_case : List[Any] = processor(images=lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) )
81
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __magic_name__ = logging.get_logger(__name__) __magic_name__ = { "transfo-xl-wt103": "https://huggingface.co/transfo-xl-wt103/resolve/main/config.json", } class SCREAMING_SNAKE_CASE_ ( _lowerCAmelCase ): """simple docstring""" __lowercase : List[str] = "transfo-xl" __lowercase : List[Any] = ["mems"] __lowercase : Optional[Any] = { "n_token": "vocab_size", "hidden_size": "d_model", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self , lowerCAmelCase__=2_6_7_7_3_5 , lowerCAmelCase__=[2_0_0_0_0, 4_0_0_0_0, 2_0_0_0_0_0] , lowerCAmelCase__=1_0_2_4 , lowerCAmelCase__=1_0_2_4 , lowerCAmelCase__=1_6 , lowerCAmelCase__=6_4 , lowerCAmelCase__=4_0_9_6 , lowerCAmelCase__=4 , lowerCAmelCase__=False , lowerCAmelCase__=1_8 , lowerCAmelCase__=1_6_0_0 , lowerCAmelCase__=1_0_0_0 , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=0 , lowerCAmelCase__=-1 , lowerCAmelCase__=True , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.0 , lowerCAmelCase__=True , lowerCAmelCase__="normal" , lowerCAmelCase__=0.01 , lowerCAmelCase__=0.01 , lowerCAmelCase__=0.02 , lowerCAmelCase__=1E-5 , lowerCAmelCase__=0 , **lowerCAmelCase__ , ): __SCREAMING_SNAKE_CASE = vocab_size __SCREAMING_SNAKE_CASE = [] self.cutoffs.extend(lowerCAmelCase__) if proj_share_all_but_first: __SCREAMING_SNAKE_CASE = [False] + [True] * len(self.cutoffs) else: __SCREAMING_SNAKE_CASE = [False] + [False] * len(self.cutoffs) __SCREAMING_SNAKE_CASE = d_model __SCREAMING_SNAKE_CASE = d_embed __SCREAMING_SNAKE_CASE = d_head __SCREAMING_SNAKE_CASE = d_inner __SCREAMING_SNAKE_CASE = div_val __SCREAMING_SNAKE_CASE = pre_lnorm __SCREAMING_SNAKE_CASE = n_layer __SCREAMING_SNAKE_CASE = n_head __SCREAMING_SNAKE_CASE = mem_len __SCREAMING_SNAKE_CASE = same_length __SCREAMING_SNAKE_CASE = attn_type __SCREAMING_SNAKE_CASE = clamp_len __SCREAMING_SNAKE_CASE = sample_softmax __SCREAMING_SNAKE_CASE = adaptive __SCREAMING_SNAKE_CASE = dropout __SCREAMING_SNAKE_CASE = dropatt __SCREAMING_SNAKE_CASE = untie_r __SCREAMING_SNAKE_CASE = init __SCREAMING_SNAKE_CASE = init_range __SCREAMING_SNAKE_CASE = proj_init_std __SCREAMING_SNAKE_CASE = init_std __SCREAMING_SNAKE_CASE = layer_norm_epsilon super().__init__(eos_token_id=lowerCAmelCase__ , **lowerCAmelCase__) @property def snake_case_ ( self): # Message copied from Transformer-XL documentation logger.info(f"The model {self.model_type} is one of the few models that has no sequence length limit.") return -1 @max_position_embeddings.setter def snake_case_ ( self , lowerCAmelCase__): # Message copied from Transformer-XL documentation raise NotImplementedError( f"The model {self.model_type} is one of the few models that has no sequence length limit.")
155
import warnings from ...utils import logging from .image_processing_deit import DeiTImageProcessor _snake_case : List[str] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[str] , *lowerCamelCase : str , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class DeiTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use DeiTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
import unittest from transformers import BertGenerationConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import BertGenerationDecoder, BertGenerationEncoder class lowerCAmelCase_ : def __init__( self : List[str] , SCREAMING_SNAKE_CASE_ : Tuple , SCREAMING_SNAKE_CASE_ : Any=13 , SCREAMING_SNAKE_CASE_ : Tuple=7 , SCREAMING_SNAKE_CASE_ : List[str]=True , SCREAMING_SNAKE_CASE_ : List[str]=True , SCREAMING_SNAKE_CASE_ : List[Any]=99 , SCREAMING_SNAKE_CASE_ : int=32 , SCREAMING_SNAKE_CASE_ : Tuple=5 , SCREAMING_SNAKE_CASE_ : Optional[int]=4 , SCREAMING_SNAKE_CASE_ : Optional[Any]=37 , SCREAMING_SNAKE_CASE_ : Optional[int]="gelu" , SCREAMING_SNAKE_CASE_ : int=0.1 , SCREAMING_SNAKE_CASE_ : Any=0.1 , SCREAMING_SNAKE_CASE_ : List[str]=50 , SCREAMING_SNAKE_CASE_ : Any=0.02 , SCREAMING_SNAKE_CASE_ : Dict=True , SCREAMING_SNAKE_CASE_ : Any=None , ): lowerCAmelCase__ = parent lowerCAmelCase__ = batch_size lowerCAmelCase__ = seq_length lowerCAmelCase__ = is_training lowerCAmelCase__ = use_input_mask lowerCAmelCase__ = vocab_size lowerCAmelCase__ = hidden_size lowerCAmelCase__ = num_hidden_layers lowerCAmelCase__ = num_attention_heads lowerCAmelCase__ = intermediate_size lowerCAmelCase__ = hidden_act lowerCAmelCase__ = hidden_dropout_prob lowerCAmelCase__ = attention_probs_dropout_prob lowerCAmelCase__ = max_position_embeddings lowerCAmelCase__ = initializer_range lowerCAmelCase__ = use_labels lowerCAmelCase__ = scope def __snake_case ( self : Optional[Any] ): lowerCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ = None if self.use_input_mask: lowerCAmelCase__ = random_attention_mask([self.batch_size, self.seq_length] ) if self.use_labels: lowerCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ = self.get_config() return config, input_ids, input_mask, token_labels def __snake_case ( self : Tuple ): return BertGenerationConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , is_decoder=SCREAMING_SNAKE_CASE_ , initializer_range=self.initializer_range , ) def __snake_case ( self : str ): ( lowerCAmelCase__ ) = self.prepare_config_and_inputs() lowerCAmelCase__ = True lowerCAmelCase__ = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) lowerCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, token_labels, encoder_hidden_states, encoder_attention_mask, ) def __snake_case ( self : List[Any] , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : List[Any] , SCREAMING_SNAKE_CASE_ : Optional[int] , **SCREAMING_SNAKE_CASE_ : List[str] , ): lowerCAmelCase__ = BertGenerationEncoder(config=SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE_ , attention_mask=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __snake_case ( self : Dict , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : Any , SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : List[str] , **SCREAMING_SNAKE_CASE_ : Union[str, Any] , ): lowerCAmelCase__ = True lowerCAmelCase__ = BertGenerationEncoder(config=SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() lowerCAmelCase__ = model( SCREAMING_SNAKE_CASE_ , attention_mask=SCREAMING_SNAKE_CASE_ , encoder_hidden_states=SCREAMING_SNAKE_CASE_ , encoder_attention_mask=SCREAMING_SNAKE_CASE_ , ) lowerCAmelCase__ = model( SCREAMING_SNAKE_CASE_ , attention_mask=SCREAMING_SNAKE_CASE_ , encoder_hidden_states=SCREAMING_SNAKE_CASE_ , ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __snake_case ( self : int , SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : Tuple , SCREAMING_SNAKE_CASE_ : Tuple , SCREAMING_SNAKE_CASE_ : str , **SCREAMING_SNAKE_CASE_ : List[str] , ): lowerCAmelCase__ = True lowerCAmelCase__ = True lowerCAmelCase__ = BertGenerationDecoder(config=SCREAMING_SNAKE_CASE_ ).to(SCREAMING_SNAKE_CASE_ ).eval() # first forward pass lowerCAmelCase__ = model( SCREAMING_SNAKE_CASE_ , attention_mask=SCREAMING_SNAKE_CASE_ , encoder_hidden_states=SCREAMING_SNAKE_CASE_ , encoder_attention_mask=SCREAMING_SNAKE_CASE_ , use_cache=SCREAMING_SNAKE_CASE_ , ) lowerCAmelCase__ = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids lowerCAmelCase__ = ids_tensor((self.batch_size, 3) , config.vocab_size ) lowerCAmelCase__ = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and lowerCAmelCase__ = torch.cat([input_ids, next_tokens] , dim=-1 ) lowerCAmelCase__ = torch.cat([input_mask, next_mask] , dim=-1 ) lowerCAmelCase__ = model( SCREAMING_SNAKE_CASE_ , attention_mask=SCREAMING_SNAKE_CASE_ , encoder_hidden_states=SCREAMING_SNAKE_CASE_ , encoder_attention_mask=SCREAMING_SNAKE_CASE_ , output_hidden_states=SCREAMING_SNAKE_CASE_ , )["hidden_states"][0] lowerCAmelCase__ = model( SCREAMING_SNAKE_CASE_ , attention_mask=SCREAMING_SNAKE_CASE_ , encoder_hidden_states=SCREAMING_SNAKE_CASE_ , encoder_attention_mask=SCREAMING_SNAKE_CASE_ , past_key_values=SCREAMING_SNAKE_CASE_ , output_hidden_states=SCREAMING_SNAKE_CASE_ , )["hidden_states"][0] # select random slice lowerCAmelCase__ = ids_tensor((1,) , output_from_past.shape[-1] ).item() lowerCAmelCase__ = output_from_no_past[:, -3:, random_slice_idx].detach() lowerCAmelCase__ = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , atol=1e-3 ) ) def __snake_case ( self : Optional[Any] , SCREAMING_SNAKE_CASE_ : Tuple , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : List[Any] , SCREAMING_SNAKE_CASE_ : List[str] , *SCREAMING_SNAKE_CASE_ : int , ): lowerCAmelCase__ = BertGenerationDecoder(SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE_ , attention_mask=SCREAMING_SNAKE_CASE_ , labels=SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __snake_case ( self : Dict ): lowerCAmelCase__ = self.prepare_config_and_inputs() lowerCAmelCase__ = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class lowerCAmelCase_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): UpperCamelCase_ :str = (BertGenerationEncoder, BertGenerationDecoder) if is_torch_available() else () UpperCamelCase_ :Dict = (BertGenerationDecoder,) if is_torch_available() else () UpperCamelCase_ :Optional[int] = ( {"feature-extraction": BertGenerationEncoder, "text-generation": BertGenerationDecoder} if is_torch_available() else {} ) def __snake_case ( self : str ): lowerCAmelCase__ = BertGenerationEncoderTester(self ) lowerCAmelCase__ = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE_ , hidden_size=37 ) def __snake_case ( self : Optional[Any] ): self.config_tester.run_common_tests() def __snake_case ( self : Union[str, Any] ): lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE_ ) def __snake_case ( self : List[Any] ): lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() lowerCAmelCase__ = "bert" self.model_tester.create_and_check_model(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def __snake_case ( self : Optional[Any] ): lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*SCREAMING_SNAKE_CASE_ ) def __snake_case ( self : int ): lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_decoder_model_past_large_inputs(*SCREAMING_SNAKE_CASE_ ) def __snake_case ( self : List[str] ): # This regression test was failing with PyTorch < 1.3 ( lowerCAmelCase__ ) = self.model_tester.prepare_config_and_inputs_for_decoder() lowerCAmelCase__ = None self.model_tester.create_and_check_model_as_decoder( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , ) def __snake_case ( self : List[str] ): lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_for_causal_lm(*SCREAMING_SNAKE_CASE_ ) @slow def __snake_case ( self : str ): lowerCAmelCase__ = BertGenerationEncoder.from_pretrained('''google/bert_for_seq_generation_L-24_bbc_encoder''' ) self.assertIsNotNone(SCREAMING_SNAKE_CASE_ ) @require_torch class lowerCAmelCase_ ( unittest.TestCase ): @slow def __snake_case ( self : List[Any] ): lowerCAmelCase__ = BertGenerationEncoder.from_pretrained('''google/bert_for_seq_generation_L-24_bbc_encoder''' ) lowerCAmelCase__ = torch.tensor([[101, 7_592, 1_010, 2_026, 3_899, 2_003, 10_140, 102]] ) with torch.no_grad(): lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE_ )[0] lowerCAmelCase__ = torch.Size([1, 8, 1_024] ) self.assertEqual(output.shape , SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ = torch.tensor( [[[0.1_775, 0.0_083, -0.0_321], [1.6_002, 0.1_287, 0.3_912], [2.1_473, 0.5_791, 0.6_066]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , SCREAMING_SNAKE_CASE_ , atol=1e-4 ) ) @require_torch class lowerCAmelCase_ ( unittest.TestCase ): @slow def __snake_case ( self : Any ): lowerCAmelCase__ = BertGenerationDecoder.from_pretrained('''google/bert_for_seq_generation_L-24_bbc_encoder''' ) lowerCAmelCase__ = torch.tensor([[101, 7_592, 1_010, 2_026, 3_899, 2_003, 10_140, 102]] ) with torch.no_grad(): lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE_ )[0] lowerCAmelCase__ = torch.Size([1, 8, 50_358] ) self.assertEqual(output.shape , SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ = torch.tensor( [[[-0.5_788, -2.5_994, -3.7_054], [0.0_438, 4.7_997, 1.8_795], [1.5_862, 6.6_409, 4.4_638]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , SCREAMING_SNAKE_CASE_ , atol=1e-4 ) )
668
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case : Union[str, Any] = { "configuration_owlvit": [ "OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "OwlViTConfig", "OwlViTOnnxConfig", "OwlViTTextConfig", "OwlViTVisionConfig", ], "processing_owlvit": ["OwlViTProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[Any] = ["OwlViTFeatureExtractor"] _snake_case : Optional[int] = ["OwlViTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "OwlViTModel", "OwlViTPreTrainedModel", "OwlViTTextModel", "OwlViTVisionModel", "OwlViTForObjectDetection", ] if TYPE_CHECKING: from .configuration_owlvit import ( OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, OwlViTConfig, OwlViTOnnxConfig, OwlViTTextConfig, OwlViTVisionConfig, ) from .processing_owlvit import OwlViTProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_owlvit import OwlViTFeatureExtractor from .image_processing_owlvit import OwlViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_owlvit import ( OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST, OwlViTForObjectDetection, OwlViTModel, OwlViTPreTrainedModel, OwlViTTextModel, OwlViTVisionModel, ) else: import sys _snake_case : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
0
'''simple docstring''' import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import CLIPImageProcessor, CLIPProcessor @require_vision class _A ( unittest.TestCase ): def lowercase__ ( self : Optional[Any] ) -> str: """simple docstring""" __snake_case : Dict = tempfile.mkdtemp() # fmt: off __snake_case : int = ["l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "lo", "l</w>", "w</w>", "r</w>", "t</w>", "low</w>", "er</w>", "lowest</w>", "newer</w>", "wider", "<unk>", "<|startoftext|>", "<|endoftext|>"] # fmt: on __snake_case : List[str] = dict(zip(__magic_name__ , range(len(__magic_name__ ) ) ) ) __snake_case : Any = ["#version: 0.2", "l o", "lo w</w>", "e r</w>", ""] __snake_case : str = {"unk_token": "<unk>"} __snake_case : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) __snake_case : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__magic_name__ ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(__magic_name__ ) ) __snake_case : Optional[Any] = { "do_resize": True, "size": 20, "do_center_crop": True, "crop_size": 18, "do_normalize": True, "image_mean": [0.48145466, 0.4578275, 0.40821073], "image_std": [0.26862954, 0.26130258, 0.27577711], } __snake_case : List[Any] = os.path.join(self.tmpdirname , __magic_name__ ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__magic_name__ , __magic_name__ ) def lowercase__ ( self : Dict , **__magic_name__ : str ) -> str: """simple docstring""" return CLIPTokenizer.from_pretrained(self.tmpdirname , **__magic_name__ ) def lowercase__ ( self : List[Any] , **__magic_name__ : Optional[int] ) -> Optional[int]: """simple docstring""" return CLIPTokenizerFast.from_pretrained(self.tmpdirname , **__magic_name__ ) def lowercase__ ( self : List[str] , **__magic_name__ : Optional[int] ) -> Optional[int]: """simple docstring""" return CLIPImageProcessor.from_pretrained(self.tmpdirname , **__magic_name__ ) def lowercase__ ( self : Union[str, Any] ) -> str: """simple docstring""" shutil.rmtree(self.tmpdirname ) def lowercase__ ( self : Tuple ) -> Union[str, Any]: """simple docstring""" __snake_case : Any = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta )] __snake_case : Optional[Any] = [Image.fromarray(np.moveaxis(__magic_name__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def lowercase__ ( self : List[str] ) -> List[str]: """simple docstring""" __snake_case : List[str] = self.get_tokenizer() __snake_case : str = self.get_rust_tokenizer() __snake_case : int = self.get_image_processor() __snake_case : Dict = CLIPProcessor(tokenizer=__magic_name__ , image_processor=__magic_name__ ) processor_slow.save_pretrained(self.tmpdirname ) __snake_case : Optional[int] = CLIPProcessor.from_pretrained(self.tmpdirname , use_fast=__magic_name__ ) __snake_case : Dict = CLIPProcessor(tokenizer=__magic_name__ , image_processor=__magic_name__ ) processor_fast.save_pretrained(self.tmpdirname ) __snake_case : Optional[Any] = CLIPProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , __magic_name__ ) self.assertIsInstance(processor_fast.tokenizer , __magic_name__ ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , __magic_name__ ) self.assertIsInstance(processor_fast.image_processor , __magic_name__ ) def lowercase__ ( self : int ) -> Optional[Any]: """simple docstring""" __snake_case : str = CLIPProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : List[str] = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) __snake_case : str = self.get_image_processor(do_normalize=__magic_name__ , padding_value=1.0 ) __snake_case : List[str] = CLIPProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__magic_name__ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , __magic_name__ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __magic_name__ ) def lowercase__ ( self : Dict ) -> str: """simple docstring""" __snake_case : List[Any] = self.get_image_processor() __snake_case : Any = self.get_tokenizer() __snake_case : List[str] = CLIPProcessor(tokenizer=__magic_name__ , image_processor=__magic_name__ ) __snake_case : Any = self.prepare_image_inputs() __snake_case : Tuple = image_processor(__magic_name__ , return_tensors="""np""" ) __snake_case : str = processor(images=__magic_name__ , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1E-2 ) def lowercase__ ( self : Dict ) -> Any: """simple docstring""" __snake_case : int = self.get_image_processor() __snake_case : List[Any] = self.get_tokenizer() __snake_case : List[str] = CLIPProcessor(tokenizer=__magic_name__ , image_processor=__magic_name__ ) __snake_case : Union[str, Any] = "lower newer" __snake_case : Any = processor(text=__magic_name__ ) __snake_case : Dict = tokenizer(__magic_name__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def lowercase__ ( self : List[str] ) -> List[str]: """simple docstring""" __snake_case : List[Any] = self.get_image_processor() __snake_case : int = self.get_tokenizer() __snake_case : Optional[Any] = CLIPProcessor(tokenizer=__magic_name__ , image_processor=__magic_name__ ) __snake_case : Tuple = "lower newer" __snake_case : Dict = self.prepare_image_inputs() __snake_case : List[str] = processor(text=__magic_name__ , images=__magic_name__ ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(__magic_name__ ): processor() def lowercase__ ( self : Optional[int] ) -> Tuple: """simple docstring""" __snake_case : List[str] = self.get_image_processor() __snake_case : List[Any] = self.get_tokenizer() __snake_case : int = CLIPProcessor(tokenizer=__magic_name__ , image_processor=__magic_name__ ) __snake_case : Union[str, Any] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __snake_case : Dict = processor.batch_decode(__magic_name__ ) __snake_case : Dict = tokenizer.batch_decode(__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) def lowercase__ ( self : Tuple ) -> Any: """simple docstring""" __snake_case : Any = self.get_image_processor() __snake_case : List[str] = self.get_tokenizer() __snake_case : Tuple = CLIPProcessor(tokenizer=__magic_name__ , image_processor=__magic_name__ ) __snake_case : Optional[Any] = "lower newer" __snake_case : List[Any] = self.prepare_image_inputs() __snake_case : Any = processor(text=__magic_name__ , images=__magic_name__ ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
26
from ....configuration_utils import PretrainedConfig from ....utils import logging _snake_case : int = logging.get_logger(__name__) _snake_case : str = { "Visual-Attention-Network/van-base": ( "https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json" ), } class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[Any] = "van" def __init__( self : Optional[int] , lowerCamelCase : Any=224 , lowerCamelCase : str=3 , lowerCamelCase : Any=[7, 3, 3, 3] , lowerCamelCase : Dict=[4, 2, 2, 2] , lowerCamelCase : List[Any]=[64, 128, 320, 512] , lowerCamelCase : str=[3, 3, 12, 3] , lowerCamelCase : Dict=[8, 8, 4, 4] , lowerCamelCase : Any="gelu" , lowerCamelCase : Optional[int]=0.02 , lowerCamelCase : Tuple=1E-6 , lowerCamelCase : Optional[int]=1E-2 , lowerCamelCase : int=0.0 , lowerCamelCase : Optional[Any]=0.0 , **lowerCamelCase : Optional[int] , ) -> int: super().__init__(**lowerCamelCase ) __snake_case : Union[str, Any] = image_size __snake_case : Any = num_channels __snake_case : Any = patch_sizes __snake_case : List[Any] = strides __snake_case : str = hidden_sizes __snake_case : Dict = depths __snake_case : Optional[int] = mlp_ratios __snake_case : Dict = hidden_act __snake_case : Union[str, Any] = initializer_range __snake_case : List[str] = layer_norm_eps __snake_case : Optional[int] = layer_scale_init_value __snake_case : List[Any] = drop_path_rate __snake_case : int = dropout_rate
81
0
'''simple docstring''' import unittest from parameterized import parameterized from transformers import AutoTokenizer, GPTNeoXConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXModel, ) class lowerCAmelCase : def __init__( self : Tuple , __lowercase : int , __lowercase : Optional[Any]=13 , __lowercase : Optional[Any]=7 , __lowercase : Union[str, Any]=True , __lowercase : Optional[Any]=True , __lowercase : Optional[int]=True , __lowercase : List[Any]=True , __lowercase : List[str]=99 , __lowercase : str=64 , __lowercase : Tuple=5 , __lowercase : str=4 , __lowercase : Any=37 , __lowercase : Union[str, Any]="gelu" , __lowercase : str=0.1 , __lowercase : str=0.1 , __lowercase : Any=512 , __lowercase : Tuple=16 , __lowercase : List[str]=2 , __lowercase : Dict=0.0_2 , __lowercase : List[Any]=3 , __lowercase : Any=4 , __lowercase : Optional[int]=None , ): """simple docstring""" __lowercase =parent __lowercase =batch_size __lowercase =seq_length __lowercase =is_training __lowercase =use_input_mask __lowercase =use_token_type_ids __lowercase =use_labels __lowercase =vocab_size __lowercase =hidden_size __lowercase =num_hidden_layers __lowercase =num_attention_heads __lowercase =intermediate_size __lowercase =hidden_act __lowercase =hidden_dropout_prob __lowercase =attention_probs_dropout_prob __lowercase =max_position_embeddings __lowercase =type_vocab_size __lowercase =type_sequence_label_size __lowercase =initializer_range __lowercase =num_labels __lowercase =num_choices __lowercase =scope __lowercase =vocab_size - 1 def snake_case ( self : Tuple ): """simple docstring""" __lowercase =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase =None if self.use_input_mask: __lowercase =random_attention_mask([self.batch_size, self.seq_length] ) __lowercase =None if self.use_labels: __lowercase =ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowercase =self.get_config() return config, input_ids, input_mask, token_labels def snake_case ( self : Optional[Any] ): """simple docstring""" return GPTNeoXConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__lowercase , initializer_range=self.initializer_range , pad_token_id=self.pad_token_id , ) def snake_case ( self : List[str] ): """simple docstring""" __lowercase =self.prepare_config_and_inputs() __lowercase =True return config, input_ids, input_mask, token_labels def snake_case ( self : Optional[Any] , __lowercase : Optional[int] , __lowercase : int , __lowercase : List[str] ): """simple docstring""" __lowercase =GPTNeoXModel(config=__lowercase ) model.to(__lowercase ) model.eval() __lowercase =model(__lowercase , attention_mask=__lowercase ) __lowercase =model(__lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case ( self : str , __lowercase : Tuple , __lowercase : Optional[int] , __lowercase : Optional[int] ): """simple docstring""" __lowercase =True __lowercase =GPTNeoXModel(__lowercase ) model.to(__lowercase ) model.eval() __lowercase =model(__lowercase , attention_mask=__lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case ( self : int , __lowercase : Any , __lowercase : List[Any] , __lowercase : List[str] , __lowercase : Optional[Any] ): """simple docstring""" __lowercase =GPTNeoXForCausalLM(config=__lowercase ) model.to(__lowercase ) model.eval() __lowercase =model(__lowercase , attention_mask=__lowercase , labels=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case ( self : Optional[int] , __lowercase : int , __lowercase : Union[str, Any] , __lowercase : Any , __lowercase : Union[str, Any] ): """simple docstring""" __lowercase =self.num_labels __lowercase =GPTNeoXForQuestionAnswering(__lowercase ) model.to(__lowercase ) model.eval() __lowercase =model(__lowercase , attention_mask=__lowercase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def snake_case ( self : Optional[Any] , __lowercase : int , __lowercase : Tuple , __lowercase : Dict , __lowercase : Dict ): """simple docstring""" __lowercase =self.num_labels __lowercase =GPTNeoXForSequenceClassification(__lowercase ) model.to(__lowercase ) model.eval() __lowercase =ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase =model(__lowercase , attention_mask=__lowercase , labels=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case ( self : Optional[Any] , __lowercase : str , __lowercase : Any , __lowercase : int , __lowercase : Any ): """simple docstring""" __lowercase =self.num_labels __lowercase =GPTNeoXForTokenClassification(__lowercase ) model.to(__lowercase ) model.eval() __lowercase =model(__lowercase , attention_mask=__lowercase , labels=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def snake_case ( self : Optional[Any] , __lowercase : Tuple , __lowercase : List[Any] , __lowercase : Dict ): """simple docstring""" __lowercase =True __lowercase =GPTNeoXForCausalLM(config=__lowercase ) model.to(__lowercase ) model.eval() # first forward pass __lowercase =model(__lowercase , attention_mask=__lowercase , use_cache=__lowercase ) __lowercase =outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids __lowercase =ids_tensor((self.batch_size, 3) , config.vocab_size ) __lowercase =ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and __lowercase =torch.cat([input_ids, next_tokens] , dim=-1 ) __lowercase =torch.cat([input_mask, next_mask] , dim=-1 ) __lowercase =model(__lowercase , attention_mask=__lowercase , output_hidden_states=__lowercase ) __lowercase =output_from_no_past["hidden_states"][0] __lowercase =model( __lowercase , attention_mask=__lowercase , past_key_values=__lowercase , output_hidden_states=__lowercase , )["hidden_states"][0] # select random slice __lowercase =ids_tensor((1,) , output_from_past.shape[-1] ).item() __lowercase =output_from_no_past[:, -3:, random_slice_idx].detach() __lowercase =output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__lowercase , __lowercase , atol=1E-3 ) ) def snake_case ( self : List[Any] ): """simple docstring""" __lowercase =self.prepare_config_and_inputs() __lowercase =config_and_inputs __lowercase ={"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class lowerCAmelCase ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): lowerCAmelCase_ = ( ( GPTNeoXModel, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, ) if is_torch_available() else () ) lowerCAmelCase_ = (GPTNeoXForCausalLM,) if is_torch_available() else () lowerCAmelCase_ = ( { "feature-extraction": GPTNeoXModel, "question-answering": GPTNeoXForQuestionAnswering, "text-classification": GPTNeoXForSequenceClassification, "text-generation": GPTNeoXForCausalLM, "token-classification": GPTNeoXForTokenClassification, "zero-shot": GPTNeoXForSequenceClassification, } if is_torch_available() else {} ) lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False def snake_case ( self : Optional[int] ): """simple docstring""" __lowercase =GPTNeoXModelTester(self ) __lowercase =ConfigTester(self , config_class=__lowercase , hidden_size=64 , num_attention_heads=8 ) def snake_case ( self : str ): """simple docstring""" self.config_tester.run_common_tests() def snake_case ( self : Optional[Any] ): """simple docstring""" __lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(__lowercase , __lowercase , __lowercase ) def snake_case ( self : Any ): """simple docstring""" __lowercase =self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(__lowercase , __lowercase , __lowercase ) def snake_case ( self : str ): """simple docstring""" __lowercase =self.model_tester.prepare_config_and_inputs_for_decoder() __lowercase =None self.model_tester.create_and_check_model_as_decoder(__lowercase , __lowercase , __lowercase ) def snake_case ( self : Tuple ): """simple docstring""" __lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(__lowercase , __lowercase , __lowercase ) def snake_case ( self : str ): """simple docstring""" __lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*__lowercase ) def snake_case ( self : str ): """simple docstring""" __lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__lowercase ) def snake_case ( self : List[Any] ): """simple docstring""" __lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__lowercase ) def snake_case ( self : Optional[Any] ): """simple docstring""" __lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__lowercase ) @unittest.skip(reason='Feed forward chunking is not implemented' ) def snake_case ( self : List[Any] ): """simple docstring""" pass @parameterized.expand([('linear',), ('dynamic',)] ) def snake_case ( self : str , __lowercase : Optional[int] ): """simple docstring""" __lowercase =self.model_tester.prepare_config_and_inputs_for_common() __lowercase =ids_tensor([1, 10] , config.vocab_size ) __lowercase =ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights __lowercase =GPTNeoXModel(__lowercase ) original_model.to(__lowercase ) original_model.eval() __lowercase =original_model(__lowercase ).last_hidden_state __lowercase =original_model(__lowercase ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights __lowercase ={"type": scaling_type, "factor": 1_0.0} __lowercase =GPTNeoXModel(__lowercase ) scaled_model.to(__lowercase ) scaled_model.eval() __lowercase =scaled_model(__lowercase ).last_hidden_state __lowercase =scaled_model(__lowercase ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(__lowercase , __lowercase , atol=1E-5 ) ) else: self.assertFalse(torch.allclose(__lowercase , __lowercase , atol=1E-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(__lowercase , __lowercase , atol=1E-5 ) ) @require_torch class lowerCAmelCase ( unittest.TestCase ): @slow def snake_case ( self : Union[str, Any] ): """simple docstring""" __lowercase =AutoTokenizer.from_pretrained('EleutherAI/pythia-410m-deduped' ) for checkpointing in [True, False]: __lowercase =GPTNeoXForCausalLM.from_pretrained('EleutherAI/pythia-410m-deduped' ) if checkpointing: model.gradient_checkpointing_enable() else: model.gradient_checkpointing_disable() model.to(__lowercase ) __lowercase =tokenizer('My favorite food is' , return_tensors='pt' ).to(__lowercase ) # The hub repo. is updated on 2023-04-04, resulting in poor outputs. # See: https://github.com/huggingface/transformers/pull/24193 __lowercase ="My favorite food is a good old-fashioned, old-fashioned, old-fashioned.\n\nI'm not sure" __lowercase =model.generate(**__lowercase , do_sample=__lowercase , max_new_tokens=20 ) __lowercase =tokenizer.batch_decode(__lowercase )[0] self.assertEqual(__lowercase , __lowercase )
119
import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer _snake_case : Union[str, Any] = logging.getLogger(__name__) def lowerCAmelCase_ ( ): __snake_case : int = argparse.ArgumentParser( description="Prepare TFRecord shards from pre-tokenized samples of the wikitext dataset." ) parser.add_argument( "--dataset_name" , type=__lowerCamelCase , default="wikitext" , help="Name of the training. Explore datasets at: hf.co/datasets." , ) parser.add_argument( "--dataset_config" , type=__lowerCamelCase , default="wikitext-103-raw-v1" , help="Configuration name of the dataset." ) parser.add_argument( "--tokenizer_name_or_path" , type=__lowerCamelCase , default="sayakpaul/unigram-tokenizer-wikitext" , help="Tokenizer identifier. Can be a local filepath or a Hub identifier." , ) parser.add_argument( "--shard_size" , type=__lowerCamelCase , default=1_0_0_0 , help="Number of entries to go in a single shard." , ) parser.add_argument("--split" , type=__lowerCamelCase , default="train" , choices=["train", "test", "validation"] ) parser.add_argument( "--limit" , default=__lowerCamelCase , type=__lowerCamelCase , help="Limit the number of shards (used for debugging)." , ) parser.add_argument( "--max_length" , type=__lowerCamelCase , default=5_1_2 , help="Maximum sequence length. For training on TPUs, it helps to have a maximum" " sequence length that is a multiple of 8." , ) parser.add_argument( "--output_dir" , default="tf-tpu" , type=__lowerCamelCase , help="Output directory where the TFRecord shards will be saved. If the" " path is appended with `gs://` ('gs://tf-tpu', for example) then the TFRecord" " shards will be directly saved to a Google Cloud Storage bucket." , ) __snake_case : List[str] = parser.parse_args() return args def lowerCAmelCase_ ( __lowerCamelCase ): def fn(__lowerCamelCase ): return tokenizer(examples["text"] ) return fn def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = [] for i in range(len(tokenized_data["input_ids"] ) ): __snake_case : Tuple = { "input_ids": tf.train.Feature(intaa_list=tf.train.IntaaList(value=tokenized_data["input_ids"][i] ) ), "attention_mask": tf.train.Feature( intaa_list=tf.train.IntaaList(value=tokenized_data["attention_mask"][i] ) ), } __snake_case : List[Any] = tf.train.Features(feature=__lowerCamelCase ) __snake_case : str = tf.train.Example(features=__lowerCamelCase ) __snake_case : List[str] = example.SerializeToString() records.append(__lowerCamelCase ) return records def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Optional[int] = datasets.load_dataset(args.dataset_name , args.dataset_config , split=args.split ) if args.limit is not None: __snake_case : Optional[Any] = min(len(__lowerCamelCase ) , args.limit ) __snake_case : Dict = dataset.select(range(__lowerCamelCase ) ) print(F'Limiting the dataset to {args.limit} entries.' ) __snake_case : Dict = AutoTokenizer.from_pretrained(args.tokenizer_name_or_path ) # Handle output directory creation. # For serializing into a Google Cloud Storage Bucket, one needs to first # create a bucket. if "gs" not in args.output_dir: if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) __snake_case : Dict = os.path.join(args.output_dir , args.split ) if not os.path.exists(__lowerCamelCase ): os.makedirs(__lowerCamelCase ) else: __snake_case : str = os.path.join(args.output_dir , args.split ) # Tokenize the whole dataset at once. __snake_case : Any = tokenize_function(__lowerCamelCase ) __snake_case : Optional[Any] = dataset.map(__lowerCamelCase , batched=__lowerCamelCase , num_proc=4 , remove_columns=["text"] ) # We need to concatenate all our texts together, and then split the result # into chunks of a fixed size, which we will call block_size. To do this, we # will use the map method again, with the option batched=True. When we use batched=True, # the function we pass to map() will be passed multiple inputs at once, allowing us # to group them into more or fewer examples than we had in the input. # This allows us to create our new fixed-length samples. The advantage of this # method is that we don't lose a whole lot of content from the dataset compared to the # case where we simply tokenize with a pre-defined max_length. def group_texts(__lowerCamelCase ): # Concatenate all texts. __snake_case : List[str] = {k: sum(examples[k] , [] ) for k in examples.keys()} __snake_case : List[Any] = len(concatenated_examples[list(examples.keys() )[0]] ) # We drop the small remainder, though you could add padding instead if the model supports it # In this, as in all things, we advise you to follow your heart 🫀 __snake_case : Any = (total_length // args.max_length) * args.max_length # Split by chunks of max_len. __snake_case : int = { k: [t[i : i + args.max_length] for i in range(0 , __lowerCamelCase , args.max_length )] for k, t in concatenated_examples.items() } return result __snake_case : Any = dataset_tokenized.map(__lowerCamelCase , batched=__lowerCamelCase , batch_size=1_0_0_0 , num_proc=4 ) __snake_case : Optional[Any] = 0 __snake_case : Optional[Any] = 0 for shard in range(0 , len(__lowerCamelCase ) , args.shard_size ): __snake_case : List[str] = grouped_dataset[shard : shard + args.shard_size] __snake_case : Any = len(dataset_snapshot["input_ids"] ) __snake_case : List[Any] = os.path.join(__lowerCamelCase , F'dataset-{shard_count}-{records_containing}.tfrecord' ) __snake_case : Optional[Any] = get_serialized_examples(__lowerCamelCase ) with tf.io.TFRecordWriter(__lowerCamelCase ) as out_file: for i in range(len(__lowerCamelCase ) ): __snake_case : Union[str, Any] = serialized_examples[i] out_file.write(__lowerCamelCase ) print("Wrote file {} containing {} records".format(__lowerCamelCase , __lowerCamelCase ) ) shard_count += 1 total_records += records_containing with open(F'split-{args.split}-records-count.txt' , "w" ) as f: print(F'Total {args.split} records: {total_records}' , file=__lowerCamelCase ) if __name__ == "__main__": _snake_case : List[Any] = parse_args() main(args)
81
0
import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def __magic_name__ ( __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : str ) -> Tuple: __lowerCamelCase = f'''{sampling_rate}''' __lowerCamelCase = "1" __lowerCamelCase = "f32le" __lowerCamelCase = [ "ffmpeg", "-i", "pipe:0", "-ac", ac, "-ar", ar, "-f", format_for_conversion, "-hide_banner", "-loglevel", "quiet", "pipe:1", ] try: with subprocess.Popen(__lowerCamelCase , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: __lowerCamelCase = ffmpeg_process.communicate(__lowerCamelCase ) except FileNotFoundError as error: raise ValueError('''ffmpeg was not found but is required to load audio files from filename''' ) from error __lowerCamelCase = output_stream[0] __lowerCamelCase = np.frombuffer(__lowerCamelCase , np.floataa ) if audio.shape[0] == 0: raise ValueError('''Malformed soundfile''' ) return audio def __magic_name__ ( __lowerCAmelCase : Tuple , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : List[str] = "f32le" , ) -> Dict: __lowerCamelCase = f'''{sampling_rate}''' __lowerCamelCase = "1" if format_for_conversion == "s16le": __lowerCamelCase = 2 elif format_for_conversion == "f32le": __lowerCamelCase = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) __lowerCamelCase = platform.system() if system == "Linux": __lowerCamelCase = "alsa" __lowerCamelCase = "default" elif system == "Darwin": __lowerCamelCase = "avfoundation" __lowerCamelCase = ":0" elif system == "Windows": __lowerCamelCase = "dshow" __lowerCamelCase = "default" __lowerCamelCase = [ "ffmpeg", "-f", format_, "-i", input_, "-ac", ac, "-ar", ar, "-f", format_for_conversion, "-fflags", "nobuffer", "-hide_banner", "-loglevel", "quiet", "pipe:1", ] __lowerCamelCase = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample __lowerCamelCase = _ffmpeg_stream(__lowerCamelCase , __lowerCamelCase ) for item in iterator: yield item def __magic_name__ ( __lowerCAmelCase : Any , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : Tuple = None , __lowerCAmelCase : str = None , __lowerCAmelCase : Optional[int] = "f32le" , ) -> Any: if stream_chunk_s is not None: __lowerCamelCase = stream_chunk_s else: __lowerCamelCase = chunk_length_s __lowerCamelCase = ffmpeg_microphone(__lowerCamelCase , __lowerCamelCase , format_for_conversion=__lowerCamelCase ) if format_for_conversion == "s16le": __lowerCamelCase = np.intaa __lowerCamelCase = 2 elif format_for_conversion == "f32le": __lowerCamelCase = np.floataa __lowerCamelCase = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) if stride_length_s is None: __lowerCamelCase = chunk_length_s / 6 __lowerCamelCase = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(__lowerCamelCase , (int, float) ): __lowerCamelCase = [stride_length_s, stride_length_s] __lowerCamelCase = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample __lowerCamelCase = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample __lowerCamelCase = datetime.datetime.now() __lowerCamelCase = datetime.timedelta(seconds=__lowerCamelCase ) for item in chunk_bytes_iter(__lowerCamelCase , __lowerCamelCase , stride=(stride_left, stride_right) , stream=__lowerCamelCase ): # Put everything back in numpy scale __lowerCamelCase = np.frombuffer(item['''raw'''] , dtype=__lowerCamelCase ) __lowerCamelCase = ( item["stride"][0] // size_of_sample, item["stride"][1] // size_of_sample, ) __lowerCamelCase = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 10 * delta: # We're late !! SKIP continue yield item def __magic_name__ ( __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : int , __lowerCAmelCase : List[Any] = False ) -> Optional[Any]: __lowerCamelCase = b"" __lowerCamelCase = stride if stride_left + stride_right >= chunk_len: raise ValueError( f'''Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}''' ) __lowerCamelCase = 0 for raw in iterator: acc += raw if stream and len(__lowerCamelCase ) < chunk_len: __lowerCamelCase = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(__lowerCamelCase ) >= chunk_len: # We are flushing the accumulator __lowerCamelCase = (_stride_left, stride_right) __lowerCamelCase = {"raw": acc[:chunk_len], "stride": stride} if stream: __lowerCamelCase = False yield item __lowerCamelCase = stride_left __lowerCamelCase = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(__lowerCamelCase ) > stride_left: __lowerCamelCase = {"raw": acc, "stride": (_stride_left, 0)} if stream: __lowerCamelCase = False yield item def __magic_name__ ( __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : Dict ) -> List[str]: __lowerCamelCase = 2**24 # 16Mo try: with subprocess.Popen(__lowerCamelCase , stdout=subprocess.PIPE , bufsize=__lowerCamelCase ) as ffmpeg_process: while True: __lowerCamelCase = ffmpeg_process.stdout.read(__lowerCamelCase ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError('''ffmpeg was not found but is required to stream audio files from filename''' ) from error
298
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) _snake_case : List[Any] = "\\n Text data.\n Second line of data." _snake_case : Tuple = "file" @pytest.fixture(scope="session" ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / (FILE_PATH + ".zstd") __snake_case : Optional[Any] = bytes(__lowerCamelCase , "utf-8" ) with zstd.open(__lowerCamelCase , "wb" ) as f: f.write(__lowerCamelCase ) return path @pytest.fixture def lowerCAmelCase_ ( __lowerCamelCase ): with open(os.path.join(tmpfs.local_root_dir , __lowerCamelCase ) , "w" ) as f: f.write(__lowerCamelCase ) return FILE_PATH @pytest.mark.parametrize("compression_format" , ["gzip", "xz", "zstd"] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Optional[int] = {"gzip": gz_file, "xz": xz_file, "zstd": zstd_path} __snake_case : str = input_paths[compression_format] __snake_case : Optional[Any] = tmp_path / "cache" __snake_case : Optional[int] = DownloadConfig(cache_dir=__lowerCamelCase , extract_compressed_file=__lowerCamelCase ) __snake_case : Union[str, Any] = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) with open(__lowerCamelCase ) as f: __snake_case : Dict = f.read() with open(__lowerCamelCase ) as f: __snake_case : Tuple = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("default_extracted" , [True, False] ) @pytest.mark.parametrize("default_cache_dir" , [True, False] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Tuple = "custom_cache" __snake_case : List[str] = "custom_extracted_dir" __snake_case : Any = tmp_path / "custom_extracted_path" if default_extracted: __snake_case : List[Any] = ("downloads" if default_cache_dir else custom_cache_dir, "extracted") else: monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_DIR" , __lowerCamelCase ) monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_PATH" , str(__lowerCamelCase ) ) __snake_case : Optional[Any] = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) __snake_case : Optional[int] = xz_file __snake_case : Optional[int] = ( DownloadConfig(extract_compressed_file=__lowerCamelCase ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=__lowerCamelCase ) ) __snake_case : str = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) assert Path(__lowerCamelCase ).parent.parts[-2:] == expected def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : Optional[Any] = str(Path(__lowerCamelCase ).resolve() ) assert cached_path(__lowerCamelCase ) == text_file # relative path __snake_case : Any = str(Path(__lowerCamelCase ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(__lowerCamelCase ) == text_file def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : List[Any] = str(tmp_path.resolve() / "__missing_file__.txt" ) with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) # relative path __snake_case : Optional[int] = "./__missing_file__.txt" with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : str = get_from_cache(F'tmp://{tmpfs_file}' ) with open(__lowerCamelCase ) as f: __snake_case : Union[str, Any] = f.read() assert output_file_content == FILE_CONTENT @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( ): with pytest.raises(__lowerCamelCase ): cached_path("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[Any] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): http_get("https://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): http_head("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[str] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): ftp_get("ftp://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): ftp_head("ftp://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): fsspec_get("s3://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): fsspec_head("s3://huggingface.co" )
81
0
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer __SCREAMING_SNAKE_CASE =logging.get_logger(__name__) __SCREAMING_SNAKE_CASE ={"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} __SCREAMING_SNAKE_CASE ={ "vocab_file": { "distilbert-base-uncased": "https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt", "distilbert-base-uncased-distilled-squad": ( "https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt" ), "distilbert-base-cased": "https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt", "distilbert-base-cased-distilled-squad": ( "https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt" ), "distilbert-base-german-cased": "https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt", "distilbert-base-multilingual-cased": ( "https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt" ), }, "tokenizer_file": { "distilbert-base-uncased": "https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json", "distilbert-base-uncased-distilled-squad": ( "https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json" ), "distilbert-base-cased": "https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json", "distilbert-base-cased-distilled-squad": ( "https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json" ), "distilbert-base-german-cased": ( "https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json" ), "distilbert-base-multilingual-cased": ( "https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json" ), }, } __SCREAMING_SNAKE_CASE ={ "distilbert-base-uncased": 512, "distilbert-base-uncased-distilled-squad": 512, "distilbert-base-cased": 512, "distilbert-base-cased-distilled-squad": 512, "distilbert-base-german-cased": 512, "distilbert-base-multilingual-cased": 512, } __SCREAMING_SNAKE_CASE ={ "distilbert-base-uncased": {"do_lower_case": True}, "distilbert-base-uncased-distilled-squad": {"do_lower_case": True}, "distilbert-base-cased": {"do_lower_case": False}, "distilbert-base-cased-distilled-squad": {"do_lower_case": False}, "distilbert-base-german-cased": {"do_lower_case": False}, "distilbert-base-multilingual-cased": {"do_lower_case": False}, } class UpperCamelCase ( _lowerCAmelCase ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = PRETRAINED_INIT_CONFIGURATION lowercase = ["input_ids", "attention_mask"] lowercase = DistilBertTokenizer def __init__( self ,__UpperCamelCase=None ,__UpperCamelCase=None ,__UpperCamelCase=True ,__UpperCamelCase="[UNK]" ,__UpperCamelCase="[SEP]" ,__UpperCamelCase="[PAD]" ,__UpperCamelCase="[CLS]" ,__UpperCamelCase="[MASK]" ,__UpperCamelCase=True ,__UpperCamelCase=None ,**__UpperCamelCase ,) -> List[Any]: '''simple docstring''' super().__init__( __UpperCamelCase ,tokenizer_file=__UpperCamelCase ,do_lower_case=__UpperCamelCase ,unk_token=__UpperCamelCase ,sep_token=__UpperCamelCase ,pad_token=__UpperCamelCase ,cls_token=__UpperCamelCase ,mask_token=__UpperCamelCase ,tokenize_chinese_chars=__UpperCamelCase ,strip_accents=__UpperCamelCase ,**__UpperCamelCase ,) lowercase_ : Dict = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('lowercase' ,__UpperCamelCase ) != do_lower_case or normalizer_state.get('strip_accents' ,__UpperCamelCase ) != strip_accents or normalizer_state.get('handle_chinese_chars' ,__UpperCamelCase ) != tokenize_chinese_chars ): lowercase_ : List[Any] = getattr(__UpperCamelCase ,normalizer_state.pop('type' ) ) lowercase_ : Union[str, Any] = do_lower_case lowercase_ : Union[str, Any] = strip_accents lowercase_ : List[str] = tokenize_chinese_chars lowercase_ : Optional[Any] = normalizer_class(**__UpperCamelCase ) lowercase_ : List[Any] = do_lower_case def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase=None ) -> Any: '''simple docstring''' lowercase_ : Tuple = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase = None ) -> List[int]: '''simple docstring''' lowercase_ : Union[str, Any] = [self.sep_token_id] lowercase_ : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase = None ) -> Tuple[str]: '''simple docstring''' lowercase_ : str = self._tokenizer.model.save(__UpperCamelCase ,name=__UpperCamelCase ) return tuple(__UpperCamelCase )
425
_snake_case : Optional[int] = {"a": ["c", "b"], "b": ["d", "e"], "c": [], "d": [], "e": []} _snake_case : Dict = ["a", "b", "c", "d", "e"] def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : List[str] = start # add current to visited visited.append(__lowerCamelCase ) __snake_case : List[Any] = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: __snake_case : Tuple = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # if all neighbors visited add current to sort sort.append(__lowerCamelCase ) # if all vertices haven't been visited select a new one to visit if len(__lowerCamelCase ) != len(__lowerCamelCase ): for vertice in vertices: if vertice not in visited: __snake_case : int = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # return sort return sort if __name__ == "__main__": _snake_case : List[Any] = topological_sort("a", [], []) print(sort)
81
0
'''simple docstring''' def __snake_case ( _UpperCAmelCase : Dict): UpperCamelCase = len(__lowerCamelCase) while cur > 1: # Find the maximum number in arr UpperCamelCase = arr.index(max(arr[0:cur])) # Reverse from 0 to mi UpperCamelCase = arr[mi::-1] + arr[mi + 1 : len(__lowerCamelCase)] # Reverse whole list UpperCamelCase = arr[cur - 1 :: -1] + arr[cur : len(__lowerCamelCase)] cur -= 1 return arr if __name__ == "__main__": snake_case_ : Tuple = input('Enter numbers separated by a comma:\n').strip() snake_case_ : List[Any] = [int(item) for item in user_input.split(',')] print(pancake_sort(unsorted))
212
def lowerCAmelCase_ ( ): return [ a * b * (1_0_0_0 - a - b) for a in range(1 , 9_9_9 ) for b in range(__lowerCamelCase , 9_9_9 ) if (a * a + b * b == (1_0_0_0 - a - b) ** 2) ][0] if __name__ == "__main__": print(f'''{solution() = }''')
81
0
def __snake_case ( _lowerCAmelCase : List[str] ) -> List[Any]: if not nums: # Makes sure that the list is not empty raise ValueError("List is empty" ) A_ : str = sum(__lowerCamelCase ) / len(__lowerCamelCase ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(__lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
454
from math import sqrt def lowerCAmelCase_ ( __lowerCamelCase = 1_0_0_0_0_0_0 ): __snake_case : int = 0 __snake_case : int = 0 __snake_case : int while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(__lowerCamelCase , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(f'''{solution() = }''')
81
0
"""simple docstring""" # Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from packaging import version from .. import __version__ from .constants import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD from .doc import ( add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, copy_func, replace_return_docstrings, ) from .generic import ( ContextManagers, ExplicitEnum, ModelOutput, PaddingStrategy, TensorType, add_model_info_to_auto_map, cached_property, can_return_loss, expand_dims, find_labels, flatten_dict, infer_framework, is_jax_tensor, is_numpy_array, is_tensor, is_tf_symbolic_tensor, is_tf_tensor, is_torch_device, is_torch_dtype, is_torch_tensor, reshape, squeeze, strtobool, tensor_size, to_numpy, to_py_obj, transpose, working_or_temp_dir, ) from .hub import ( CLOUDFRONT_DISTRIB_PREFIX, DISABLE_TELEMETRY, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, EntryNotFoundError, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, cached_file, default_cache_path, define_sagemaker_information, download_url, extract_commit_hash, get_cached_models, get_file_from_repo, get_full_repo_name, has_file, http_user_agent, is_offline_mode, is_remote_url, move_cache, send_example_telemetry, try_to_load_from_cache, ) from .import_utils import ( ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, TORCH_FX_REQUIRED_VERSION, USE_JAX, USE_TF, USE_TORCH, DummyObject, OptionalDependencyNotAvailable, _LazyModule, ccl_version, direct_transformers_import, get_torch_version, is_accelerate_available, is_apex_available, is_bitsandbytes_available, is_bsa_available, is_coloredlogs_available, is_cython_available, is_datasets_available, is_decord_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_jieba_available, is_jumanpp_available, is_kenlm_available, is_keras_nlp_available, is_librosa_available, is_natten_available, is_ninja_available, is_onnx_available, is_openai_available, is_optimum_available, is_pandas_available, is_peft_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytest_available, is_pytorch_quantization_available, is_rjieba_available, is_sacremoses_available, is_safetensors_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_sudachi_available, is_tensorflow_probability_available, is_tensorflow_text_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_bfaa_cpu_available, is_torch_bfaa_gpu_available, is_torch_compile_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_neuroncore_available, is_torch_tensorrt_fx_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_torchdistx_available, is_torchdynamo_available, is_torchvision_available, is_training_run_on_sagemaker, is_vision_available, requires_backends, torch_only_method, ) lowerCamelCase__ = "pytorch_model.bin" lowerCamelCase__ = "pytorch_model.bin.index.json" lowerCamelCase__ = "adapter_config.json" lowerCamelCase__ = "adapter_model.bin" lowerCamelCase__ = "adapter_model.safetensors" lowerCamelCase__ = "tf_model.h5" lowerCamelCase__ = "tf_model.h5.index.json" lowerCamelCase__ = "model.ckpt" lowerCamelCase__ = "flax_model.msgpack" lowerCamelCase__ = "flax_model.msgpack.index.json" lowerCamelCase__ = "model.safetensors" lowerCamelCase__ = "model.safetensors.index.json" lowerCamelCase__ = "config.json" lowerCamelCase__ = "preprocessor_config.json" lowerCamelCase__ = FEATURE_EXTRACTOR_NAME lowerCamelCase__ = "generation_config.json" lowerCamelCase__ = "modelcard.json" lowerCamelCase__ = "▁" lowerCamelCase__ = SENTENCEPIECE_UNDERLINE # Kept for backward compatibility lowerCamelCase__ = [ [[0, 1, 0, 1], [1, 0, 0, 1]] ] * 2 # Needs to have 0s and 1s only since XLM uses it for langs too. lowerCamelCase__ = [[7, 6, 0, 0, 1], [1, 2, 3, 0, 0], [0, 0, 0, 4, 5]] lowerCamelCase__ = [[1, 1, 1, 1, 1], [1, 1, 1, 0, 0], [0, 0, 0, 1, 1]] def _SCREAMING_SNAKE_CASE ( UpperCamelCase : Tuple ): if version.parse(__lowerCamelCase ) < version.parse(__lowerCamelCase ): if "dev" in min_version: A__ = ( "This example requires a source install from HuggingFace Transformers (see " "`https://huggingface.co/docs/transformers/installation#install-from-source`)," ) else: A__ = F"""This example requires a minimum version of {min_version},""" error_message += F""" but the version found is {__version__}.\n""" raise ImportError( error_message + """Check out https://github.com/huggingface/transformers/tree/main/examples#important-note for the examples corresponding to other """ """versions of HuggingFace Transformers.""" )
574
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class a (_lowerCAmelCase ): """simple docstring""" def __snake_case ( self : str ) -> str: __snake_case : Tuple = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCamelCase , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "neck_hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "num_attention_heads" ) ) class a : """simple docstring""" def __init__( self : Optional[int] , lowerCamelCase : List[str] , lowerCamelCase : Tuple=13 , lowerCamelCase : str=32 , lowerCamelCase : Dict=2 , lowerCamelCase : List[str]=3 , lowerCamelCase : Any=640 , lowerCamelCase : Optional[Any]=4 , lowerCamelCase : Tuple="silu" , lowerCamelCase : int=3 , lowerCamelCase : Dict=32 , lowerCamelCase : str=0.1 , lowerCamelCase : Optional[int]=0.1 , lowerCamelCase : Optional[Any]=0.1 , lowerCamelCase : Dict=0.02 , lowerCamelCase : Union[str, Any]=True , lowerCamelCase : Optional[int]=True , lowerCamelCase : Union[str, Any]=10 , lowerCamelCase : int=None , ) -> str: __snake_case : Optional[Any] = parent __snake_case : Optional[Any] = batch_size __snake_case : Any = image_size __snake_case : List[Any] = patch_size __snake_case : Any = num_channels __snake_case : Union[str, Any] = last_hidden_size __snake_case : Any = num_attention_heads __snake_case : Any = hidden_act __snake_case : Tuple = conv_kernel_size __snake_case : Any = output_stride __snake_case : Any = hidden_dropout_prob __snake_case : List[Any] = attention_probs_dropout_prob __snake_case : Optional[Any] = classifier_dropout_prob __snake_case : Union[str, Any] = use_labels __snake_case : Optional[int] = is_training __snake_case : Dict = num_labels __snake_case : Any = initializer_range __snake_case : Optional[int] = scope def __snake_case ( self : str ) -> Union[str, Any]: __snake_case : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : List[Any] = None __snake_case : Optional[int] = None if self.use_labels: __snake_case : Optional[int] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : Any = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __snake_case : Optional[Any] = self.get_config() return config, pixel_values, labels, pixel_labels def __snake_case ( self : Any ) -> Union[str, Any]: return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __snake_case ( self : Any , lowerCamelCase : Tuple , lowerCamelCase : int , lowerCamelCase : Tuple , lowerCamelCase : Optional[int] ) -> Dict: __snake_case : List[Any] = MobileViTModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[str] = model(lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[Any] , lowerCamelCase : List[str] , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] , lowerCamelCase : Tuple ) -> List[str]: __snake_case : str = self.num_labels __snake_case : List[Any] = MobileViTForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : Dict , lowerCamelCase : Any , lowerCamelCase : Any , lowerCamelCase : Optional[Any] , lowerCamelCase : Dict ) -> Dict: __snake_case : Union[str, Any] = self.num_labels __snake_case : Optional[int] = MobileViTForSemanticSegmentation(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Tuple = model(lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[int] ) -> List[Any]: __snake_case : Optional[Any] = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case , __snake_case : Union[str, Any] = config_and_inputs __snake_case : Dict = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) __UpperCAmelCase : Optional[Any] = ( { "feature-extraction": MobileViTModel, "image-classification": MobileViTForImageClassification, "image-segmentation": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) __UpperCAmelCase : List[str] = False __UpperCAmelCase : int = False __UpperCAmelCase : Optional[int] = False __UpperCAmelCase : Optional[int] = False def __snake_case ( self : Optional[int] ) -> Dict: __snake_case : Tuple = MobileViTModelTester(self ) __snake_case : Any = MobileViTConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase ) def __snake_case ( self : Optional[int] ) -> Dict: self.config_tester.run_common_tests() @unittest.skip(reason="MobileViT does not use inputs_embeds" ) def __snake_case ( self : Dict ) -> Any: pass @unittest.skip(reason="MobileViT does not support input and output embeddings" ) def __snake_case ( self : Dict ) -> List[Any]: pass @unittest.skip(reason="MobileViT does not output attentions" ) def __snake_case ( self : int ) -> Dict: pass def __snake_case ( self : int ) -> Union[str, Any]: __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Optional[Any] = model_class(lowerCamelCase ) __snake_case : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : int = [*signature.parameters.keys()] __snake_case : List[str] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __snake_case ( self : int ) -> Tuple: pass def __snake_case ( self : Any ) -> Tuple: __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def __snake_case ( self : Any ) -> str: def check_hidden_states_output(lowerCamelCase : Union[str, Any] , lowerCamelCase : Dict , lowerCamelCase : Any ): __snake_case : int = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : int = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : Union[str, Any] = outputs.hidden_states __snake_case : int = 5 self.assertEqual(len(lowerCamelCase ) , lowerCamelCase ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. __snake_case : List[Any] = 2 for i in range(len(lowerCamelCase ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) __snake_case , __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : str = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : List[Any] = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Any ) -> Any: __snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) def __snake_case ( self : List[str] ) -> List[str]: __snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowerCamelCase ) @slow def __snake_case ( self : List[str] ) -> Any: for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : List[str] = MobileViTModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def lowerCAmelCase_ ( ): __snake_case : Optional[int] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a (unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self : str ) -> Dict: return MobileViTImageProcessor.from_pretrained("apple/mobilevit-xx-small" ) if is_vision_available() else None @slow def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : Tuple = MobileViTForImageClassification.from_pretrained("apple/mobilevit-xx-small" ).to(lowerCamelCase ) __snake_case : Optional[Any] = self.default_image_processor __snake_case : Union[str, Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Dict = model(**lowerCamelCase ) # verify the logits __snake_case : Union[str, Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([-1.93_64, -1.23_27, -0.46_53] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : str ) -> Optional[int]: __snake_case : Optional[int] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : str = model.to(lowerCamelCase ) __snake_case : int = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Optional[int] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : List[str] = model(**lowerCamelCase ) __snake_case : Union[str, Any] = outputs.logits # verify the logits __snake_case : Tuple = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , lowerCamelCase ) __snake_case : Union[str, Any] = torch.tensor( [ [[6.97_13, 6.97_86, 7.24_22], [7.28_93, 7.28_25, 7.44_46], [7.65_80, 7.87_97, 7.94_20]], [[-10.68_69, -10.32_50, -10.34_71], [-10.42_28, -9.98_68, -9.71_32], [-11.04_05, -11.02_21, -10.73_18]], [[-3.30_89, -2.85_39, -2.67_40], [-3.27_06, -2.56_21, -2.51_08], [-3.25_34, -2.66_15, -2.66_51]], ] , device=lowerCamelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : Union[str, Any] ) -> Optional[int]: __snake_case : Optional[Any] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Tuple = model.to(lowerCamelCase ) __snake_case : Dict = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : List[Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Any = model(**lowerCamelCase ) __snake_case : Dict = outputs.logits.detach().cpu() __snake_case : Any = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase , target_sizes=[(50, 60)] ) __snake_case : int = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , lowerCamelCase ) __snake_case : List[str] = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase ) __snake_case : Optional[int] = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , lowerCamelCase )
81
0
'''simple docstring''' import argparse import collections import numpy as np import torch from flax import traverse_util from tax import checkpoints from transformers import MTaConfig, UMTaEncoderModel, UMTaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): return params[F"""{prefix}/{prefix}/relpos_bias/rel_embedding"""][:, i, :] def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase="attention" ): lowercase__ : Tuple = np.ascontiguousarray(params[F"""{prefix}/{prefix}/{layer_name}/key/kernel"""][:, i, :, :] ) lowercase__ : Dict = k_tmp.reshape(k_tmp.shape[0] , k_tmp.shape[1] * k_tmp.shape[2] ) lowercase__ : str = np.ascontiguousarray(params[F"""{prefix}/{prefix}/{layer_name}/out/kernel"""][:, i, :, :] ) lowercase__ : List[str] = o_tmp.reshape(o_tmp.shape[0] * o_tmp.shape[1] , o_tmp.shape[2] ) lowercase__ : Any = np.ascontiguousarray(params[F"""{prefix}/{prefix}/{layer_name}/query/kernel"""][:, i, :, :] ) lowercase__ : Optional[int] = q_tmp.reshape(q_tmp.shape[0] , q_tmp.shape[1] * q_tmp.shape[2] ) lowercase__ : List[str] = np.ascontiguousarray(params[F"""{prefix}/{prefix}/{layer_name}/value/kernel"""][:, i, :, :] ) lowercase__ : List[Any] = v_tmp.reshape(v_tmp.shape[0] , v_tmp.shape[1] * v_tmp.shape[2] ) return k, o, q, v def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=False ): if split_mlp_wi: lowercase__ : Any = params[F"""{prefix}/{prefix}/mlp/wi_0/kernel"""][:, i, :] lowercase__ : List[Any] = params[F"""{prefix}/{prefix}/mlp/wi_1/kernel"""][:, i, :] lowercase__ : Optional[Any] = (wi_a, wi_a) else: lowercase__ : Optional[int] = params[F"""{prefix}/{prefix}/mlp/wi/kernel"""][:, i, :] lowercase__ : List[Any] = params[F"""{prefix}/{prefix}/mlp/wo/kernel"""][:, i, :] return wi, wo def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): return params[F"""{prefix}/{prefix}/{layer_name}/scale"""][:, i] def __UpperCamelCase ( UpperCAmelCase , *, UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = False ): lowercase__ : Union[str, Any] = traverse_util.flatten_dict(variables['''target'''] ) lowercase__ : str = {"/".join(__lowerCamelCase ): v for k, v in old.items()} # v1.1 models have a gated GeLU with wi_0 and wi_1 instead of wi lowercase__ : Any = "encoder/encoder/mlp/wi_0/kernel" in old print('''Split MLP:''' , __lowerCamelCase ) lowercase__ : Tuple = collections.OrderedDict() # Shared embeddings. lowercase__ : Optional[int] = old["token_embedder/embedding"] # Encoder. for i in range(__lowerCamelCase ): # Block i, layer 0 (Self Attention). lowercase__ : Union[str, Any] = tax_layer_norm_lookup(__lowerCamelCase , __lowerCamelCase , '''encoder''' , '''pre_attention_layer_norm''' ) lowercase__ : List[str] = tax_attention_lookup(__lowerCamelCase , __lowerCamelCase , '''encoder''' , '''attention''' ) lowercase__ : List[str] = layer_norm lowercase__ : Union[str, Any] = k.T lowercase__ : List[str] = o.T lowercase__ : List[str] = q.T lowercase__ : Union[str, Any] = v.T # Block i, layer 1 (MLP). lowercase__ : str = tax_layer_norm_lookup(__lowerCamelCase , __lowerCamelCase , '''encoder''' , '''pre_mlp_layer_norm''' ) lowercase__ : Any = tax_mlp_lookup(__lowerCamelCase , __lowerCamelCase , '''encoder''' , __lowerCamelCase ) lowercase__ : str = layer_norm if split_mlp_wi: lowercase__ : Optional[int] = wi[0].T lowercase__ : List[Any] = wi[1].T else: lowercase__ : Any = wi.T lowercase__ : Union[str, Any] = wo.T if scalable_attention: # convert the rel_embedding of each layer lowercase__ : List[Any] = tax_relpos_bias_lookup( __lowerCamelCase , __lowerCamelCase , '''encoder''' ).T lowercase__ : int = old["encoder/encoder_norm/scale"] if not scalable_attention: lowercase__ : Optional[int] = tax_relpos_bias_lookup( __lowerCamelCase , 0 , '''encoder''' ).T lowercase__ : Union[str, Any] = tax_relpos_bias_lookup( __lowerCamelCase , 0 , '''decoder''' ).T if not is_encoder_only: # Decoder. for i in range(__lowerCamelCase ): # Block i, layer 0 (Self Attention). lowercase__ : Tuple = tax_layer_norm_lookup(__lowerCamelCase , __lowerCamelCase , '''decoder''' , '''pre_self_attention_layer_norm''' ) lowercase__ : Tuple = tax_attention_lookup(__lowerCamelCase , __lowerCamelCase , '''decoder''' , '''self_attention''' ) lowercase__ : int = layer_norm lowercase__ : Tuple = k.T lowercase__ : List[Any] = o.T lowercase__ : str = q.T lowercase__ : Optional[Any] = v.T # Block i, layer 1 (Cross Attention). lowercase__ : Optional[Any] = tax_layer_norm_lookup(__lowerCamelCase , __lowerCamelCase , '''decoder''' , '''pre_cross_attention_layer_norm''' ) lowercase__ : List[Any] = tax_attention_lookup(__lowerCamelCase , __lowerCamelCase , '''decoder''' , '''encoder_decoder_attention''' ) lowercase__ : Optional[int] = layer_norm lowercase__ : int = k.T lowercase__ : Optional[Any] = o.T lowercase__ : Dict = q.T lowercase__ : List[Any] = v.T # Block i, layer 2 (MLP). lowercase__ : Dict = tax_layer_norm_lookup(__lowerCamelCase , __lowerCamelCase , '''decoder''' , '''pre_mlp_layer_norm''' ) lowercase__ : Dict = tax_mlp_lookup(__lowerCamelCase , __lowerCamelCase , '''decoder''' , __lowerCamelCase ) lowercase__ : Dict = layer_norm if split_mlp_wi: lowercase__ : List[Any] = wi[0].T lowercase__ : int = wi[1].T else: lowercase__ : Dict = wi.T lowercase__ : Any = wo.T if scalable_attention: # convert the rel_embedding of each layer lowercase__ : Any = tax_relpos_bias_lookup(__lowerCamelCase , __lowerCamelCase , '''decoder''' ).T lowercase__ : Dict = old["decoder/decoder_norm/scale"] # LM Head (only in v1.1 checkpoints, in v1.0 embeddings are used instead) if "decoder/logits_dense/kernel" in old: lowercase__ : List[Any] = old["decoder/logits_dense/kernel"].T return new def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase ): lowercase__ : Optional[Any] = collections.OrderedDict([(k, torch.from_numpy(v.copy() )) for (k, v) in converted_params.items()] ) # Add what is missing. if "encoder.embed_tokens.weight" not in state_dict: lowercase__ : int = state_dict["shared.weight"] if not is_encoder_only: if "decoder.embed_tokens.weight" not in state_dict: lowercase__ : Optional[int] = state_dict["shared.weight"] if "lm_head.weight" not in state_dict: # For old 1.0 models. print('''Using shared word embeddings as lm_head.''' ) lowercase__ : int = state_dict["shared.weight"] return state_dict def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): lowercase__ : Union[str, Any] = checkpoints.load_tax_checkpoint(__lowerCamelCase ) lowercase__ : int = convert_tax_to_pytorch( __lowerCamelCase , num_layers=config.num_layers , is_encoder_only=__lowerCamelCase , scalable_attention=__lowerCamelCase ) lowercase__ : List[str] = make_state_dict(__lowerCamelCase , __lowerCamelCase ) model.load_state_dict(__lowerCamelCase , strict=__lowerCamelCase ) def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = False , UpperCAmelCase = False , ): lowercase__ : str = MTaConfig.from_json_file(__lowerCamelCase ) print(F"""Building PyTorch model from configuration: {config}""" ) # Non-v1.1 checkpoints could also use T5Model, but this works for all. # The v1.0 checkpoints will simply have an LM head that is the word embeddings. if is_encoder_only: lowercase__ : Optional[int] = UMTaEncoderModel(__lowerCamelCase ) else: lowercase__ : int = UMTaForConditionalGeneration(__lowerCamelCase ) # Load weights from tf checkpoint load_tax_weights_in_ta(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # Save pytorch-model print(F"""Save PyTorch model to {pytorch_dump_path}""" ) model.save_pretrained(__lowerCamelCase ) # Verify that we can load the checkpoint. model.from_pretrained(__lowerCamelCase ) print('''Done''' ) if __name__ == "__main__": __a: Optional[Any] = argparse.ArgumentParser(description="""Converts a native T5X checkpoint into a PyTorch checkpoint.""") # Required parameters parser.add_argument( """--t5x_checkpoint_path""", default=None, type=str, required=True, help="""Path to the T5X checkpoint.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help="""The config json file corresponding to the pre-trained T5 model.\nThis specifies the model architecture.""", ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--is_encoder_only""", action="""store_true""", help="""Check if the model is encoder-decoder model""", default=False ) parser.add_argument( """--scalable_attention""", action="""store_true""", help="""Whether the model uses scaled attention (umt5 model)""", default=False, ) __a: Optional[Any] = parser.parse_args() convert_tax_checkpoint_to_pytorch( args.tax_checkpoint_path, args.config_file, args.pytorch_dump_path, args.is_encoder_only, args.scalable_attention, )
152
import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor _snake_case : int = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[int] , *lowerCamelCase : List[Any] , **lowerCamelCase : int ) -> None: warnings.warn( "The class ImageGPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use ImageGPTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
import string def A_ ( lowercase_ ) -> int: for key in range(len(string.ascii_uppercase ) ): _snake_case : Optional[int] = "" for symbol in message: if symbol in string.ascii_uppercase: _snake_case : Dict = string.ascii_uppercase.find(__lowerCamelCase ) _snake_case : Any = num - key if num < 0: _snake_case : Optional[Any] = num + len(string.ascii_uppercase ) _snake_case : List[Any] = translated + string.ascii_uppercase[num] else: _snake_case : Dict = translated + symbol print(f'''Decryption using Key #{key}: {translated}''' ) def A_ ( ) -> Dict: _snake_case : Tuple = input('''Encrypted message: ''' ) _snake_case : int = message.upper() decrypt(__lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() main()
326
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _snake_case : Optional[Any] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[str] = ["pixel_values"] def __init__( self : List[Any] , lowerCamelCase : bool = True , lowerCamelCase : Union[int, float] = 1 / 255 , lowerCamelCase : bool = True , lowerCamelCase : int = 8 , **lowerCamelCase : Tuple , ) -> None: super().__init__(**lowerCamelCase ) __snake_case : Dict = do_rescale __snake_case : Dict = rescale_factor __snake_case : Optional[Any] = do_pad __snake_case : Tuple = pad_size def __snake_case ( self : Dict , lowerCamelCase : np.ndarray , lowerCamelCase : float , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase : Optional[int] ) -> np.ndarray: return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def __snake_case ( self : Optional[Any] , lowerCamelCase : np.ndarray , lowerCamelCase : int , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None ) -> Tuple: __snake_case , __snake_case : List[str] = get_image_size(lowerCamelCase ) __snake_case : Optional[Any] = (old_height // size + 1) * size - old_height __snake_case : List[Any] = (old_width // size + 1) * size - old_width return pad(lowerCamelCase , ((0, pad_height), (0, pad_width)) , mode="symmetric" , data_format=lowerCamelCase ) def __snake_case ( self : Tuple , lowerCamelCase : ImageInput , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[float] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[int] = None , lowerCamelCase : Optional[Union[str, TensorType]] = None , lowerCamelCase : Union[str, ChannelDimension] = ChannelDimension.FIRST , **lowerCamelCase : Union[str, Any] , ) -> List[str]: __snake_case : int = do_rescale if do_rescale is not None else self.do_rescale __snake_case : Dict = rescale_factor if rescale_factor is not None else self.rescale_factor __snake_case : str = do_pad if do_pad is not None else self.do_pad __snake_case : Any = pad_size if pad_size is not None else self.pad_size __snake_case : int = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) # All transformations expect numpy arrays. __snake_case : str = [to_numpy_array(lowerCamelCase ) for image in images] if do_rescale: __snake_case : Optional[int] = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_pad: __snake_case : Optional[Any] = [self.pad(lowerCamelCase , size=lowerCamelCase ) for image in images] __snake_case : int = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __snake_case : Union[str, Any] = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
81
0
"""simple docstring""" import inspect import unittest from transformers import SegformerConfig, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_MAPPING, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerModel, ) from transformers.models.segformer.modeling_segformer import SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import SegformerImageProcessor class SCREAMING_SNAKE_CASE_ ( _lowerCAmelCase ): """simple docstring""" def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.config_class(**self.inputs_dict) self.parent.assertTrue(hasattr(lowerCAmelCase__ , """hidden_sizes""")) self.parent.assertTrue(hasattr(lowerCAmelCase__ , """num_attention_heads""")) self.parent.assertTrue(hasattr(lowerCAmelCase__ , """num_encoder_blocks""")) class SCREAMING_SNAKE_CASE_ : """simple docstring""" def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=1_3 , lowerCAmelCase__=6_4 , lowerCAmelCase__=3 , lowerCAmelCase__=4 , lowerCAmelCase__=[2, 2, 2, 2] , lowerCAmelCase__=[8, 4, 2, 1] , lowerCAmelCase__=[1_6, 3_2, 6_4, 1_2_8] , lowerCAmelCase__=[1, 4, 8, 1_6] , lowerCAmelCase__=[1, 2, 4, 8] , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.02 , lowerCAmelCase__=3 , lowerCAmelCase__=None , ): __SCREAMING_SNAKE_CASE = parent __SCREAMING_SNAKE_CASE = batch_size __SCREAMING_SNAKE_CASE = image_size __SCREAMING_SNAKE_CASE = num_channels __SCREAMING_SNAKE_CASE = num_encoder_blocks __SCREAMING_SNAKE_CASE = sr_ratios __SCREAMING_SNAKE_CASE = depths __SCREAMING_SNAKE_CASE = hidden_sizes __SCREAMING_SNAKE_CASE = downsampling_rates __SCREAMING_SNAKE_CASE = num_attention_heads __SCREAMING_SNAKE_CASE = is_training __SCREAMING_SNAKE_CASE = use_labels __SCREAMING_SNAKE_CASE = hidden_act __SCREAMING_SNAKE_CASE = hidden_dropout_prob __SCREAMING_SNAKE_CASE = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE = initializer_range __SCREAMING_SNAKE_CASE = num_labels __SCREAMING_SNAKE_CASE = scope def snake_case_ ( self): __SCREAMING_SNAKE_CASE = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) __SCREAMING_SNAKE_CASE = None if self.use_labels: __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels) __SCREAMING_SNAKE_CASE = self.get_config() return config, pixel_values, labels def snake_case_ ( self): return SegformerConfig( image_size=self.image_size , num_channels=self.num_channels , num_encoder_blocks=self.num_encoder_blocks , depths=self.depths , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , ) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = SegformerModel(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = self.image_size // (self.downsampling_rates[-1] * 2) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], expected_height, expected_width)) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = self.num_labels __SCREAMING_SNAKE_CASE = SegformerForSemanticSegmentation(lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4)) __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , labels=lowerCAmelCase__) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4)) self.parent.assertGreater(result.loss , 0.0) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = 1 __SCREAMING_SNAKE_CASE = SegformerForSemanticSegmentation(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = torch.randint(0 , 1 , (self.batch_size, self.image_size, self.image_size)).to(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , labels=lowerCAmelCase__) self.parent.assertGreater(result.loss , 0.0) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() __SCREAMING_SNAKE_CASE = config_and_inputs __SCREAMING_SNAKE_CASE = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE_ ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __lowercase : Any = ( ( SegformerModel, SegformerForSemanticSegmentation, SegformerForImageClassification, ) if is_torch_available() else () ) __lowercase : Tuple = ( { "feature-extraction": SegformerModel, "image-classification": SegformerForImageClassification, "image-segmentation": SegformerForSemanticSegmentation, } if is_torch_available() else {} ) __lowercase : Dict = True __lowercase : Optional[Any] = False __lowercase : Optional[int] = False __lowercase : Any = False def snake_case_ ( self): __SCREAMING_SNAKE_CASE = SegformerModelTester(self) __SCREAMING_SNAKE_CASE = SegformerConfigTester(self , config_class=lowerCAmelCase__) def snake_case_ ( self): self.config_tester.run_common_tests() def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_binary_image_segmentation(*lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_segmentation(*lowerCAmelCase__) @unittest.skip("""SegFormer does not use inputs_embeds""") def snake_case_ ( self): pass @unittest.skip("""SegFormer does not have get_input_embeddings method and get_output_embeddings methods""") def snake_case_ ( self): pass def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __SCREAMING_SNAKE_CASE = model_class(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic __SCREAMING_SNAKE_CASE = [*signature.parameters.keys()] __SCREAMING_SNAKE_CASE = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() __SCREAMING_SNAKE_CASE = True for model_class in self.all_model_classes: __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = model_class(lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(**self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__)) __SCREAMING_SNAKE_CASE = outputs.attentions __SCREAMING_SNAKE_CASE = sum(self.model_tester.depths) self.assertEqual(len(lowerCAmelCase__) , lowerCAmelCase__) # check that output_attentions also work using config del inputs_dict["output_attentions"] __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = model_class(lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(**self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__)) __SCREAMING_SNAKE_CASE = outputs.attentions self.assertEqual(len(lowerCAmelCase__) , lowerCAmelCase__) # verify the first attentions (first block, first layer) __SCREAMING_SNAKE_CASE = (self.model_tester.image_size // 4) ** 2 __SCREAMING_SNAKE_CASE = (self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(attentions[0].shape[-3:]) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) # verify the last attentions (last block, last layer) __SCREAMING_SNAKE_CASE = (self.model_tester.image_size // 3_2) ** 2 __SCREAMING_SNAKE_CASE = (self.model_tester.image_size // (3_2 * self.model_tester.sr_ratios[-1])) ** 2 self.assertListEqual( list(attentions[-1].shape[-3:]) , [self.model_tester.num_attention_heads[-1], expected_seq_len, expected_reduced_seq_len] , ) __SCREAMING_SNAKE_CASE = len(lowerCAmelCase__) # Check attention is always last and order is fine __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = model_class(lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(**self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__)) self.assertEqual(out_len + 1 , len(lowerCAmelCase__)) __SCREAMING_SNAKE_CASE = outputs.attentions self.assertEqual(len(lowerCAmelCase__) , lowerCAmelCase__) # verify the first attentions (first block, first layer) __SCREAMING_SNAKE_CASE = (self.model_tester.image_size // 4) ** 2 __SCREAMING_SNAKE_CASE = (self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(self_attentions[0].shape[-3:]) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) def snake_case_ ( self): def check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = model_class(lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(**self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__)) __SCREAMING_SNAKE_CASE = outputs.hidden_states __SCREAMING_SNAKE_CASE = self.model_tester.num_encoder_blocks self.assertEqual(len(lowerCAmelCase__) , lowerCAmelCase__) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:]) , [ self.model_tester.hidden_sizes[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __SCREAMING_SNAKE_CASE = True check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __SCREAMING_SNAKE_CASE = True check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__) def snake_case_ ( self): if not self.model_tester.is_training: return __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() __SCREAMING_SNAKE_CASE = True for model_class in self.all_model_classes: if model_class in get_values(lowerCAmelCase__): continue __SCREAMING_SNAKE_CASE = model_class(lowerCAmelCase__) model.to(lowerCAmelCase__) model.train() __SCREAMING_SNAKE_CASE = self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ , return_labels=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = model(**lowerCAmelCase__).loss loss.backward() @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""") def snake_case_ ( self): pass @slow def snake_case_ ( self): for model_name in SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __SCREAMING_SNAKE_CASE = SegformerModel.from_pretrained(lowerCAmelCase__) self.assertIsNotNone(lowerCAmelCase__) def _lowerCAmelCase ( ): __SCREAMING_SNAKE_CASE = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): """simple docstring""" @slow def snake_case_ ( self): # only resize + normalize __SCREAMING_SNAKE_CASE = SegformerImageProcessor( image_scale=(5_1_2, 5_1_2) , keep_ratio=lowerCAmelCase__ , align=lowerCAmelCase__ , do_random_crop=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = SegformerForSemanticSegmentation.from_pretrained("""nvidia/segformer-b0-finetuned-ade-512-512""").to( lowerCAmelCase__) __SCREAMING_SNAKE_CASE = prepare_img() __SCREAMING_SNAKE_CASE = image_processor(images=lowerCAmelCase__ , return_tensors="""pt""") __SCREAMING_SNAKE_CASE = encoded_inputs.pixel_values.to(lowerCAmelCase__) with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = torch.Size((1, model.config.num_labels, 1_2_8, 1_2_8)) self.assertEqual(outputs.logits.shape , lowerCAmelCase__) __SCREAMING_SNAKE_CASE = torch.tensor( [ [[-4.63_10, -5.52_32, -6.23_56], [-5.19_21, -6.14_44, -6.59_96], [-5.44_24, -6.27_90, -6.75_74]], [[-12.13_91, -13.31_22, -13.95_54], [-12.87_32, -13.93_52, -14.35_63], [-12.94_38, -13.82_26, -14.25_13]], [[-12.51_34, -13.46_86, -14.49_15], [-12.86_69, -14.43_43, -14.77_58], [-13.25_23, -14.58_19, -15.06_94]], ]).to(lowerCAmelCase__) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , lowerCAmelCase__ , atol=1E-4)) @slow def snake_case_ ( self): # only resize + normalize __SCREAMING_SNAKE_CASE = SegformerImageProcessor( image_scale=(5_1_2, 5_1_2) , keep_ratio=lowerCAmelCase__ , align=lowerCAmelCase__ , do_random_crop=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = SegformerForSemanticSegmentation.from_pretrained( """nvidia/segformer-b1-finetuned-cityscapes-1024-1024""").to(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = prepare_img() __SCREAMING_SNAKE_CASE = image_processor(images=lowerCAmelCase__ , return_tensors="""pt""") __SCREAMING_SNAKE_CASE = encoded_inputs.pixel_values.to(lowerCAmelCase__) with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = torch.Size((1, model.config.num_labels, 1_2_8, 1_2_8)) self.assertEqual(outputs.logits.shape , lowerCAmelCase__) __SCREAMING_SNAKE_CASE = torch.tensor( [ [[-13.57_48, -13.91_11, -12.65_00], [-14.35_00, -15.36_83, -14.23_28], [-14.75_32, -16.04_24, -15.60_87]], [[-17.16_51, -15.87_25, -12.96_53], [-17.25_80, -17.37_18, -14.82_23], [-16.60_58, -16.87_83, -16.74_52]], [[-3.64_56, -3.02_09, -1.42_03], [-3.07_97, -3.19_59, -2.00_00], [-1.87_57, -1.92_17, -1.69_97]], ]).to(lowerCAmelCase__) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , lowerCAmelCase__ , atol=1E-1)) @slow def snake_case_ ( self): # only resize + normalize __SCREAMING_SNAKE_CASE = SegformerImageProcessor( image_scale=(5_1_2, 5_1_2) , keep_ratio=lowerCAmelCase__ , align=lowerCAmelCase__ , do_random_crop=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = SegformerForSemanticSegmentation.from_pretrained("""nvidia/segformer-b0-finetuned-ade-512-512""").to( lowerCAmelCase__) __SCREAMING_SNAKE_CASE = prepare_img() __SCREAMING_SNAKE_CASE = image_processor(images=lowerCAmelCase__ , return_tensors="""pt""") __SCREAMING_SNAKE_CASE = encoded_inputs.pixel_values.to(lowerCAmelCase__) with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = outputs.logits.detach().cpu() __SCREAMING_SNAKE_CASE = image_processor.post_process_semantic_segmentation(outputs=lowerCAmelCase__ , target_sizes=[(5_0_0, 3_0_0)]) __SCREAMING_SNAKE_CASE = torch.Size((5_0_0, 3_0_0)) self.assertEqual(segmentation[0].shape , lowerCAmelCase__) __SCREAMING_SNAKE_CASE = image_processor.post_process_semantic_segmentation(outputs=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = torch.Size((1_2_8, 1_2_8)) self.assertEqual(segmentation[0].shape , lowerCAmelCase__)
155
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case : Any = { "configuration_informer": [ "INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "InformerConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "InformerForPrediction", "InformerModel", "InformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys _snake_case : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
0
import inspect import unittest from transformers import BitConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import BitBackbone, BitForImageClassification, BitImageProcessor, BitModel from transformers.models.bit.modeling_bit import BIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class lowerCAmelCase_ : def __init__( self : Optional[Any] , SCREAMING_SNAKE_CASE_ : Tuple , SCREAMING_SNAKE_CASE_ : str=3 , SCREAMING_SNAKE_CASE_ : Tuple=32 , SCREAMING_SNAKE_CASE_ : Tuple=3 , SCREAMING_SNAKE_CASE_ : Any=10 , SCREAMING_SNAKE_CASE_ : int=[8, 16, 32, 64] , SCREAMING_SNAKE_CASE_ : Optional[int]=[1, 1, 2, 1] , SCREAMING_SNAKE_CASE_ : int=True , SCREAMING_SNAKE_CASE_ : Optional[Any]=True , SCREAMING_SNAKE_CASE_ : Any="relu" , SCREAMING_SNAKE_CASE_ : Tuple=3 , SCREAMING_SNAKE_CASE_ : str=None , SCREAMING_SNAKE_CASE_ : Optional[Any]=["stage2", "stage3", "stage4"] , SCREAMING_SNAKE_CASE_ : Union[str, Any]=[2, 3, 4] , SCREAMING_SNAKE_CASE_ : str=1 , ): lowerCAmelCase__ = parent lowerCAmelCase__ = batch_size lowerCAmelCase__ = image_size lowerCAmelCase__ = num_channels lowerCAmelCase__ = embeddings_size lowerCAmelCase__ = hidden_sizes lowerCAmelCase__ = depths lowerCAmelCase__ = is_training lowerCAmelCase__ = use_labels lowerCAmelCase__ = hidden_act lowerCAmelCase__ = num_labels lowerCAmelCase__ = scope lowerCAmelCase__ = len(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ = out_features lowerCAmelCase__ = out_indices lowerCAmelCase__ = num_groups def __snake_case ( self : Optional[Any] ): lowerCAmelCase__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase__ = None if self.use_labels: lowerCAmelCase__ = ids_tensor([self.batch_size] , self.num_labels ) lowerCAmelCase__ = self.get_config() return config, pixel_values, labels def __snake_case ( self : int ): return BitConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , out_features=self.out_features , out_indices=self.out_indices , num_groups=self.num_groups , ) def __snake_case ( self : Optional[Any] , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : List[Any] ): lowerCAmelCase__ = BitModel(config=SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def __snake_case ( self : List[str] , SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : Any , SCREAMING_SNAKE_CASE_ : Optional[int] ): lowerCAmelCase__ = self.num_labels lowerCAmelCase__ = BitForImageClassification(SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE_ , labels=SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : Dict , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : Optional[Any] , SCREAMING_SNAKE_CASE_ : int ): lowerCAmelCase__ = BitBackbone(config=SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE_ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None lowerCAmelCase__ = None lowerCAmelCase__ = BitBackbone(config=SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() lowerCAmelCase__ = model(SCREAMING_SNAKE_CASE_ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def __snake_case ( self : Union[str, Any] ): lowerCAmelCase__ = self.prepare_config_and_inputs() lowerCAmelCase__ = config_and_inputs lowerCAmelCase__ = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class lowerCAmelCase_ ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): UpperCamelCase_ :str = (BitModel, BitForImageClassification, BitBackbone) if is_torch_available() else () UpperCamelCase_ :List[str] = ( {"feature-extraction": BitModel, "image-classification": BitForImageClassification} if is_torch_available() else {} ) UpperCamelCase_ :int = False UpperCamelCase_ :Tuple = False UpperCamelCase_ :Any = False UpperCamelCase_ :List[str] = False UpperCamelCase_ :Optional[int] = False def __snake_case ( self : int ): lowerCAmelCase__ = BitModelTester(self ) lowerCAmelCase__ = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE_ , has_text_modality=SCREAMING_SNAKE_CASE_ ) def __snake_case ( self : str ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __snake_case ( self : Optional[int] ): return @unittest.skip(reason='''Bit does not output attentions''' ) def __snake_case ( self : int ): pass @unittest.skip(reason='''Bit does not use inputs_embeds''' ) def __snake_case ( self : int ): pass @unittest.skip(reason='''Bit does not support input and output embeddings''' ) def __snake_case ( self : str ): pass def __snake_case ( self : List[Any] ): lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ = model_class(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ = [*signature.parameters.keys()] lowerCAmelCase__ = ["pixel_values"] self.assertListEqual(arg_names[:1] , SCREAMING_SNAKE_CASE_ ) def __snake_case ( self : str ): lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE_ ) def __snake_case ( self : str ): lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*SCREAMING_SNAKE_CASE_ ) def __snake_case ( self : List[Any] ): lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ = model_class(config=SCREAMING_SNAKE_CASE_ ) for name, module in model.named_modules(): if isinstance(SCREAMING_SNAKE_CASE_ , (nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) , msg=f'Parameter {name} of model {model_class} seems not properly initialized' , ) self.assertTrue( torch.all(module.bias == 0 ) , msg=f'Parameter {name} of model {model_class} seems not properly initialized' , ) def __snake_case ( self : Dict ): def check_hidden_states_output(SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : Any ): lowerCAmelCase__ = model_class(SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() with torch.no_grad(): lowerCAmelCase__ = model(**self._prepare_for_class(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) lowerCAmelCase__ = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states lowerCAmelCase__ = self.model_tester.num_stages self.assertEqual(len(SCREAMING_SNAKE_CASE_ ) , expected_num_stages + 1 ) # Bit's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ = ["preactivation", "bottleneck"] for model_class in self.all_model_classes: for layer_type in layers_type: lowerCAmelCase__ = layer_type lowerCAmelCase__ = True check_hidden_states_output(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ = True check_hidden_states_output(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) @unittest.skip(reason='''Bit does not use feedforward chunking''' ) def __snake_case ( self : str ): pass def __snake_case ( self : Dict ): lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*SCREAMING_SNAKE_CASE_ ) @slow def __snake_case ( self : List[Any] ): for model_name in BIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ = BitModel.from_pretrained(SCREAMING_SNAKE_CASE_ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE_ ) def lowerCAmelCase_ () -> Optional[int]: '''simple docstring''' lowerCAmelCase__ = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowerCAmelCase_ ( unittest.TestCase ): @cached_property def __snake_case ( self : Any ): return ( BitImageProcessor.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def __snake_case ( self : List[Any] ): lowerCAmelCase__ = BitForImageClassification.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ = self.default_image_processor lowerCAmelCase__ = prepare_img() lowerCAmelCase__ = image_processor(images=SCREAMING_SNAKE_CASE_ , return_tensors='''pt''' ).to(SCREAMING_SNAKE_CASE_ ) # forward pass with torch.no_grad(): lowerCAmelCase__ = model(**SCREAMING_SNAKE_CASE_ ) # verify the logits lowerCAmelCase__ = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ = torch.tensor([[-0.6_526, -0.5_263, -1.4_398]] ).to(SCREAMING_SNAKE_CASE_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , SCREAMING_SNAKE_CASE_ , atol=1e-4 ) ) @require_torch class lowerCAmelCase_ ( _lowerCAmelCase , unittest.TestCase ): UpperCamelCase_ :str = (BitBackbone,) if is_torch_available() else () UpperCamelCase_ :str = BitConfig UpperCamelCase_ :List[Any] = False def __snake_case ( self : Union[str, Any] ): lowerCAmelCase__ = BitModelTester(self )
668
from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging _snake_case : Any = logging.get_logger(__name__) # pylint: disable=invalid-name class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[Any] , lowerCamelCase : CLIPSegForImageSegmentation , lowerCamelCase : CLIPSegProcessor , lowerCamelCase : AutoencoderKL , lowerCamelCase : CLIPTextModel , lowerCamelCase : CLIPTokenizer , lowerCamelCase : UNetaDConditionModel , lowerCamelCase : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , lowerCamelCase : StableDiffusionSafetyChecker , lowerCamelCase : CLIPImageProcessor , ) -> Tuple: super().__init__() if hasattr(scheduler.config , "steps_offset" ) and scheduler.config.steps_offset != 1: __snake_case : Tuple = ( F'The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`' F' should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure ' "to update the config accordingly as leaving `steps_offset` might led to incorrect results" " in future versions. If you have downloaded this checkpoint from the Hugging Face Hub," " it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`" " file" ) deprecate("steps_offset!=1" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : Any = dict(scheduler.config ) __snake_case : List[Any] = 1 __snake_case : Tuple = FrozenDict(lowerCamelCase ) if hasattr(scheduler.config , "skip_prk_steps" ) and scheduler.config.skip_prk_steps is False: __snake_case : List[str] = ( F'The configuration file of this scheduler: {scheduler} has not set the configuration' " `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make" " sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to" " incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face" " Hub, it would be very nice if you could open a Pull request for the" " `scheduler/scheduler_config.json` file" ) deprecate("skip_prk_steps not set" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : List[str] = dict(scheduler.config ) __snake_case : List[str] = True __snake_case : Any = FrozenDict(lowerCamelCase ) if safety_checker is None: logger.warning( F'You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure' " that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered" " results in services or applications open to the public. Both the diffusers team and Hugging Face" " strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling" " it only for use-cases that involve analyzing network behavior or auditing its results. For more" " information, please have a look at https://github.com/huggingface/diffusers/pull/254 ." ) self.register_modules( segmentation_model=lowerCamelCase , segmentation_processor=lowerCamelCase , vae=lowerCamelCase , text_encoder=lowerCamelCase , tokenizer=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase , safety_checker=lowerCamelCase , feature_extractor=lowerCamelCase , ) def __snake_case ( self : Dict , lowerCamelCase : Optional[Union[str, int]] = "auto" ) -> Tuple: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __snake_case : Any = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowerCamelCase ) def __snake_case ( self : List[Any] ) -> Any: self.enable_attention_slicing(lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> str: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) __snake_case : Optional[int] = torch.device("cuda" ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(lowerCamelCase , lowerCamelCase ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __snake_case ( self : int ) -> Any: if self.device != torch.device("meta" ) or not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowerCamelCase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self : List[Any] , lowerCamelCase : Union[str, List[str]] , lowerCamelCase : Union[torch.FloatTensor, PIL.Image.Image] , lowerCamelCase : str , lowerCamelCase : int = 512 , lowerCamelCase : int = 512 , lowerCamelCase : int = 50 , lowerCamelCase : float = 7.5 , lowerCamelCase : Optional[Union[str, List[str]]] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : float = 0.0 , lowerCamelCase : Optional[torch.Generator] = None , lowerCamelCase : Optional[torch.FloatTensor] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , lowerCamelCase : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , lowerCamelCase : int = 1 , **lowerCamelCase : Dict , ) -> List[str]: __snake_case : Tuple = self.segmentation_processor( text=[text] , images=[image] , padding="max_length" , return_tensors="pt" ).to(self.device ) __snake_case : str = self.segmentation_model(**lowerCamelCase ) __snake_case : Dict = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() __snake_case : List[Any] = self.numpy_to_pil(lowerCamelCase )[0].resize(image.size ) # Run inpainting pipeline with the generated mask __snake_case : Tuple = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=lowerCamelCase , image=lowerCamelCase , mask_image=lowerCamelCase , height=lowerCamelCase , width=lowerCamelCase , num_inference_steps=lowerCamelCase , guidance_scale=lowerCamelCase , negative_prompt=lowerCamelCase , num_images_per_prompt=lowerCamelCase , eta=lowerCamelCase , generator=lowerCamelCase , latents=lowerCamelCase , output_type=lowerCamelCase , return_dict=lowerCamelCase , callback=lowerCamelCase , callback_steps=lowerCamelCase , )
81
0
'''simple docstring''' # Note: if you intend to run this script make sure you look under scripts/fsmt/ # to locate the appropriate script to do the work correctly. There is a set of scripts to: # - download and prepare data and run the conversion script # - perform eval to get the best hparam into the config # - generate model_cards - useful if you have multiple models from the same paper import argparse import json import os import re from collections import OrderedDict from os.path import basename, dirname import fairseq import torch from fairseq import hub_utils from fairseq.data.dictionary import Dictionary from transformers import FSMTConfig, FSMTForConditionalGeneration from transformers.models.fsmt.tokenization_fsmt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from transformers.utils import WEIGHTS_NAME, logging logging.set_verbosity_warning() __UpperCamelCase = 2 # based on the results of a search on a range of `num_beams`, `length_penalty` and `early_stopping` # values against wmt19 test data to obtain the best BLEU scores, we will use the following defaults: # # * `num_beams`: 5 (higher scores better, but requires more memory/is slower, can be adjusted by users) # * `early_stopping`: `False` consistently scored better # * `length_penalty` varied, so will assign the best one depending on the model __UpperCamelCase = { # fairseq: "wmt19-ru-en": {"length_penalty": 1.1}, "wmt19-en-ru": {"length_penalty": 1.15}, "wmt19-en-de": {"length_penalty": 1.0}, "wmt19-de-en": {"length_penalty": 1.1}, # allenai: "wmt16-en-de-dist-12-1": {"length_penalty": 0.6}, "wmt16-en-de-dist-6-1": {"length_penalty": 0.6}, "wmt16-en-de-12-1": {"length_penalty": 0.8}, "wmt19-de-en-6-6-base": {"length_penalty": 0.6}, "wmt19-de-en-6-6-big": {"length_penalty": 0.6}, } # this remaps the different models to their organization names __UpperCamelCase = {} for m in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: __UpperCamelCase = "facebook" for m in [ "wmt16-en-de-dist-12-1", "wmt16-en-de-dist-6-1", "wmt16-en-de-12-1", "wmt19-de-en-6-6-base", "wmt19-de-en-6-6-big", ]: __UpperCamelCase = "allenai" def _a ( _lowerCamelCase ) -> Optional[Any]: """simple docstring""" __snake_case : List[Any] = dict((re.sub(R"""@@$""" , """""" , __lowerCamelCase ), v) if k.endswith("""@@""" ) else (re.sub(R"""$""" , """</w>""" , __lowerCamelCase ), v) for k, v in d.items() ) __snake_case : Tuple = "<s> <pad> </s> <unk>".split() # restore the special tokens for k in keep_keys: del da[F'''{k}</w>'''] __snake_case : Optional[int] = d[k] # restore return da def _a ( _lowerCamelCase , _lowerCamelCase ) -> List[Any]: """simple docstring""" assert os.path.exists(__lowerCamelCase ) os.makedirs(__lowerCamelCase , exist_ok=__lowerCamelCase ) print(F'''Writing results to {pytorch_dump_folder_path}''' ) # handle various types of models __snake_case : Any = basename(__lowerCamelCase ) __snake_case : Union[str, Any] = dirname(__lowerCamelCase ) __snake_case : int = fairseq.model_parallel.models.transformer.ModelParallelTransformerModel __snake_case : Optional[Any] = cls.hub_models() __snake_case : Union[str, Any] = {"bpe": "fastbpe", "tokenizer": "moses"} __snake_case : Any = "." # note: since the model dump is old, fairseq has upgraded its model some # time later, and it does a whole lot of rewrites and splits on the saved # weights, therefore we can't use torch.load() directly on the model file. # see: upgrade_state_dict(state_dict) in fairseq_model.py print(F'''using checkpoint {checkpoint_file}''' ) __snake_case : Dict = hub_utils.from_pretrained( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , archive_map=__lowerCamelCase , **__lowerCamelCase ) __snake_case : List[Any] = vars(chkpt["""args"""]["""model"""] ) __snake_case : int = args["source_lang"] __snake_case : str = args["target_lang"] __snake_case : List[str] = dirname(__lowerCamelCase ) __snake_case : str = basename(__lowerCamelCase ) # dicts __snake_case : str = os.path.join(__lowerCamelCase , F'''dict.{src_lang}.txt''' ) __snake_case : List[str] = os.path.join(__lowerCamelCase , F'''dict.{tgt_lang}.txt''' ) __snake_case : List[str] = Dictionary.load(__lowerCamelCase ) __snake_case : Any = rewrite_dict_keys(src_dict.indices ) __snake_case : str = len(__lowerCamelCase ) __snake_case : List[str] = os.path.join(__lowerCamelCase , """vocab-src.json""" ) print(F'''Generating {src_vocab_file} of {src_vocab_size} of {src_lang} records''' ) with open(__lowerCamelCase , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(__lowerCamelCase , ensure_ascii=__lowerCamelCase , indent=__lowerCamelCase ) ) # detect whether this is a do_lower_case situation, which can be derived by checking whether we # have at least one uppercase letter in the source vocab __snake_case : Tuple = True for k in src_vocab.keys(): if not k.islower(): __snake_case : Tuple = False break __snake_case : Dict = Dictionary.load(__lowerCamelCase ) __snake_case : List[Any] = rewrite_dict_keys(tgt_dict.indices ) __snake_case : Any = len(__lowerCamelCase ) __snake_case : List[str] = os.path.join(__lowerCamelCase , """vocab-tgt.json""" ) print(F'''Generating {tgt_vocab_file} of {tgt_vocab_size} of {tgt_lang} records''' ) with open(__lowerCamelCase , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(__lowerCamelCase , ensure_ascii=__lowerCamelCase , indent=__lowerCamelCase ) ) # merges_file (bpecodes) __snake_case : Union[str, Any] = os.path.join(__lowerCamelCase , VOCAB_FILES_NAMES["""merges_file"""] ) for fn in ["bpecodes", "code"]: # older fairseq called the merges file "code" __snake_case : Any = os.path.join(__lowerCamelCase , __lowerCamelCase ) if os.path.exists(__lowerCamelCase ): break with open(__lowerCamelCase , encoding="""utf-8""" ) as fin: __snake_case : int = fin.read() __snake_case : Any = re.sub(R""" \d+$""" , """""" , __lowerCamelCase , 0 , re.M ) # remove frequency number print(F'''Generating {merges_file}''' ) with open(__lowerCamelCase , """w""" , encoding="""utf-8""" ) as fout: fout.write(__lowerCamelCase ) # model config __snake_case : str = os.path.join(__lowerCamelCase , """config.json""" ) # validate bpe/tokenizer config, as currently it's hardcoded to moses+fastbpe - # may have to modify the tokenizer if a different type is used by a future model assert args["bpe"] == "fastbpe", F'''need to extend tokenizer to support bpe={args["bpe"]}''' assert args["tokenizer"] == "moses", F'''need to extend tokenizer to support bpe={args["tokenizer"]}''' __snake_case : List[Any] = { "architectures": ["FSMTForConditionalGeneration"], "model_type": "fsmt", "activation_dropout": args["activation_dropout"], "activation_function": "relu", "attention_dropout": args["attention_dropout"], "d_model": args["decoder_embed_dim"], "dropout": args["dropout"], "init_std": 0.02, "max_position_embeddings": args["max_source_positions"], "num_hidden_layers": args["encoder_layers"], "src_vocab_size": src_vocab_size, "tgt_vocab_size": tgt_vocab_size, "langs": [src_lang, tgt_lang], "encoder_attention_heads": args["encoder_attention_heads"], "encoder_ffn_dim": args["encoder_ffn_embed_dim"], "encoder_layerdrop": args["encoder_layerdrop"], "encoder_layers": args["encoder_layers"], "decoder_attention_heads": args["decoder_attention_heads"], "decoder_ffn_dim": args["decoder_ffn_embed_dim"], "decoder_layerdrop": args["decoder_layerdrop"], "decoder_layers": args["decoder_layers"], "bos_token_id": 0, "pad_token_id": 1, "eos_token_id": 2, "is_encoder_decoder": True, "scale_embedding": not args["no_scale_embedding"], "tie_word_embeddings": args["share_all_embeddings"], } # good hparam defaults to start with __snake_case : Optional[Any] = 5 __snake_case : List[Any] = False if model_dir in best_score_hparams and "length_penalty" in best_score_hparams[model_dir]: __snake_case : List[str] = best_score_hparams[model_dir]["length_penalty"] else: __snake_case : Tuple = 1.0 print(F'''Generating {fsmt_model_config_file}''' ) with open(__lowerCamelCase , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(__lowerCamelCase , ensure_ascii=__lowerCamelCase , indent=__lowerCamelCase ) ) # tokenizer config __snake_case : List[str] = os.path.join(__lowerCamelCase , __lowerCamelCase ) __snake_case : Tuple = { "langs": [src_lang, tgt_lang], "model_max_length": 1024, "do_lower_case": do_lower_case, } print(F'''Generating {fsmt_tokenizer_config_file}''' ) with open(__lowerCamelCase , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(__lowerCamelCase , ensure_ascii=__lowerCamelCase , indent=__lowerCamelCase ) ) # model __snake_case : List[Any] = chkpt["models"][0] __snake_case : Tuple = model.state_dict() # rename keys to start with 'model.' __snake_case : Union[str, Any] = OrderedDict(("""model.""" + k, v) for k, v in model_state_dict.items() ) # remove unneeded keys __snake_case : Any = [ "model.model", "model.encoder.version", "model.decoder.version", "model.encoder_embed_tokens.weight", "model.decoder_embed_tokens.weight", "model.encoder.embed_positions._float_tensor", "model.decoder.embed_positions._float_tensor", ] for k in ignore_keys: model_state_dict.pop(__lowerCamelCase , __lowerCamelCase ) __snake_case : Union[str, Any] = FSMTConfig.from_pretrained(__lowerCamelCase ) __snake_case : Union[str, Any] = FSMTForConditionalGeneration(__lowerCamelCase ) # check that it loads ok model_new.load_state_dict(__lowerCamelCase , strict=__lowerCamelCase ) # save __snake_case : int = os.path.join(__lowerCamelCase , __lowerCamelCase ) print(F'''Generating {pytorch_weights_dump_path}''' ) torch.save(__lowerCamelCase , __lowerCamelCase ) print("""Conversion is done!""" ) print("""\nLast step is to upload the files to s3""" ) print(F'''cd {data_root}''' ) print(F'''transformers-cli upload {model_dir}''' ) if __name__ == "__main__": __UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--fsmt_checkpoint_path", default=None, type=str, required=True, help=( "Path to the official PyTorch checkpoint file which is expected to reside in the dump dir with dicts," " bpecodes, etc." ), ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) __UpperCamelCase = parser.parse_args() convert_fsmt_checkpoint_to_pytorch(args.fsmt_checkpoint_path, args.pytorch_dump_folder_path)
26
class a : """simple docstring""" def __init__( self : Tuple , lowerCamelCase : list ) -> None: __snake_case : str = set_counts __snake_case : Union[str, Any] = max(lowerCamelCase ) __snake_case : List[Any] = len(lowerCamelCase ) __snake_case : Tuple = [1] * num_sets __snake_case : Dict = list(range(lowerCamelCase ) ) def __snake_case ( self : str , lowerCamelCase : int , lowerCamelCase : int ) -> bool: __snake_case : List[Any] = self.get_parent(lowerCamelCase ) __snake_case : Tuple = self.get_parent(lowerCamelCase ) if src_parent == dst_parent: return False if self.ranks[dst_parent] >= self.ranks[src_parent]: self.set_counts[dst_parent] += self.set_counts[src_parent] __snake_case : List[str] = 0 __snake_case : List[Any] = dst_parent if self.ranks[dst_parent] == self.ranks[src_parent]: self.ranks[dst_parent] += 1 __snake_case : Dict = self.set_counts[dst_parent] else: self.set_counts[src_parent] += self.set_counts[dst_parent] __snake_case : Union[str, Any] = 0 __snake_case : Optional[int] = src_parent __snake_case : Tuple = self.set_counts[src_parent] __snake_case : str = max(self.max_set , lowerCamelCase ) return True def __snake_case ( self : int , lowerCamelCase : int ) -> int: if self.parents[disj_set] == disj_set: return disj_set __snake_case : Optional[int] = self.get_parent(self.parents[disj_set] ) return self.parents[disj_set]
81
0
'''simple docstring''' from queue import PriorityQueue from typing import Any import numpy as np def __UpperCamelCase ( lowercase__ : str, lowercase__ : Union[str, Any], lowercase__ : str, lowercase__ : int, lowercase__ : Union[str, Any], lowercase__ : Union[str, Any], lowercase__ : List[str], lowercase__ : Tuple, lowercase__ : Dict, ): '''simple docstring''' for nxt, d in graph[v]: if nxt in visited_forward: continue __lowercase =cst_fwd.get(__lowerCamelCase, np.inf ) __lowercase =cst_fwd[v] + d if new_cost_f < old_cost_f: queue.put((new_cost_f, nxt) ) __lowercase =new_cost_f __lowercase =v if nxt in visited_backward: if cst_fwd[v] + d + cst_bwd[nxt] < shortest_distance: __lowercase =cst_fwd[v] + d + cst_bwd[nxt] return shortest_distance def __UpperCamelCase ( lowercase__ : List[Any], lowercase__ : int, lowercase__ : Dict, lowercase__ : Tuple ): '''simple docstring''' __lowercase =-1 __lowercase =set() __lowercase =set() __lowercase ={source: 0} __lowercase ={destination: 0} __lowercase ={source: None} __lowercase ={destination: None} __lowercase =PriorityQueue() __lowercase =PriorityQueue() __lowercase =np.inf queue_forward.put((0, source) ) queue_backward.put((0, destination) ) if source == destination: return 0 while not queue_forward.empty() and not queue_backward.empty(): __lowercase =queue_forward.get() visited_forward.add(__lowerCamelCase ) __lowercase =queue_backward.get() visited_backward.add(__lowerCamelCase ) __lowercase =pass_and_relaxation( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, ) __lowercase =pass_and_relaxation( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, ) if cst_fwd[v_fwd] + cst_bwd[v_bwd] >= shortest_distance: break if shortest_distance != np.inf: __lowercase =shortest_distance return shortest_path_distance UpperCAmelCase = { "B": [["C", 1]], "C": [["D", 1]], "D": [["F", 1]], "E": [["B", 1], ["G", 2]], "F": [], "G": [["F", 1]], } UpperCAmelCase = { "B": [["E", 1]], "C": [["B", 1]], "D": [["C", 1]], "F": [["D", 1], ["G", 1]], "E": [[None, np.inf]], "G": [["E", 2]], } if __name__ == "__main__": import doctest doctest.testmod()
119
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class a (unittest.TestCase ): """simple docstring""" def __init__( self : int , lowerCamelCase : Any , lowerCamelCase : int=7 , lowerCamelCase : str=3 , lowerCamelCase : Optional[int]=30 , lowerCamelCase : Dict=400 , lowerCamelCase : str=True , lowerCamelCase : str=None , lowerCamelCase : Any=True , lowerCamelCase : Union[str, Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[str]=True , lowerCamelCase : Optional[int]=1 / 255 , lowerCamelCase : Any=True , ) -> str: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __snake_case : Optional[Any] = size if size is not None else {"shortest_edge": 18, "longest_edge": 1333} __snake_case : Optional[Any] = parent __snake_case : List[Any] = batch_size __snake_case : Optional[int] = num_channels __snake_case : str = min_resolution __snake_case : int = max_resolution __snake_case : int = do_resize __snake_case : Tuple = size __snake_case : Any = do_normalize __snake_case : int = image_mean __snake_case : Tuple = image_std __snake_case : Dict = do_rescale __snake_case : Optional[Any] = rescale_factor __snake_case : str = do_pad def __snake_case ( self : Any ) -> int: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def __snake_case ( self : Dict , lowerCamelCase : List[Any] , lowerCamelCase : Optional[Any]=False ) -> List[str]: if not batched: __snake_case : Dict = image_inputs[0] if isinstance(lowerCamelCase , Image.Image ): __snake_case , __snake_case : Dict = image.size else: __snake_case , __snake_case : List[str] = image.shape[1], image.shape[2] if w < h: __snake_case : Optional[int] = int(self.size["shortest_edge"] * h / w ) __snake_case : int = self.size["shortest_edge"] elif w > h: __snake_case : List[str] = self.size["shortest_edge"] __snake_case : Optional[Any] = int(self.size["shortest_edge"] * w / h ) else: __snake_case : List[Any] = self.size["shortest_edge"] __snake_case : Any = self.size["shortest_edge"] else: __snake_case : int = [] for image in image_inputs: __snake_case , __snake_case : List[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[0] )[0] __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class a (_lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ConditionalDetrImageProcessor if is_vision_available() else None def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : Optional[Any] = ConditionalDetrImageProcessingTester(self ) @property def __snake_case ( self : Any ) -> str: return self.image_processor_tester.prepare_image_processor_dict() def __snake_case ( self : Optional[Any] ) -> Optional[int]: __snake_case : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase , "image_mean" ) ) self.assertTrue(hasattr(lowerCamelCase , "image_std" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_normalize" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase , "size" ) ) def __snake_case ( self : Any ) -> Dict: __snake_case : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 1333} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) __snake_case : str = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=lowerCamelCase ) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> Dict: pass def __snake_case ( self : Tuple ) -> str: # Initialize image_processing __snake_case : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , Image.Image ) # Test not batched input __snake_case : int = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) __snake_case : Dict = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> str: # Initialize image_processing __snake_case : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __snake_case : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , numpify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , np.ndarray ) # Test not batched input __snake_case : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : List[Any] = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> List[str]: # Initialize image_processing __snake_case : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , torchify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , torch.Tensor ) # Test not batched input __snake_case : Optional[Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : List[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : int = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def __snake_case ( self : Any ) -> Optional[int]: # prepare image and target __snake_case : List[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : List[Any] = {"image_id": 39769, "annotations": target} # encode them __snake_case : List[str] = ConditionalDetrImageProcessor.from_pretrained("microsoft/conditional-detr-resnet-50" ) __snake_case : List[str] = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : Tuple = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Optional[int] = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : List[Any] = torch.tensor([58_87.96_00, 1_12_50.20_61, 48_93_53.84_38, 83_71_22.75_00, 14_79_67.51_56, 16_57_32.34_38] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : Optional[int] = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([0.55_03, 0.27_65, 0.06_04, 0.22_15] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Optional[Any] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : List[Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : Dict = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify orig_size __snake_case : int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Tuple = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) ) @slow def __snake_case ( self : str ) -> Tuple: # prepare image, target and masks_path __snake_case : Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : str = {"file_name": "000000039769.png", "image_id": 39769, "segments_info": target} __snake_case : Optional[int] = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them __snake_case : int = ConditionalDetrImageProcessor(format="coco_panoptic" ) __snake_case : str = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , masks_path=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : List[str] = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Dict = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : Any = torch.tensor([14_79_79.68_75, 16_55_27.04_69, 48_46_38.59_38, 1_12_92.93_75, 58_79.65_62, 76_34.11_47] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : str = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : Optional[Any] = torch.tensor([0.26_25, 0.54_37, 0.46_88, 0.86_25] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Tuple = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : Dict = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : int = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify masks __snake_case : List[Any] = 822873 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , lowerCamelCase ) # verify orig_size __snake_case : List[str] = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Any = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) )
81
0
def __magic_name__ ( __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Optional[Any] ) -> str: while second != 0: __lowerCamelCase = first & second first ^= second __lowerCamelCase = c << 1 return first if __name__ == "__main__": import doctest doctest.testmod() SCREAMING_SNAKE_CASE__ : List[Any] = int(input("Enter the first number: ").strip()) SCREAMING_SNAKE_CASE__ : Optional[int] = int(input("Enter the second number: ").strip()) print(F'{add(first, second) = }')
298
from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from ... import AutoBackbone from ...modeling_outputs import SemanticSegmenterOutput from ...modeling_utils import PreTrainedModel from ...utils import add_start_docstrings, add_start_docstrings_to_model_forward, replace_return_docstrings from ...utils.backbone_utils import BackboneMixin from .configuration_upernet import UperNetConfig _snake_case : Optional[Any] = [ "openmmlab/upernet-convnext-tiny", # See all UperNet models at https://huggingface.co/models?filter=upernet ] # General docstring _snake_case : Dict = "UperNetConfig" class a (nn.Module ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : Union[int, Tuple[int, int]] , lowerCamelCase : Union[int, Tuple[int, int], str] = 0 , lowerCamelCase : bool = False , lowerCamelCase : Union[int, Tuple[int, int]] = 1 , ) -> None: super().__init__() __snake_case : Union[str, Any] = nn.Convad( in_channels=lowerCamelCase , out_channels=lowerCamelCase , kernel_size=lowerCamelCase , padding=lowerCamelCase , bias=lowerCamelCase , dilation=lowerCamelCase , ) __snake_case : Dict = nn.BatchNormad(lowerCamelCase ) __snake_case : List[Any] = nn.ReLU() def __snake_case ( self : List[Any] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : Dict = self.conv(lowerCamelCase ) __snake_case : int = self.batch_norm(lowerCamelCase ) __snake_case : Optional[Any] = self.activation(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : str , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : int ) -> None: super().__init__() __snake_case : Tuple = [ nn.AdaptiveAvgPoolad(lowerCamelCase ), UperNetConvModule(lowerCamelCase , lowerCamelCase , kernel_size=1 ), ] for i, layer in enumerate(self.layers ): self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : Dict , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : List[str] = input for layer in self.layers: __snake_case : Tuple = layer(lowerCamelCase ) return hidden_state class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Tuple[int, ...] , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : bool ) -> None: super().__init__() __snake_case : Dict = pool_scales __snake_case : List[str] = align_corners __snake_case : List[Any] = in_channels __snake_case : str = channels __snake_case : Optional[Any] = [] for i, pool_scale in enumerate(lowerCamelCase ): __snake_case : Union[str, Any] = UperNetPyramidPoolingBlock(pool_scale=lowerCamelCase , in_channels=lowerCamelCase , channels=lowerCamelCase ) self.blocks.append(lowerCamelCase ) self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> List[torch.Tensor]: __snake_case : Tuple = [] for ppm in self.blocks: __snake_case : Any = ppm(lowerCamelCase ) __snake_case : List[Any] = nn.functional.interpolate( lowerCamelCase , size=x.size()[2:] , mode="bilinear" , align_corners=self.align_corners ) ppm_outs.append(lowerCamelCase ) return ppm_outs class a (nn.Module ): """simple docstring""" def __init__( self : int , lowerCamelCase : str , lowerCamelCase : str ) -> List[str]: super().__init__() __snake_case : Dict = config __snake_case : List[str] = config.pool_scales # e.g. (1, 2, 3, 6) __snake_case : Tuple = in_channels __snake_case : str = config.hidden_size __snake_case : List[str] = False __snake_case : Any = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) # PSP Module __snake_case : Tuple = UperNetPyramidPoolingModule( self.pool_scales , self.in_channels[-1] , self.channels , align_corners=self.align_corners , ) __snake_case : List[str] = UperNetConvModule( self.in_channels[-1] + len(self.pool_scales ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) # FPN Module __snake_case : List[Any] = nn.ModuleList() __snake_case : Dict = nn.ModuleList() for in_channels in self.in_channels[:-1]: # skip the top layer __snake_case : Union[str, Any] = UperNetConvModule(lowerCamelCase , self.channels , kernel_size=1 ) __snake_case : Optional[int] = UperNetConvModule(self.channels , self.channels , kernel_size=3 , padding=1 ) self.lateral_convs.append(lowerCamelCase ) self.fpn_convs.append(lowerCamelCase ) __snake_case : int = UperNetConvModule( len(self.in_channels ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) def __snake_case ( self : List[str] ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> str: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : List[Any] , lowerCamelCase : Tuple ) -> Optional[int]: __snake_case : str = inputs[-1] __snake_case : int = [x] psp_outs.extend(self.psp_modules(lowerCamelCase ) ) __snake_case : Tuple = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Union[str, Any] = self.bottleneck(lowerCamelCase ) return output def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # build laterals __snake_case : Any = [lateral_conv(encoder_hidden_states[i] ) for i, lateral_conv in enumerate(self.lateral_convs )] laterals.append(self.psp_forward(lowerCamelCase ) ) # build top-down path __snake_case : Dict = len(lowerCamelCase ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Union[str, Any] = laterals[i - 1].shape[2:] __snake_case : Optional[int] = laterals[i - 1] + nn.functional.interpolate( laterals[i] , size=lowerCamelCase , mode="bilinear" , align_corners=self.align_corners ) # build outputs __snake_case : str = [self.fpn_convs[i](laterals[i] ) for i in range(used_backbone_levels - 1 )] # append psp feature fpn_outs.append(laterals[-1] ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Tuple = nn.functional.interpolate( fpn_outs[i] , size=fpn_outs[0].shape[2:] , mode="bilinear" , align_corners=self.align_corners ) __snake_case : str = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Optional[Any] = self.fpn_bottleneck(lowerCamelCase ) __snake_case : Tuple = self.classifier(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Optional[int] , lowerCamelCase : int = 2 , lowerCamelCase : int = 3 , lowerCamelCase : Union[int, Tuple[int, int]] = 1 ) -> None: super().__init__() __snake_case : List[Any] = config __snake_case : List[str] = config.auxiliary_in_channels __snake_case : List[Any] = config.auxiliary_channels __snake_case : Tuple = config.auxiliary_num_convs __snake_case : int = config.auxiliary_concat_input __snake_case : Optional[int] = in_index __snake_case : Tuple = (kernel_size // 2) * dilation __snake_case : Optional[int] = [] convs.append( UperNetConvModule( self.in_channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) for i in range(self.num_convs - 1 ): convs.append( UperNetConvModule( self.channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) if self.num_convs == 0: __snake_case : Union[str, Any] = nn.Identity() else: __snake_case : Any = nn.Sequential(*lowerCamelCase ) if self.concat_input: __snake_case : int = UperNetConvModule( self.in_channels + self.channels , self.channels , kernel_size=lowerCamelCase , padding=kernel_size // 2 ) __snake_case : Optional[Any] = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) def __snake_case ( self : Dict ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : Tuple ) -> Optional[int]: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : Optional[int] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # just take the relevant feature maps __snake_case : List[str] = encoder_hidden_states[self.in_index] __snake_case : Optional[Any] = self.convs(lowerCamelCase ) if self.concat_input: __snake_case : Optional[Any] = self.conv_cat(torch.cat([hidden_states, output] , dim=1 ) ) __snake_case : Union[str, Any] = self.classifier(lowerCamelCase ) return output class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = UperNetConfig __UpperCAmelCase : int = "pixel_values" __UpperCAmelCase : str = True def __snake_case ( self : Union[str, Any] , lowerCamelCase : List[Any] ) -> List[Any]: if isinstance(lowerCamelCase , lowerCamelCase ): module.backbone.init_weights() module.decode_head.init_weights() module.auxiliary_head.init_weights() def __snake_case ( self : Optional[Any] ) -> List[str]: self.backbone.init_weights() self.decode_head.init_weights() self.auxiliary_head.init_weights() def __snake_case ( self : List[Any] , lowerCamelCase : Any , lowerCamelCase : Optional[Any]=False ) -> Dict: if isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : Union[str, Any] = value _snake_case : Dict = R"\n Parameters:\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n config ([`UperNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" _snake_case : Tuple = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using\n [`AutoImageProcessor`]. See [`SegformerImageProcessor.__call__`] for details.\n output_attentions (`bool`, *optional*):\n Whether or not to return the attentions tensors of all attention layers in case the backbone has them. See\n `attentions` under returned tensors for more detail.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers of the backbone. See `hidden_states` under\n returned tensors for more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n" @add_start_docstrings( "UperNet framework leveraging any vision backbone e.g. for ADE20k, CityScapes." , _lowerCAmelCase , ) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Dict , lowerCamelCase : int ) -> Optional[int]: super().__init__(lowerCamelCase ) __snake_case : Any = AutoBackbone.from_config(config.backbone_config ) # Semantic segmentation head(s) __snake_case : Union[str, Any] = UperNetHead(lowerCamelCase , in_channels=self.backbone.channels ) __snake_case : Any = UperNetFCNHead(lowerCamelCase ) if config.use_auxiliary_head else None # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UPERNET_INPUTS_DOCSTRING.format("batch_size, sequence_length" ) ) @replace_return_docstrings(output_type=lowerCamelCase , config_class=_CONFIG_FOR_DOC ) def __snake_case ( self : Union[str, Any] , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , ) -> Union[tuple, SemanticSegmenterOutput]: __snake_case : Any = return_dict if return_dict is not None else self.config.use_return_dict __snake_case : int = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __snake_case : str = output_attentions if output_attentions is not None else self.config.output_attentions __snake_case : Tuple = self.backbone.forward_with_filtered_kwargs( lowerCamelCase , output_hidden_states=lowerCamelCase , output_attentions=lowerCamelCase ) __snake_case : List[Any] = outputs.feature_maps __snake_case : List[Any] = self.decode_head(lowerCamelCase ) __snake_case : List[str] = nn.functional.interpolate(lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : Optional[int] = None if self.auxiliary_head is not None: __snake_case : Dict = self.auxiliary_head(lowerCamelCase ) __snake_case : Dict = nn.functional.interpolate( lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : int = None if labels is not None: if self.config.num_labels == 1: raise ValueError("The number of labels should be greater than one" ) else: # compute weighted loss __snake_case : Any = CrossEntropyLoss(ignore_index=self.config.loss_ignore_index ) __snake_case : Union[str, Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = main_loss + self.config.auxiliary_loss_weight * auxiliary_loss if not return_dict: if output_hidden_states: __snake_case : Any = (logits,) + outputs[1:] else: __snake_case : str = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return SemanticSegmenterOutput( loss=lowerCamelCase , logits=lowerCamelCase , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
81
0
"""simple docstring""" __SCREAMING_SNAKE_CASE ="Tobias Carryer" from time import time class UpperCamelCase : def __init__( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase=int(time() ) ) -> Optional[Any]: # noqa: B008 '''simple docstring''' lowercase_ : int = multiplier lowercase_ : int = increment lowercase_ : Optional[Any] = modulo lowercase_ : List[str] = seed def _UpperCAmelCase ( self ) -> Optional[Any]: '''simple docstring''' lowercase_ : Any = (self.multiplier * self.seed + self.increment) % self.modulo return self.seed if __name__ == "__main__": # Show the LCG in action. __SCREAMING_SNAKE_CASE =LinearCongruentialGenerator(166_4525, 10_1390_4223, 2 << 31) while True: print(lcg.next_number())
425
from __future__ import annotations def lowerCAmelCase_ ( __lowerCamelCase ): if not nums: return 0 __snake_case : Any = nums[0] __snake_case : str = 0 for num in nums[1:]: __snake_case , __snake_case : List[str] = ( max_excluding + num, max(__lowerCamelCase , __lowerCamelCase ), ) return max(__lowerCamelCase , __lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
81
0
'''simple docstring''' from __future__ import annotations def __snake_case ( _UpperCAmelCase : Optional[Any]): return len(set(__lowerCamelCase)) == len(__lowerCamelCase) if __name__ == "__main__": import doctest doctest.testmod()
212
from __future__ import annotations from typing import Any def lowerCAmelCase_ ( __lowerCamelCase ): create_state_space_tree(__lowerCamelCase , [] , 0 ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): if index == len(__lowerCamelCase ): print(__lowerCamelCase ) return create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.pop() if __name__ == "__main__": _snake_case : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["A", "B", "C"]) generate_all_subsequences(seq)
81
0
import argparse import logging import pickle from collections import Counter logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) _lowerCAmelCase : Dict = logging.getLogger(__name__) if __name__ == "__main__": _lowerCAmelCase : int = argparse.ArgumentParser( description='''Token Counts for smoothing the masking probabilities in MLM (cf XLM/word2vec)''' ) parser.add_argument( '''--data_file''', type=str, default='''data/dump.bert-base-uncased.pickle''', help='''The binarized dataset.''' ) parser.add_argument( '''--token_counts_dump''', type=str, default='''data/token_counts.bert-base-uncased.pickle''', help='''The dump file.''' ) parser.add_argument('''--vocab_size''', default=30_522, type=int) _lowerCAmelCase : str = parser.parse_args() logger.info(F'''Loading data from {args.data_file}''') with open(args.data_file, '''rb''') as fp: _lowerCAmelCase : str = pickle.load(fp) logger.info('''Counting occurrences for MLM.''') _lowerCAmelCase : Optional[int] = Counter() for tk_ids in data: counter.update(tk_ids) _lowerCAmelCase : int = [0] * args.vocab_size for k, v in counter.items(): _lowerCAmelCase : List[Any] = v logger.info(F'''Dump to {args.token_counts_dump}''') with open(args.token_counts_dump, '''wb''') as handle: pickle.dump(counts, handle, protocol=pickle.HIGHEST_PROTOCOL)
454
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case , __snake_case : List[Any] = image.size __snake_case , __snake_case : Tuple = (x - x % 3_2 for x in (w, h)) # resize to integer multiple of 32 __snake_case : str = image.resize((w, h) , resample=PIL_INTERPOLATION["lanczos"] ) __snake_case : int = np.array(__lowerCamelCase ).astype(np.floataa ) / 2_5_5.0 __snake_case : Union[str, Any] = image[None].transpose(0 , 3 , 1 , 2 ) __snake_case : Union[str, Any] = torch.from_numpy(__lowerCamelCase ) return 2.0 * image - 1.0 class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : VQModel , lowerCamelCase : UNetaDModel , lowerCamelCase : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ) -> Union[str, Any]: super().__init__() self.register_modules(vqvae=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase ) @torch.no_grad() def __call__( self : List[str] , lowerCamelCase : Union[torch.Tensor, PIL.Image.Image] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : Optional[int] = 100 , lowerCamelCase : Optional[float] = 0.0 , lowerCamelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , ) -> Union[Tuple, ImagePipelineOutput]: if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : Any = 1 elif isinstance(lowerCamelCase , torch.Tensor ): __snake_case : Any = image.shape[0] else: raise ValueError(F'`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(lowerCamelCase )}' ) if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : List[Any] = preprocess(lowerCamelCase ) __snake_case , __snake_case : int = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image __snake_case : str = (batch_size, self.unet.config.in_channels // 2, height, width) __snake_case : str = next(self.unet.parameters() ).dtype __snake_case : Tuple = randn_tensor(lowerCamelCase , generator=lowerCamelCase , device=self.device , dtype=lowerCamelCase ) __snake_case : List[Any] = image.to(device=self.device , dtype=lowerCamelCase ) # set timesteps and move to the correct device self.scheduler.set_timesteps(lowerCamelCase , device=self.device ) __snake_case : str = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler __snake_case : Dict = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] __snake_case : Union[str, Any] = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __snake_case : int = {} if accepts_eta: __snake_case : List[str] = eta for t in self.progress_bar(lowerCamelCase ): # concat latents and low resolution image in the channel dimension. __snake_case : Union[str, Any] = torch.cat([latents, image] , dim=1 ) __snake_case : Optional[Any] = self.scheduler.scale_model_input(lowerCamelCase , lowerCamelCase ) # predict the noise residual __snake_case : int = self.unet(lowerCamelCase , lowerCamelCase ).sample # compute the previous noisy sample x_t -> x_t-1 __snake_case : Union[str, Any] = self.scheduler.step(lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase ).prev_sample # decode the image latents with the VQVAE __snake_case : List[Any] = self.vqvae.decode(lowerCamelCase ).sample __snake_case : Dict = torch.clamp(lowerCamelCase , -1.0 , 1.0 ) __snake_case : Any = image / 2 + 0.5 __snake_case : Tuple = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __snake_case : Tuple = self.numpy_to_pil(lowerCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCamelCase )
81
0
"""simple docstring""" import copy import os from typing import TYPE_CHECKING, List, Union if TYPE_CHECKING: pass from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { "kakaobrain/align-base": "https://huggingface.co/kakaobrain/align-base/resolve/main/config.json", } class _UpperCamelCase ( _lowerCAmelCase): __lowerCamelCase = "align_text_model" def __init__(self , lowerCamelCase__=3_0_5_2_2 , lowerCamelCase__=7_6_8 , lowerCamelCase__=1_2 , lowerCamelCase__=1_2 , lowerCamelCase__=3_0_7_2 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_2 , lowerCamelCase__=2 , lowerCamelCase__=0.0_2 , lowerCamelCase__=1E-12 , lowerCamelCase__=0 , lowerCamelCase__="absolute" , lowerCamelCase__=True , **lowerCamelCase__ , ): """simple docstring""" super().__init__(**lowerCamelCase__ ) A__ = vocab_size A__ = hidden_size A__ = num_hidden_layers A__ = num_attention_heads A__ = hidden_act A__ = intermediate_size A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = max_position_embeddings A__ = type_vocab_size A__ = initializer_range A__ = layer_norm_eps A__ = position_embedding_type A__ = use_cache A__ = pad_token_id @classmethod def A (cls , lowerCamelCase__ , **lowerCamelCase__ ): """simple docstring""" cls._set_token_in_kwargs(lowerCamelCase__ ) A__ = cls.get_config_dict(lowerCamelCase__ , **lowerCamelCase__ ) # get the text config dict if we are loading from AlignConfig if config_dict.get("""model_type""" ) == "align": A__ = config_dict["text_config"] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(lowerCamelCase__ , **lowerCamelCase__ ) class _UpperCamelCase ( _lowerCAmelCase): __lowerCamelCase = "align_vision_model" def __init__(self , lowerCamelCase__ = 3 , lowerCamelCase__ = 6_0_0 , lowerCamelCase__ = 2.0 , lowerCamelCase__ = 3.1 , lowerCamelCase__ = 8 , lowerCamelCase__ = [3, 3, 5, 3, 5, 5, 3] , lowerCamelCase__ = [3_2, 1_6, 2_4, 4_0, 8_0, 1_1_2, 1_9_2] , lowerCamelCase__ = [1_6, 2_4, 4_0, 8_0, 1_1_2, 1_9_2, 3_2_0] , lowerCamelCase__ = [] , lowerCamelCase__ = [1, 2, 2, 2, 1, 2, 1] , lowerCamelCase__ = [1, 2, 2, 3, 3, 4, 1] , lowerCamelCase__ = [1, 6, 6, 6, 6, 6, 6] , lowerCamelCase__ = 0.2_5 , lowerCamelCase__ = "swish" , lowerCamelCase__ = 2_5_6_0 , lowerCamelCase__ = "mean" , lowerCamelCase__ = 0.0_2 , lowerCamelCase__ = 0.0_0_1 , lowerCamelCase__ = 0.9_9 , lowerCamelCase__ = 0.2 , **lowerCamelCase__ , ): """simple docstring""" super().__init__(**lowerCamelCase__ ) A__ = num_channels A__ = image_size A__ = width_coefficient A__ = depth_coefficient A__ = depth_divisor A__ = kernel_sizes A__ = in_channels A__ = out_channels A__ = depthwise_padding A__ = strides A__ = num_block_repeats A__ = expand_ratios A__ = squeeze_expansion_ratio A__ = hidden_act A__ = hidden_dim A__ = pooling_type A__ = initializer_range A__ = batch_norm_eps A__ = batch_norm_momentum A__ = drop_connect_rate A__ = sum(lowerCamelCase__ ) * 4 @classmethod def A (cls , lowerCamelCase__ , **lowerCamelCase__ ): """simple docstring""" cls._set_token_in_kwargs(lowerCamelCase__ ) A__ = cls.get_config_dict(lowerCamelCase__ , **lowerCamelCase__ ) # get the vision config dict if we are loading from AlignConfig if config_dict.get("""model_type""" ) == "align": A__ = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(lowerCamelCase__ , **lowerCamelCase__ ) class _UpperCamelCase ( _lowerCAmelCase): __lowerCamelCase = "align" __lowerCamelCase = True def __init__(self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=6_4_0 , lowerCamelCase__=1.0 , lowerCamelCase__=0.0_2 , **lowerCamelCase__ , ): """simple docstring""" super().__init__(**lowerCamelCase__ ) if text_config is None: A__ = {} logger.info("""text_config is None. Initializing the AlignTextConfig with default values.""" ) if vision_config is None: A__ = {} logger.info("""vision_config is None. Initializing the AlignVisionConfig with default values.""" ) A__ = AlignTextConfig(**lowerCamelCase__ ) A__ = AlignVisionConfig(**lowerCamelCase__ ) A__ = projection_dim A__ = temperature_init_value A__ = initializer_range @classmethod def A (cls , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ): """simple docstring""" return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **lowerCamelCase__ ) def A (self ): """simple docstring""" A__ = copy.deepcopy(self.__dict__ ) A__ = self.text_config.to_dict() A__ = self.vision_config.to_dict() A__ = self.__class__.model_type return output
574
import gc import unittest import torch from parameterized import parameterized from diffusers import AutoencoderKL from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = AutoencoderKL __UpperCAmelCase : Optional[Any] = "sample" __UpperCAmelCase : Optional[int] = 1e-2 @property def __snake_case ( self : Dict ) -> Optional[Any]: __snake_case : Optional[Any] = 4 __snake_case : Tuple = 3 __snake_case : List[str] = (32, 32) __snake_case : str = floats_tensor((batch_size, num_channels) + sizes ).to(lowerCamelCase ) return {"sample": image} @property def __snake_case ( self : Union[str, Any] ) -> Tuple: return (3, 32, 32) @property def __snake_case ( self : int ) -> int: return (3, 32, 32) def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Optional[Any] = { "block_out_channels": [32, 64], "in_channels": 3, "out_channels": 3, "down_block_types": ["DownEncoderBlock2D", "DownEncoderBlock2D"], "up_block_types": ["UpDecoderBlock2D", "UpDecoderBlock2D"], "latent_channels": 4, } __snake_case : Any = self.dummy_input return init_dict, inputs_dict def __snake_case ( self : str ) -> Dict: pass def __snake_case ( self : Tuple ) -> List[str]: pass @unittest.skipIf(torch_device == "mps" , "Gradient checkpointing skipped on MPS" ) def __snake_case ( self : Any ) -> Optional[Any]: # enable deterministic behavior for gradient checkpointing __snake_case , __snake_case : int = self.prepare_init_args_and_inputs_for_common() __snake_case : str = self.model_class(**lowerCamelCase ) model.to(lowerCamelCase ) assert not model.is_gradient_checkpointing and model.training __snake_case : str = model(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model.zero_grad() __snake_case : Any = torch.randn_like(lowerCamelCase ) __snake_case : str = (out - labels).mean() loss.backward() # re-instantiate the model now enabling gradient checkpointing __snake_case : Optional[int] = self.model_class(**lowerCamelCase ) # clone model model_a.load_state_dict(model.state_dict() ) model_a.to(lowerCamelCase ) model_a.enable_gradient_checkpointing() assert model_a.is_gradient_checkpointing and model_a.training __snake_case : int = model_a(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model_a.zero_grad() __snake_case : Union[str, Any] = (out_a - labels).mean() loss_a.backward() # compare the output and parameters gradients self.assertTrue((loss - loss_a).abs() < 1E-5 ) __snake_case : Optional[int] = dict(model.named_parameters() ) __snake_case : List[Any] = dict(model_a.named_parameters() ) for name, param in named_params.items(): self.assertTrue(torch_all_close(param.grad.data , named_params_a[name].grad.data , atol=5E-5 ) ) def __snake_case ( self : List[Any] ) -> Optional[int]: __snake_case , __snake_case : Optional[Any] = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" , output_loading_info=lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(lowerCamelCase ) __snake_case : Optional[Any] = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def __snake_case ( self : Optional[Any] ) -> Union[str, Any]: __snake_case : Tuple = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" ) __snake_case : Dict = model.to(lowerCamelCase ) model.eval() if torch_device == "mps": __snake_case : int = torch.manual_seed(0 ) else: __snake_case : str = torch.Generator(device=lowerCamelCase ).manual_seed(0 ) __snake_case : List[str] = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) __snake_case : Union[str, Any] = image.to(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , sample_posterior=lowerCamelCase , generator=lowerCamelCase ).sample __snake_case : List[Any] = output[0, -1, -3:, -3:].flatten().cpu() # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. if torch_device == "mps": __snake_case : Union[str, Any] = torch.tensor( [ -4.0078E-01, -3.8323E-04, -1.2681E-01, -1.1462E-01, 2.0095E-01, 1.0893E-01, -8.8247E-02, -3.0361E-01, -9.8644E-03, ] ) elif torch_device == "cpu": __snake_case : Tuple = torch.tensor( [-0.13_52, 0.08_78, 0.04_19, -0.08_18, -0.10_69, 0.06_88, -0.14_58, -0.44_46, -0.00_26] ) else: __snake_case : List[str] = torch.tensor( [-0.24_21, 0.46_42, 0.25_07, -0.04_38, 0.06_82, 0.31_60, -0.20_18, -0.07_27, 0.24_85] ) self.assertTrue(torch_all_close(lowerCamelCase , lowerCamelCase , rtol=1E-2 ) ) @slow class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : int , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] ) -> List[str]: return F'gaussian_noise_s={seed}_shape={"_".join([str(lowerCamelCase ) for s in shape] )}.npy' def __snake_case ( self : List[Any] ) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __snake_case ( self : Tuple , lowerCamelCase : List[Any]=0 , lowerCamelCase : Tuple=(4, 3, 512, 512) , lowerCamelCase : Optional[int]=False ) -> str: __snake_case : List[Any] = torch.floataa if fpaa else torch.floataa __snake_case : Tuple = torch.from_numpy(load_hf_numpy(self.get_file_format(lowerCamelCase , lowerCamelCase ) ) ).to(lowerCamelCase ).to(lowerCamelCase ) return image def __snake_case ( self : Optional[Any] , lowerCamelCase : int="CompVis/stable-diffusion-v1-4" , lowerCamelCase : int=False ) -> int: __snake_case : str = "fp16" if fpaa else None __snake_case : int = torch.floataa if fpaa else torch.floataa __snake_case : int = AutoencoderKL.from_pretrained( lowerCamelCase , subfolder="vae" , torch_dtype=lowerCamelCase , revision=lowerCamelCase , ) model.to(lowerCamelCase ).eval() return model def __snake_case ( self : str , lowerCamelCase : int=0 ) -> Optional[Any]: if torch_device == "mps": return torch.manual_seed(lowerCamelCase ) return torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) @parameterized.expand( [ # fmt: off [33, [-0.16_03, 0.98_78, -0.04_95, -0.07_90, -0.27_09, 0.83_75, -0.20_60, -0.08_24], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_76, 0.11_68, 0.13_32, -0.48_40, -0.25_08, -0.07_91, -0.04_93, -0.40_89], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[str] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] ) -> List[Any]: __snake_case : Optional[Any] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) __snake_case : Tuple = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : List[Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : int = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [33, [-0.05_13, 0.02_89, 1.37_99, 0.21_66, -0.25_73, -0.08_71, 0.51_03, -0.09_99]], [47, [-0.41_28, -0.13_20, -0.37_04, 0.19_65, -0.41_16, -0.23_32, -0.33_40, 0.22_47]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : Any , lowerCamelCase : List[str] , lowerCamelCase : List[str] ) -> Tuple: __snake_case : Any = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Any = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.16_09, 0.98_66, -0.04_87, -0.07_77, -0.27_16, 0.83_68, -0.20_55, -0.08_14], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_77, 0.11_47, 0.13_33, -0.48_41, -0.25_06, -0.08_05, -0.04_91, -0.40_85], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Any , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) with torch.no_grad(): __snake_case : int = model(lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Union[str, Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : List[str] = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [13, [-0.20_51, -0.18_03, -0.23_11, -0.21_14, -0.32_92, -0.35_74, -0.29_53, -0.33_23]], [37, [-0.26_32, -0.26_25, -0.21_99, -0.27_41, -0.45_39, -0.49_90, -0.37_20, -0.49_25]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : List[str] , lowerCamelCase : Tuple , lowerCamelCase : Any ) -> Optional[Any]: __snake_case : List[str] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : str = sample[-1, -2:, :2, -2:].flatten().cpu() __snake_case : Optional[int] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-3 ) @parameterized.expand( [ # fmt: off [27, [-0.03_69, 0.02_07, -0.07_76, -0.06_82, -0.17_47, -0.19_30, -0.14_65, -0.20_39]], [16, [-0.16_28, -0.21_34, -0.27_47, -0.26_42, -0.37_74, -0.44_04, -0.36_87, -0.42_77]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : str , lowerCamelCase : Optional[int] , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[str] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : Union[str, Any] = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Optional[Any] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=5E-3 ) @parameterized.expand([(13,), (16,), (27,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> Tuple: __snake_case : Dict = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : Any = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Any = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-1 ) @parameterized.expand([(13,), (16,), (37,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : List[Any] , lowerCamelCase : Any ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : Union[str, Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : List[Any] = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Dict = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.30_01, 0.09_18, -2.69_84, -3.97_20, -3.20_99, -5.03_53, 1.73_38, -0.20_65, 3.42_67]], [47, [-1.50_30, -4.38_71, -6.03_55, -9.11_57, -1.66_61, -2.78_53, 2.16_07, -5.08_23, 2.56_33]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Dict ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : int = self.get_sd_image(lowerCamelCase ) __snake_case : int = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model.encode(lowerCamelCase ).latent_dist __snake_case : Dict = dist.sample(generator=lowerCamelCase ) assert list(sample.shape ) == [image.shape[0], 4] + [i // 8 for i in image.shape[2:]] __snake_case : List[str] = sample[0, -1, -3:, -3:].flatten().cpu() __snake_case : Dict = torch.tensor(lowerCamelCase ) __snake_case : Dict = 3E-3 if torch_device != "mps" else 1E-2 assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=lowerCamelCase )
81
0
'''simple docstring''' class UpperCAmelCase : '''simple docstring''' def __init__( self , __lowerCAmelCase = "" , __lowerCAmelCase = False ) -> None: # Mapping from the first character of the prefix of the node lowercase__ : dict[str, RadixNode] = {} # A node will be a leaf if the tree contains its word lowercase__ : int = is_leaf lowercase__ : Union[str, Any] = prefix def _lowerCAmelCase( self , __lowerCAmelCase ) -> tuple[str, str, str]: lowercase__ : Any = 0 for q, w in zip(self.prefix , __lowerCAmelCase ): if q != w: break x += 1 return self.prefix[:x], self.prefix[x:], word[x:] def _lowerCAmelCase( self , __lowerCAmelCase ) -> None: for word in words: self.insert(__lowerCAmelCase ) def _lowerCAmelCase( self , __lowerCAmelCase ) -> None: # Case 1: If the word is the prefix of the node # Solution: We set the current node as leaf if self.prefix == word: lowercase__ : Optional[int] = True # Case 2: The node has no edges that have a prefix to the word # Solution: We create an edge from the current node to a new one # containing the word elif word[0] not in self.nodes: lowercase__ : Optional[Any] = RadixNode(prefix=__lowerCAmelCase , is_leaf=__lowerCAmelCase ) else: lowercase__ : int = self.nodes[word[0]] lowercase__ : Optional[Any] = incoming_node.match( __lowerCAmelCase ) # Case 3: The node prefix is equal to the matching # Solution: We insert remaining word on the next node if remaining_prefix == "": self.nodes[matching_string[0]].insert(__lowerCAmelCase ) # Case 4: The word is greater equal to the matching # Solution: Create a node in between both nodes, change # prefixes and add the new node for the remaining word else: lowercase__ : List[Any] = remaining_prefix lowercase__ : Dict = self.nodes[matching_string[0]] lowercase__ : List[str] = RadixNode(__lowerCAmelCase , __lowerCAmelCase ) lowercase__ : Dict = aux_node if remaining_word == "": lowercase__ : Tuple = True else: self.nodes[matching_string[0]].insert(__lowerCAmelCase ) def _lowerCAmelCase( self , __lowerCAmelCase ) -> bool: lowercase__ : str = self.nodes.get(word[0] , __lowerCAmelCase ) if not incoming_node: return False else: lowercase__ : str = incoming_node.match( __lowerCAmelCase ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # This applies when the word and the prefix are equal elif remaining_word == "": return incoming_node.is_leaf # We have word remaining so we check the next node else: return incoming_node.find(__lowerCAmelCase ) def _lowerCAmelCase( self , __lowerCAmelCase ) -> bool: lowercase__ : Any = self.nodes.get(word[0] , __lowerCAmelCase ) if not incoming_node: return False else: lowercase__ : Dict = incoming_node.match( __lowerCAmelCase ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # We have word remaining so we check the next node elif remaining_word != "": return incoming_node.delete(__lowerCAmelCase ) else: # If it is not a leaf, we don't have to delete if not incoming_node.is_leaf: return False else: # We delete the nodes if no edges go from it if len(incoming_node.nodes ) == 0: del self.nodes[word[0]] # We merge the current node with its only child if len(self.nodes ) == 1 and not self.is_leaf: lowercase__ : Dict = list(self.nodes.values() )[0] lowercase__ : Any = merging_node.is_leaf self.prefix += merging_node.prefix lowercase__ : Tuple = merging_node.nodes # If there is more than 1 edge, we just mark it as non-leaf elif len(incoming_node.nodes ) > 1: lowercase__ : Optional[int] = False # If there is 1 edge, we merge it with its child else: lowercase__ : Dict = list(incoming_node.nodes.values() )[0] lowercase__ : Optional[int] = merging_node.is_leaf incoming_node.prefix += merging_node.prefix lowercase__ : List[str] = merging_node.nodes return True def _lowerCAmelCase( self , __lowerCAmelCase = 0 ) -> None: if self.prefix != "": print('''-''' * height , self.prefix , ''' (leaf)''' if self.is_leaf else '''''' ) for value in self.nodes.values(): value.print_tree(height + 1 ) def __UpperCamelCase ( ): lowercase__ : Dict = "banana bananas bandana band apple all beast".split() lowercase__ : int = RadixNode() root.insert_many(__lowerCamelCase ) assert all(root.find(__lowerCamelCase ) for word in words ) assert not root.find('''bandanas''' ) assert not root.find('''apps''' ) root.delete('''all''' ) assert not root.find('''all''' ) root.delete('''banana''' ) assert not root.find('''banana''' ) assert root.find('''bananas''' ) return True def __UpperCamelCase ( ): assert test_trie() def __UpperCamelCase ( ): lowercase__ : Optional[Any] = RadixNode() lowercase__ : Union[str, Any] = "banana bananas bandanas bandana band apple all beast".split() root.insert_many(__lowerCamelCase ) print('''Words:''' , __lowerCamelCase ) print('''Tree:''' ) root.print_tree() if __name__ == "__main__": main()
152
import warnings from ...utils import logging from .image_processing_flava import FlavaImageProcessor _snake_case : Optional[int] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[Any] , *lowerCamelCase : Any , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class FlavaFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use FlavaImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
import warnings from ...utils import logging from .image_processing_flava import FlavaImageProcessor lowerCAmelCase_ = logging.get_logger(__name__) class A (_lowerCAmelCase ): def __init__( self , *lowercase_ , **lowercase_ ) -> None: '''simple docstring''' warnings.warn( '''The class FlavaFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use FlavaImageProcessor instead.''' , lowercase_ , ) super().__init__(*lowercase_ , **lowercase_ )
326
import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : Tuple ) -> Optional[Any]: __snake_case : Dict = tempfile.mkdtemp() __snake_case : Any = SamImageProcessor() __snake_case : Optional[int] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : Optional[Any] , **lowerCamelCase : Optional[int] ) -> Optional[Any]: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[Any] ) -> Dict: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : int ) -> List[Any]: __snake_case : List[str] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : int = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : List[Any] ) -> Dict: __snake_case : int = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Dict = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Optional[Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : List[str] ) -> Tuple: __snake_case : int = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Optional[int] = self.prepare_image_inputs() __snake_case : List[str] = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : Dict = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_torch def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Tuple = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[str] = [torch.ones((1, 3, 5, 5) )] __snake_case : Tuple = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : int = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , torch.tensor(lowerCamelCase ) , torch.tensor(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : List[str] = [np.ones((1, 3, 5, 5) )] __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : str = [[1, 0], [0, 1]] with self.assertRaises(lowerCamelCase ): __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) @require_vision @require_tf class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[Any] ) -> Union[str, Any]: __snake_case : int = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : str , **lowerCamelCase : Any ) -> Tuple: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> Any: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : str ) -> List[Any]: __snake_case : int = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : Dict = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : int ) -> List[str]: __snake_case : List[Any] = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Optional[Any] = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Tuple = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : Union[str, Any] ) -> List[Any]: __snake_case : str = self.get_image_processor() __snake_case : Union[str, Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : int = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : List[str] = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_tf def __snake_case ( self : Any ) -> Optional[int]: __snake_case : List[str] = self.get_image_processor() __snake_case : Dict = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Union[str, Any] = [tf.ones((1, 3, 5, 5) )] __snake_case : List[Any] = [[1764, 2646]] __snake_case : Dict = [[683, 1024]] __snake_case : List[str] = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , tf.convert_to_tensor(lowerCamelCase ) , tf.convert_to_tensor(lowerCamelCase ) , return_tensors="tf" , ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : Union[str, Any] = [np.ones((1, 3, 5, 5) )] __snake_case : List[str] = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Tuple = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): __snake_case : Dict = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[str] ) -> str: __snake_case : Optional[int] = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : List[str] , **lowerCamelCase : Any ) -> Dict: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> List[Any]: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : str = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : List[Any] = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : str = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[Any] = np.random.randint(0 , 2 , size=(1, 3, 5, 5) ).astype(np.floataa ) __snake_case : Dict = [tf.convert_to_tensor(lowerCamelCase )] __snake_case : List[Any] = [torch.tensor(lowerCamelCase )] __snake_case : Optional[Any] = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : Union[str, Any] = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) __snake_case : Dict = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="pt" ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def __snake_case ( self : List[Any] ) -> List[str]: __snake_case : Any = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : Any = image_processor(lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Optional[Any] = processor(images=lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Tuple = image_processor(lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() __snake_case : List[Any] = processor(images=lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) )
81
0
"""simple docstring""" import os import tempfile import unittest from transformers import FlaubertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FlaubertForMultipleChoice, FlaubertForQuestionAnswering, FlaubertForQuestionAnsweringSimple, FlaubertForSequenceClassification, FlaubertForTokenClassification, FlaubertModel, FlaubertWithLMHeadModel, ) from transformers.models.flaubert.modeling_flaubert import FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST class SCREAMING_SNAKE_CASE_ ( _lowerCAmelCase ): """simple docstring""" def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=1_3 , lowerCAmelCase__=7 , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__=False , lowerCAmelCase__=False , lowerCAmelCase__=2 , lowerCAmelCase__=9_9 , lowerCAmelCase__=0 , lowerCAmelCase__=3_2 , lowerCAmelCase__=5 , lowerCAmelCase__=4 , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=5_1_2 , lowerCAmelCase__=1_2 , lowerCAmelCase__=2 , lowerCAmelCase__=0.02 , lowerCAmelCase__=3 , lowerCAmelCase__=4 , lowerCAmelCase__="last" , lowerCAmelCase__=None , lowerCAmelCase__=None , ): __SCREAMING_SNAKE_CASE = parent __SCREAMING_SNAKE_CASE = batch_size __SCREAMING_SNAKE_CASE = seq_length __SCREAMING_SNAKE_CASE = is_training __SCREAMING_SNAKE_CASE = use_input_lengths __SCREAMING_SNAKE_CASE = use_token_type_ids __SCREAMING_SNAKE_CASE = use_labels __SCREAMING_SNAKE_CASE = gelu_activation __SCREAMING_SNAKE_CASE = sinusoidal_embeddings __SCREAMING_SNAKE_CASE = causal __SCREAMING_SNAKE_CASE = asm __SCREAMING_SNAKE_CASE = n_langs __SCREAMING_SNAKE_CASE = vocab_size __SCREAMING_SNAKE_CASE = n_special __SCREAMING_SNAKE_CASE = hidden_size __SCREAMING_SNAKE_CASE = num_hidden_layers __SCREAMING_SNAKE_CASE = num_attention_heads __SCREAMING_SNAKE_CASE = hidden_dropout_prob __SCREAMING_SNAKE_CASE = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE = max_position_embeddings __SCREAMING_SNAKE_CASE = type_vocab_size __SCREAMING_SNAKE_CASE = type_sequence_label_size __SCREAMING_SNAKE_CASE = initializer_range __SCREAMING_SNAKE_CASE = num_labels __SCREAMING_SNAKE_CASE = num_choices __SCREAMING_SNAKE_CASE = summary_type __SCREAMING_SNAKE_CASE = use_proj __SCREAMING_SNAKE_CASE = scope def snake_case_ ( self): __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) __SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length]) __SCREAMING_SNAKE_CASE = None if self.use_input_lengths: __SCREAMING_SNAKE_CASE = ( ids_tensor([self.batch_size] , vocab_size=2) + self.seq_length - 2 ) # small variation of seq_length __SCREAMING_SNAKE_CASE = None if self.use_token_type_ids: __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.n_langs) __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = None if self.use_labels: __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.type_sequence_label_size) __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , 2).float() __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.num_choices) __SCREAMING_SNAKE_CASE = self.get_config() return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def snake_case_ ( self): return FlaubertConfig( vocab_size=self.vocab_size , n_special=self.n_special , emb_dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , gelu_activation=self.gelu_activation , sinusoidal_embeddings=self.sinusoidal_embeddings , asm=self.asm , causal=self.causal , n_langs=self.n_langs , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , summary_type=self.summary_type , use_proj=self.use_proj , ) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ): __SCREAMING_SNAKE_CASE = FlaubertModel(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , lengths=lowerCAmelCase__ , langs=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , langs=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ): __SCREAMING_SNAKE_CASE = FlaubertWithLMHeadModel(lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , labels=lowerCAmelCase__) self.parent.assertEqual(result.loss.shape , ()) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ): __SCREAMING_SNAKE_CASE = FlaubertForQuestionAnsweringSimple(lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , start_positions=lowerCAmelCase__ , end_positions=lowerCAmelCase__) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ): __SCREAMING_SNAKE_CASE = FlaubertForQuestionAnswering(lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = model( lowerCAmelCase__ , start_positions=lowerCAmelCase__ , end_positions=lowerCAmelCase__ , cls_index=lowerCAmelCase__ , is_impossible=lowerCAmelCase__ , p_mask=lowerCAmelCase__ , ) __SCREAMING_SNAKE_CASE = model( lowerCAmelCase__ , start_positions=lowerCAmelCase__ , end_positions=lowerCAmelCase__ , cls_index=lowerCAmelCase__ , is_impossible=lowerCAmelCase__ , ) (__SCREAMING_SNAKE_CASE ) = result_with_labels.to_tuple() __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , start_positions=lowerCAmelCase__ , end_positions=lowerCAmelCase__) (__SCREAMING_SNAKE_CASE ) = result_with_labels.to_tuple() self.parent.assertEqual(result_with_labels.loss.shape , ()) self.parent.assertEqual(result.start_top_log_probs.shape , (self.batch_size, model.config.start_n_top)) self.parent.assertEqual(result.start_top_index.shape , (self.batch_size, model.config.start_n_top)) self.parent.assertEqual( result.end_top_log_probs.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top)) self.parent.assertEqual( result.end_top_index.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top)) self.parent.assertEqual(result.cls_logits.shape , (self.batch_size,)) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ): __SCREAMING_SNAKE_CASE = FlaubertForSequenceClassification(lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , labels=lowerCAmelCase__) self.parent.assertEqual(result.loss.shape , ()) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size)) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ): __SCREAMING_SNAKE_CASE = self.num_labels __SCREAMING_SNAKE_CASE = FlaubertForTokenClassification(lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , labels=lowerCAmelCase__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ): __SCREAMING_SNAKE_CASE = self.num_choices __SCREAMING_SNAKE_CASE = FlaubertForMultipleChoice(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = input_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() __SCREAMING_SNAKE_CASE = token_type_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() __SCREAMING_SNAKE_CASE = input_mask.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() __SCREAMING_SNAKE_CASE = model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , labels=lowerCAmelCase__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices)) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() ( __SCREAMING_SNAKE_CASE ) = config_and_inputs __SCREAMING_SNAKE_CASE = { "input_ids": input_ids, "token_type_ids": token_type_ids, "lengths": input_lengths, "attention_mask": input_mask, } return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE_ ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __lowercase : Dict = ( ( FlaubertModel, FlaubertWithLMHeadModel, FlaubertForQuestionAnswering, FlaubertForQuestionAnsweringSimple, FlaubertForSequenceClassification, FlaubertForTokenClassification, FlaubertForMultipleChoice, ) if is_torch_available() else () ) __lowercase : Optional[Any] = ( { "feature-extraction": FlaubertModel, "fill-mask": FlaubertWithLMHeadModel, "question-answering": FlaubertForQuestionAnsweringSimple, "text-classification": FlaubertForSequenceClassification, "token-classification": FlaubertForTokenClassification, "zero-shot": FlaubertForSequenceClassification, } if is_torch_available() else {} ) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith("""Fast""") ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=False): __SCREAMING_SNAKE_CASE = super()._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ , return_labels=lowerCAmelCase__) if return_labels: if model_class.__name__ == "FlaubertForQuestionAnswering": __SCREAMING_SNAKE_CASE = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowerCAmelCase__) return inputs_dict def snake_case_ ( self): __SCREAMING_SNAKE_CASE = FlaubertModelTester(self) __SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=lowerCAmelCase__ , emb_dim=3_7) def snake_case_ ( self): self.config_tester.run_common_tests() def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_model(*lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_lm_head(*lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_simple_qa(*lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_qa(*lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_sequence_classif(*lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_token_classif(*lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_multiple_choice(*lowerCAmelCase__) @slow def snake_case_ ( self): for model_name in FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __SCREAMING_SNAKE_CASE = FlaubertModel.from_pretrained(lowerCAmelCase__) self.assertIsNotNone(lowerCAmelCase__) @slow @require_torch_gpu def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # FlauBertForMultipleChoice behaves incorrectly in JIT environments. if model_class == FlaubertForMultipleChoice: return __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = model_class(config=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__) __SCREAMING_SNAKE_CASE = torch.jit.trace( lowerCAmelCase__ , (inputs_dict["""input_ids"""].to("""cpu"""), inputs_dict["""attention_mask"""].to("""cpu"""))) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(lowerCAmelCase__ , os.path.join(lowerCAmelCase__ , """traced_model.pt""")) __SCREAMING_SNAKE_CASE = torch.jit.load(os.path.join(lowerCAmelCase__ , """traced_model.pt""") , map_location=lowerCAmelCase__) loaded(inputs_dict["""input_ids"""].to(lowerCAmelCase__) , inputs_dict["""attention_mask"""].to(lowerCAmelCase__)) @require_torch class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): """simple docstring""" @slow def snake_case_ ( self): __SCREAMING_SNAKE_CASE = FlaubertModel.from_pretrained("""flaubert/flaubert_base_cased""") __SCREAMING_SNAKE_CASE = torch.tensor([[0, 3_4_5, 2_3_2, 3_2_8, 7_4_0, 1_4_0, 1_6_9_5, 6_9, 6_0_7_8, 1_5_8_8, 2]]) with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__)[0] __SCREAMING_SNAKE_CASE = torch.Size((1, 1_1, 7_6_8)) self.assertEqual(output.shape , lowerCAmelCase__) __SCREAMING_SNAKE_CASE = torch.tensor( [[[-2.62_51, -1.42_98, -0.02_27], [-2.85_10, -1.63_87, 0.22_58], [-2.81_14, -1.18_32, -0.30_66]]]) self.assertTrue(torch.allclose(output[:, :3, :3] , lowerCAmelCase__ , atol=1E-4))
155
import warnings from ...utils import logging from .image_processing_deit import DeiTImageProcessor _snake_case : List[str] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[str] , *lowerCamelCase : str , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class DeiTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use DeiTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _UpperCAmelCase : str = { "configuration_whisper": ["WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP", "WhisperConfig", "WhisperOnnxConfig"], "feature_extraction_whisper": ["WhisperFeatureExtractor"], "processing_whisper": ["WhisperProcessor"], "tokenization_whisper": ["WhisperTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[Any] = ["WhisperTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[str] = [ "WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST", "WhisperForConditionalGeneration", "WhisperModel", "WhisperPreTrainedModel", "WhisperForAudioClassification", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[int] = [ "TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST", "TFWhisperForConditionalGeneration", "TFWhisperModel", "TFWhisperPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[Any] = [ "FlaxWhisperForConditionalGeneration", "FlaxWhisperModel", "FlaxWhisperPreTrainedModel", "FlaxWhisperForAudioClassification", ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys _UpperCAmelCase : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
668
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case : Union[str, Any] = { "configuration_owlvit": [ "OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "OwlViTConfig", "OwlViTOnnxConfig", "OwlViTTextConfig", "OwlViTVisionConfig", ], "processing_owlvit": ["OwlViTProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[Any] = ["OwlViTFeatureExtractor"] _snake_case : Optional[int] = ["OwlViTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "OwlViTModel", "OwlViTPreTrainedModel", "OwlViTTextModel", "OwlViTVisionModel", "OwlViTForObjectDetection", ] if TYPE_CHECKING: from .configuration_owlvit import ( OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, OwlViTConfig, OwlViTOnnxConfig, OwlViTTextConfig, OwlViTVisionConfig, ) from .processing_owlvit import OwlViTProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_owlvit import OwlViTFeatureExtractor from .image_processing_owlvit import OwlViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_owlvit import ( OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST, OwlViTForObjectDetection, OwlViTModel, OwlViTPreTrainedModel, OwlViTTextModel, OwlViTVisionModel, ) else: import sys _snake_case : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = { "tiiuae/falcon-40b": "https://huggingface.co/tiiuae/falcon-40b/resolve/main/config.json", "tiiuae/falcon-7b": "https://huggingface.co/tiiuae/falcon-7b/resolve/main/config.json", } class _A ( _lowerCAmelCase ): lowercase__: Optional[int] = "falcon" lowercase__: str = ["past_key_values"] def __init__( self : Optional[Any] , __magic_name__ : str=6_50_24 , __magic_name__ : str=45_44 , __magic_name__ : Union[str, Any]=32 , __magic_name__ : Any=71 , __magic_name__ : Optional[Any]=1E-5 , __magic_name__ : Dict=0.02 , __magic_name__ : Any=True , __magic_name__ : Tuple=0.0 , __magic_name__ : str=0.0 , __magic_name__ : str=None , __magic_name__ : List[str]=False , __magic_name__ : int=False , __magic_name__ : str=True , __magic_name__ : List[str]=True , __magic_name__ : List[str]=False , __magic_name__ : Optional[int]=11 , __magic_name__ : Optional[int]=11 , **__magic_name__ : Dict , ) -> Optional[Any]: """simple docstring""" __snake_case : List[Any] = vocab_size # Backward compatibility with n_embed kwarg __snake_case : Optional[Any] = kwargs.pop("""n_embed""" , __magic_name__ ) __snake_case : str = hidden_size if n_embed is None else n_embed __snake_case : List[str] = num_hidden_layers __snake_case : Optional[int] = num_attention_heads __snake_case : Tuple = layer_norm_epsilon __snake_case : List[str] = initializer_range __snake_case : Tuple = use_cache __snake_case : List[Any] = hidden_dropout __snake_case : Tuple = attention_dropout __snake_case : int = bos_token_id __snake_case : Any = eos_token_id __snake_case : Optional[int] = num_attention_heads if num_kv_heads is None else num_kv_heads __snake_case : Optional[Any] = alibi __snake_case : Any = new_decoder_architecture __snake_case : List[str] = multi_query # Ignored when new_decoder_architecture is True __snake_case : int = parallel_attn __snake_case : Union[str, Any] = bias super().__init__(bos_token_id=__magic_name__ , eos_token_id=__magic_name__ , **__magic_name__ ) @property def lowercase__ ( self : Dict ) -> List[Any]: """simple docstring""" return self.hidden_size // self.num_attention_heads @property def lowercase__ ( self : Any ) -> List[Any]: """simple docstring""" return not self.alibi
26
from ....configuration_utils import PretrainedConfig from ....utils import logging _snake_case : int = logging.get_logger(__name__) _snake_case : str = { "Visual-Attention-Network/van-base": ( "https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json" ), } class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[Any] = "van" def __init__( self : Optional[int] , lowerCamelCase : Any=224 , lowerCamelCase : str=3 , lowerCamelCase : Any=[7, 3, 3, 3] , lowerCamelCase : Dict=[4, 2, 2, 2] , lowerCamelCase : List[Any]=[64, 128, 320, 512] , lowerCamelCase : str=[3, 3, 12, 3] , lowerCamelCase : Dict=[8, 8, 4, 4] , lowerCamelCase : Any="gelu" , lowerCamelCase : Optional[int]=0.02 , lowerCamelCase : Tuple=1E-6 , lowerCamelCase : Optional[int]=1E-2 , lowerCamelCase : int=0.0 , lowerCamelCase : Optional[Any]=0.0 , **lowerCamelCase : Optional[int] , ) -> int: super().__init__(**lowerCamelCase ) __snake_case : Union[str, Any] = image_size __snake_case : Any = num_channels __snake_case : Any = patch_sizes __snake_case : List[Any] = strides __snake_case : str = hidden_sizes __snake_case : Dict = depths __snake_case : Optional[int] = mlp_ratios __snake_case : Dict = hidden_act __snake_case : Union[str, Any] = initializer_range __snake_case : List[str] = layer_norm_eps __snake_case : Optional[int] = layer_scale_init_value __snake_case : List[Any] = drop_path_rate __snake_case : int = dropout_rate
81
0
'''simple docstring''' import numpy as np UpperCAmelCase = [ ["a", "b", "c", "d", "e"], ["f", "g", "h", "i", "k"], ["l", "m", "n", "o", "p"], ["q", "r", "s", "t", "u"], ["v", "w", "x", "y", "z"], ] class lowerCAmelCase : def __init__( self : Optional[int] ): """simple docstring""" __lowercase =np.array(__lowercase ) def snake_case ( self : Union[str, Any] , __lowercase : str ): """simple docstring""" __lowercase =np.where(letter == self.SQUARE ) __lowercase =np.concatenate([indexa + 1, indexa + 1] ) return indexes def snake_case ( self : Union[str, Any] , __lowercase : int , __lowercase : int ): """simple docstring""" __lowercase =self.SQUARE[indexa - 1, indexa - 1] return letter def snake_case ( self : Union[str, Any] , __lowercase : str ): """simple docstring""" __lowercase =message.lower() __lowercase =message.replace(' ' , '' ) __lowercase =message.replace('j' , 'i' ) __lowercase =np.empty((2, len(__lowercase )) ) for letter_index in range(len(__lowercase ) ): __lowercase =self.letter_to_numbers(message[letter_index] ) __lowercase =numbers[0] __lowercase =numbers[1] __lowercase =first_step.reshape(2 * len(__lowercase ) ) __lowercase ="" for numbers_index in range(len(__lowercase ) ): __lowercase =int(second_step[numbers_index * 2] ) __lowercase =int(second_step[(numbers_index * 2) + 1] ) __lowercase =self.numbers_to_letter(__lowercase , __lowercase ) __lowercase =encoded_message + letter return encoded_message def snake_case ( self : Any , __lowercase : str ): """simple docstring""" __lowercase =message.lower() message.replace(' ' , '' ) __lowercase =np.empty(2 * len(__lowercase ) ) for letter_index in range(len(__lowercase ) ): __lowercase =self.letter_to_numbers(message[letter_index] ) __lowercase =numbers[0] __lowercase =numbers[1] __lowercase =first_step.reshape((2, len(__lowercase )) ) __lowercase ="" for numbers_index in range(len(__lowercase ) ): __lowercase =int(second_step[0, numbers_index] ) __lowercase =int(second_step[1, numbers_index] ) __lowercase =self.numbers_to_letter(__lowercase , __lowercase ) __lowercase =decoded_message + letter return decoded_message
119
import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer _snake_case : Union[str, Any] = logging.getLogger(__name__) def lowerCAmelCase_ ( ): __snake_case : int = argparse.ArgumentParser( description="Prepare TFRecord shards from pre-tokenized samples of the wikitext dataset." ) parser.add_argument( "--dataset_name" , type=__lowerCamelCase , default="wikitext" , help="Name of the training. Explore datasets at: hf.co/datasets." , ) parser.add_argument( "--dataset_config" , type=__lowerCamelCase , default="wikitext-103-raw-v1" , help="Configuration name of the dataset." ) parser.add_argument( "--tokenizer_name_or_path" , type=__lowerCamelCase , default="sayakpaul/unigram-tokenizer-wikitext" , help="Tokenizer identifier. Can be a local filepath or a Hub identifier." , ) parser.add_argument( "--shard_size" , type=__lowerCamelCase , default=1_0_0_0 , help="Number of entries to go in a single shard." , ) parser.add_argument("--split" , type=__lowerCamelCase , default="train" , choices=["train", "test", "validation"] ) parser.add_argument( "--limit" , default=__lowerCamelCase , type=__lowerCamelCase , help="Limit the number of shards (used for debugging)." , ) parser.add_argument( "--max_length" , type=__lowerCamelCase , default=5_1_2 , help="Maximum sequence length. For training on TPUs, it helps to have a maximum" " sequence length that is a multiple of 8." , ) parser.add_argument( "--output_dir" , default="tf-tpu" , type=__lowerCamelCase , help="Output directory where the TFRecord shards will be saved. If the" " path is appended with `gs://` ('gs://tf-tpu', for example) then the TFRecord" " shards will be directly saved to a Google Cloud Storage bucket." , ) __snake_case : List[str] = parser.parse_args() return args def lowerCAmelCase_ ( __lowerCamelCase ): def fn(__lowerCamelCase ): return tokenizer(examples["text"] ) return fn def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = [] for i in range(len(tokenized_data["input_ids"] ) ): __snake_case : Tuple = { "input_ids": tf.train.Feature(intaa_list=tf.train.IntaaList(value=tokenized_data["input_ids"][i] ) ), "attention_mask": tf.train.Feature( intaa_list=tf.train.IntaaList(value=tokenized_data["attention_mask"][i] ) ), } __snake_case : List[Any] = tf.train.Features(feature=__lowerCamelCase ) __snake_case : str = tf.train.Example(features=__lowerCamelCase ) __snake_case : List[str] = example.SerializeToString() records.append(__lowerCamelCase ) return records def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Optional[int] = datasets.load_dataset(args.dataset_name , args.dataset_config , split=args.split ) if args.limit is not None: __snake_case : Optional[Any] = min(len(__lowerCamelCase ) , args.limit ) __snake_case : Dict = dataset.select(range(__lowerCamelCase ) ) print(F'Limiting the dataset to {args.limit} entries.' ) __snake_case : Dict = AutoTokenizer.from_pretrained(args.tokenizer_name_or_path ) # Handle output directory creation. # For serializing into a Google Cloud Storage Bucket, one needs to first # create a bucket. if "gs" not in args.output_dir: if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) __snake_case : Dict = os.path.join(args.output_dir , args.split ) if not os.path.exists(__lowerCamelCase ): os.makedirs(__lowerCamelCase ) else: __snake_case : str = os.path.join(args.output_dir , args.split ) # Tokenize the whole dataset at once. __snake_case : Any = tokenize_function(__lowerCamelCase ) __snake_case : Optional[Any] = dataset.map(__lowerCamelCase , batched=__lowerCamelCase , num_proc=4 , remove_columns=["text"] ) # We need to concatenate all our texts together, and then split the result # into chunks of a fixed size, which we will call block_size. To do this, we # will use the map method again, with the option batched=True. When we use batched=True, # the function we pass to map() will be passed multiple inputs at once, allowing us # to group them into more or fewer examples than we had in the input. # This allows us to create our new fixed-length samples. The advantage of this # method is that we don't lose a whole lot of content from the dataset compared to the # case where we simply tokenize with a pre-defined max_length. def group_texts(__lowerCamelCase ): # Concatenate all texts. __snake_case : List[str] = {k: sum(examples[k] , [] ) for k in examples.keys()} __snake_case : List[Any] = len(concatenated_examples[list(examples.keys() )[0]] ) # We drop the small remainder, though you could add padding instead if the model supports it # In this, as in all things, we advise you to follow your heart 🫀 __snake_case : Any = (total_length // args.max_length) * args.max_length # Split by chunks of max_len. __snake_case : int = { k: [t[i : i + args.max_length] for i in range(0 , __lowerCamelCase , args.max_length )] for k, t in concatenated_examples.items() } return result __snake_case : Any = dataset_tokenized.map(__lowerCamelCase , batched=__lowerCamelCase , batch_size=1_0_0_0 , num_proc=4 ) __snake_case : Optional[Any] = 0 __snake_case : Optional[Any] = 0 for shard in range(0 , len(__lowerCamelCase ) , args.shard_size ): __snake_case : List[str] = grouped_dataset[shard : shard + args.shard_size] __snake_case : Any = len(dataset_snapshot["input_ids"] ) __snake_case : List[Any] = os.path.join(__lowerCamelCase , F'dataset-{shard_count}-{records_containing}.tfrecord' ) __snake_case : Optional[Any] = get_serialized_examples(__lowerCamelCase ) with tf.io.TFRecordWriter(__lowerCamelCase ) as out_file: for i in range(len(__lowerCamelCase ) ): __snake_case : Union[str, Any] = serialized_examples[i] out_file.write(__lowerCamelCase ) print("Wrote file {} containing {} records".format(__lowerCamelCase , __lowerCamelCase ) ) shard_count += 1 total_records += records_containing with open(F'split-{args.split}-records-count.txt' , "w" ) as f: print(F'Total {args.split} records: {total_records}' , file=__lowerCamelCase ) if __name__ == "__main__": _snake_case : List[Any] = parse_args() main(args)
81
0
import os import re from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging SCREAMING_SNAKE_CASE__ : Dict = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : Optional[Any] = { "vocab_file": "vocab.txt", "merges_file": "bpe.codes", } SCREAMING_SNAKE_CASE__ : str = { "vocab_file": { "vinai/phobert-base": "https://huggingface.co/vinai/phobert-base/resolve/main/vocab.txt", "vinai/phobert-large": "https://huggingface.co/vinai/phobert-large/resolve/main/vocab.txt", }, "merges_file": { "vinai/phobert-base": "https://huggingface.co/vinai/phobert-base/resolve/main/bpe.codes", "vinai/phobert-large": "https://huggingface.co/vinai/phobert-large/resolve/main/bpe.codes", }, } SCREAMING_SNAKE_CASE__ : int = { "vinai/phobert-base": 256, "vinai/phobert-large": 256, } def __magic_name__ ( __lowerCAmelCase : List[Any] ) -> int: __lowerCamelCase = set() __lowerCamelCase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __lowerCamelCase = char __lowerCamelCase = set(__lowerCamelCase ) return pairs class lowerCAmelCase__ ( _lowerCAmelCase ): a__ : int = VOCAB_FILES_NAMES a__ : Optional[int] = PRETRAINED_VOCAB_FILES_MAP a__ : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : List[str] , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Union[str, Any]="<s>" , SCREAMING_SNAKE_CASE__ : Optional[Any]="</s>" , SCREAMING_SNAKE_CASE__ : Union[str, Any]="</s>" , SCREAMING_SNAKE_CASE__ : Any="<s>" , SCREAMING_SNAKE_CASE__ : List[Any]="<unk>" , SCREAMING_SNAKE_CASE__ : int="<pad>" , SCREAMING_SNAKE_CASE__ : Optional[Any]="<mask>" , **SCREAMING_SNAKE_CASE__ : List[str] , ) -> str: super().__init__( bos_token=SCREAMING_SNAKE_CASE__ , eos_token=SCREAMING_SNAKE_CASE__ , unk_token=SCREAMING_SNAKE_CASE__ , sep_token=SCREAMING_SNAKE_CASE__ , cls_token=SCREAMING_SNAKE_CASE__ , pad_token=SCREAMING_SNAKE_CASE__ , mask_token=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ , ) __lowerCamelCase = vocab_file __lowerCamelCase = merges_file __lowerCamelCase = {} __lowerCamelCase = 0 __lowerCamelCase = 1 __lowerCamelCase = 2 __lowerCamelCase = 3 self.add_from_file(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = {v: k for k, v in self.encoder.items()} with open(SCREAMING_SNAKE_CASE__ , encoding='''utf-8''' ) as merges_handle: __lowerCamelCase = merges_handle.read().split('''\n''' )[:-1] __lowerCamelCase = [tuple(merge.split()[:-1] ) for merge in merges] __lowerCamelCase = dict(zip(SCREAMING_SNAKE_CASE__ , range(len(SCREAMING_SNAKE_CASE__ ) ) ) ) __lowerCamelCase = {} def __A ( self : Tuple , SCREAMING_SNAKE_CASE__ : List[int] , SCREAMING_SNAKE_CASE__ : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] __lowerCamelCase = [self.cls_token_id] __lowerCamelCase = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __A ( self : Optional[int] , SCREAMING_SNAKE_CASE__ : List[int] , SCREAMING_SNAKE_CASE__ : Optional[List[int]] = None , SCREAMING_SNAKE_CASE__ : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=SCREAMING_SNAKE_CASE__ , token_ids_a=SCREAMING_SNAKE_CASE__ , already_has_special_tokens=SCREAMING_SNAKE_CASE__ ) if token_ids_a is None: return [1] + ([0] * len(SCREAMING_SNAKE_CASE__ )) + [1] return [1] + ([0] * len(SCREAMING_SNAKE_CASE__ )) + [1, 1] + ([0] * len(SCREAMING_SNAKE_CASE__ )) + [1] def __A ( self : List[str] , SCREAMING_SNAKE_CASE__ : List[int] , SCREAMING_SNAKE_CASE__ : Optional[List[int]] = None ) -> List[int]: __lowerCamelCase = [self.sep_token_id] __lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __A ( self : Optional[int] ) -> Union[str, Any]: return len(self.encoder ) def __A ( self : Dict ) -> int: return dict(self.encoder , **self.added_tokens_encoder ) def __A ( self : List[Any] , SCREAMING_SNAKE_CASE__ : Any ) -> Dict: if token in self.cache: return self.cache[token] __lowerCamelCase = tuple(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = tuple(list(word[:-1] ) + [word[-1] + '''</w>'''] ) __lowerCamelCase = get_pairs(SCREAMING_SNAKE_CASE__ ) if not pairs: return token while True: __lowerCamelCase = min(SCREAMING_SNAKE_CASE__ , key=lambda SCREAMING_SNAKE_CASE__ : self.bpe_ranks.get(SCREAMING_SNAKE_CASE__ , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break __lowerCamelCase = bigram __lowerCamelCase = [] __lowerCamelCase = 0 while i < len(SCREAMING_SNAKE_CASE__ ): try: __lowerCamelCase = word.index(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) __lowerCamelCase = j if word[i] == first and i < len(SCREAMING_SNAKE_CASE__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __lowerCamelCase = tuple(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = new_word if len(SCREAMING_SNAKE_CASE__ ) == 1: break else: __lowerCamelCase = get_pairs(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = "@@ ".join(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = word[:-4] __lowerCamelCase = word return word def __A ( self : int , SCREAMING_SNAKE_CASE__ : Union[str, Any] ) -> Optional[int]: __lowerCamelCase = [] __lowerCamelCase = re.findall(R'''\S+\n?''' , SCREAMING_SNAKE_CASE__ ) for token in words: split_tokens.extend(list(self.bpe(SCREAMING_SNAKE_CASE__ ).split(''' ''' ) ) ) return split_tokens def __A ( self : int , SCREAMING_SNAKE_CASE__ : List[Any] ) -> List[Any]: return self.encoder.get(SCREAMING_SNAKE_CASE__ , self.encoder.get(self.unk_token ) ) def __A ( self : List[str] , SCREAMING_SNAKE_CASE__ : Tuple ) -> int: return self.decoder.get(SCREAMING_SNAKE_CASE__ , self.unk_token ) def __A ( self : Optional[Any] , SCREAMING_SNAKE_CASE__ : Optional[Any] ) -> str: __lowerCamelCase = " ".join(SCREAMING_SNAKE_CASE__ ).replace('''@@ ''' , '''''' ).strip() return out_string def __A ( self : List[str] , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(SCREAMING_SNAKE_CASE__ ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return __lowerCamelCase = os.path.join( SCREAMING_SNAKE_CASE__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) __lowerCamelCase = os.path.join( SCREAMING_SNAKE_CASE__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(SCREAMING_SNAKE_CASE__ ): copyfile(self.vocab_file , SCREAMING_SNAKE_CASE__ ) if os.path.abspath(self.merges_file ) != os.path.abspath(SCREAMING_SNAKE_CASE__ ): copyfile(self.merges_file , SCREAMING_SNAKE_CASE__ ) return out_vocab_file, out_merge_file def __A ( self : Optional[int] , SCREAMING_SNAKE_CASE__ : Optional[int] ) -> int: if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): try: with open(SCREAMING_SNAKE_CASE__ , '''r''' , encoding='''utf-8''' ) as fd: self.add_from_file(SCREAMING_SNAKE_CASE__ ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception(f'''Incorrect encoding detected in {f}, please rebuild the dataset''' ) return __lowerCamelCase = f.readlines() for lineTmp in lines: __lowerCamelCase = lineTmp.strip() __lowerCamelCase = line.rfind(''' ''' ) if idx == -1: raise ValueError('''Incorrect dictionary format, expected \'<token> <cnt>\'''' ) __lowerCamelCase = line[:idx] __lowerCamelCase = len(self.encoder )
298
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) _snake_case : List[Any] = "\\n Text data.\n Second line of data." _snake_case : Tuple = "file" @pytest.fixture(scope="session" ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / (FILE_PATH + ".zstd") __snake_case : Optional[Any] = bytes(__lowerCamelCase , "utf-8" ) with zstd.open(__lowerCamelCase , "wb" ) as f: f.write(__lowerCamelCase ) return path @pytest.fixture def lowerCAmelCase_ ( __lowerCamelCase ): with open(os.path.join(tmpfs.local_root_dir , __lowerCamelCase ) , "w" ) as f: f.write(__lowerCamelCase ) return FILE_PATH @pytest.mark.parametrize("compression_format" , ["gzip", "xz", "zstd"] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Optional[int] = {"gzip": gz_file, "xz": xz_file, "zstd": zstd_path} __snake_case : str = input_paths[compression_format] __snake_case : Optional[Any] = tmp_path / "cache" __snake_case : Optional[int] = DownloadConfig(cache_dir=__lowerCamelCase , extract_compressed_file=__lowerCamelCase ) __snake_case : Union[str, Any] = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) with open(__lowerCamelCase ) as f: __snake_case : Dict = f.read() with open(__lowerCamelCase ) as f: __snake_case : Tuple = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("default_extracted" , [True, False] ) @pytest.mark.parametrize("default_cache_dir" , [True, False] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Tuple = "custom_cache" __snake_case : List[str] = "custom_extracted_dir" __snake_case : Any = tmp_path / "custom_extracted_path" if default_extracted: __snake_case : List[Any] = ("downloads" if default_cache_dir else custom_cache_dir, "extracted") else: monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_DIR" , __lowerCamelCase ) monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_PATH" , str(__lowerCamelCase ) ) __snake_case : Optional[Any] = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) __snake_case : Optional[int] = xz_file __snake_case : Optional[int] = ( DownloadConfig(extract_compressed_file=__lowerCamelCase ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=__lowerCamelCase ) ) __snake_case : str = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) assert Path(__lowerCamelCase ).parent.parts[-2:] == expected def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : Optional[Any] = str(Path(__lowerCamelCase ).resolve() ) assert cached_path(__lowerCamelCase ) == text_file # relative path __snake_case : Any = str(Path(__lowerCamelCase ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(__lowerCamelCase ) == text_file def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : List[Any] = str(tmp_path.resolve() / "__missing_file__.txt" ) with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) # relative path __snake_case : Optional[int] = "./__missing_file__.txt" with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : str = get_from_cache(F'tmp://{tmpfs_file}' ) with open(__lowerCamelCase ) as f: __snake_case : Union[str, Any] = f.read() assert output_file_content == FILE_CONTENT @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( ): with pytest.raises(__lowerCamelCase ): cached_path("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[Any] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): http_get("https://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): http_head("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[str] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): ftp_get("ftp://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): ftp_head("ftp://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): fsspec_get("s3://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): fsspec_head("s3://huggingface.co" )
81
0
"""simple docstring""" from __future__ import annotations import copy import inspect import json import math import os import tempfile import unittest from importlib import import_module import numpy as np from transformers import ViTMAEConfig from transformers.file_utils import cached_property, is_tf_available, is_vision_available from transformers.testing_utils import require_tf, require_vision, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTMAEForPreTraining, TFViTMAEModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class UpperCamelCase : def __init__( self ,__UpperCamelCase ,__UpperCamelCase=13 ,__UpperCamelCase=30 ,__UpperCamelCase=2 ,__UpperCamelCase=3 ,__UpperCamelCase=True ,__UpperCamelCase=True ,__UpperCamelCase=32 ,__UpperCamelCase=2 ,__UpperCamelCase=4 ,__UpperCamelCase=37 ,__UpperCamelCase="gelu" ,__UpperCamelCase=0.1 ,__UpperCamelCase=0.1 ,__UpperCamelCase=10 ,__UpperCamelCase=0.02 ,__UpperCamelCase=3 ,__UpperCamelCase=0.6 ,__UpperCamelCase=None ,) -> Union[str, Any]: '''simple docstring''' lowercase_ : List[str] = parent lowercase_ : List[str] = batch_size lowercase_ : Optional[int] = image_size lowercase_ : List[Any] = patch_size lowercase_ : Union[str, Any] = num_channels lowercase_ : Any = is_training lowercase_ : List[str] = use_labels lowercase_ : Optional[Any] = hidden_size lowercase_ : int = num_hidden_layers lowercase_ : List[str] = num_attention_heads lowercase_ : Tuple = intermediate_size lowercase_ : Tuple = hidden_act lowercase_ : Dict = hidden_dropout_prob lowercase_ : str = attention_probs_dropout_prob lowercase_ : Any = type_sequence_label_size lowercase_ : str = initializer_range lowercase_ : int = mask_ratio lowercase_ : int = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) lowercase_ : Optional[int] = (image_size // patch_size) ** 2 lowercase_ : Union[str, Any] = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def _UpperCAmelCase ( self ) -> str: '''simple docstring''' lowercase_ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase_ : List[Any] = None if self.use_labels: lowercase_ : Optional[Any] = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) lowercase_ : Tuple = self.get_config() return config, pixel_values, labels def _UpperCAmelCase ( self ) -> Union[str, Any]: '''simple docstring''' return ViTMAEConfig( image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,decoder_hidden_size=self.hidden_size ,decoder_num_hidden_layers=self.num_hidden_layers ,decoder_num_attention_heads=self.num_attention_heads ,decoder_intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,is_decoder=__UpperCamelCase ,initializer_range=self.initializer_range ,mask_ratio=self.mask_ratio ,) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> List[Any]: '''simple docstring''' lowercase_ : List[Any] = TFViTMAEModel(config=__UpperCamelCase ) lowercase_ : Optional[Any] = model(__UpperCamelCase ,training=__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> Union[str, Any]: '''simple docstring''' lowercase_ : Tuple = TFViTMAEForPreTraining(__UpperCamelCase ) lowercase_ : Tuple = model(__UpperCamelCase ,training=__UpperCamelCase ) # expected sequence length = num_patches lowercase_ : Union[str, Any] = (self.image_size // self.patch_size) ** 2 lowercase_ : Union[str, Any] = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape ,(self.batch_size, num_patches, expected_num_channels) ) # test greyscale images lowercase_ : Optional[int] = 1 lowercase_ : Optional[int] = TFViTMAEForPreTraining(__UpperCamelCase ) lowercase_ : List[str] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowercase_ : List[Any] = model(__UpperCamelCase ,training=__UpperCamelCase ) lowercase_ : Dict = self.patch_size**2 self.parent.assertEqual(result.logits.shape ,(self.batch_size, num_patches, expected_num_channels) ) def _UpperCAmelCase ( self ) -> int: '''simple docstring''' lowercase_ : List[str] = self.prepare_config_and_inputs() (lowercase_) : Optional[Any] = config_and_inputs lowercase_ : Optional[int] = {"pixel_values": pixel_values} return config, inputs_dict @require_tf class UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): lowercase = (TFViTMAEModel, TFViTMAEForPreTraining) if is_tf_available() else () lowercase = {"feature-extraction": TFViTMAEModel} if is_tf_available() else {} lowercase = False lowercase = False lowercase = False lowercase = False def _UpperCAmelCase ( self ) -> Tuple: '''simple docstring''' lowercase_ : Any = TFViTMAEModelTester(self ) lowercase_ : List[str] = ConfigTester(self ,config_class=__UpperCamelCase ,has_text_modality=__UpperCamelCase ,hidden_size=37 ) def _UpperCAmelCase ( self ) -> str: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='ViTMAE does not use inputs_embeds' ) def _UpperCAmelCase ( self ) -> int: '''simple docstring''' pass def _UpperCAmelCase ( self ) -> Union[str, Any]: '''simple docstring''' lowercase_ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase_ : int = model_class(__UpperCamelCase ) self.assertIsInstance(model.get_input_embeddings() ,(tf.keras.layers.Layer) ) lowercase_ : Optional[Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__UpperCamelCase ,tf.keras.layers.Layer ) ) def _UpperCAmelCase ( self ) -> Tuple: '''simple docstring''' lowercase_ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase_ : List[str] = model_class(__UpperCamelCase ) lowercase_ : str = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase_ : Tuple = [*signature.parameters.keys()] lowercase_ : Optional[Any] = ["pixel_values"] self.assertListEqual(arg_names[:1] ,__UpperCamelCase ) def _UpperCAmelCase ( self ) -> int: '''simple docstring''' lowercase_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def _UpperCAmelCase ( self ) -> Optional[int]: '''simple docstring''' lowercase_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*__UpperCamelCase ) def _UpperCAmelCase ( self ) -> Dict: '''simple docstring''' np.random.seed(2 ) lowercase_ : int = self.model_tester.prepare_config_and_inputs_for_common() lowercase_ : List[Any] = int((config.image_size // config.patch_size) ** 2 ) lowercase_ : str = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: lowercase_ : Any = model_class(__UpperCamelCase ) lowercase_ : List[str] = self._prepare_for_class(__UpperCamelCase ,__UpperCamelCase ) lowercase_ : int = model(__UpperCamelCase ,noise=__UpperCamelCase ) lowercase_ : List[Any] = copy.deepcopy(self._prepare_for_class(__UpperCamelCase ,__UpperCamelCase ) ) lowercase_ : List[str] = model(**__UpperCamelCase ,noise=__UpperCamelCase ) lowercase_ : str = outputs_dict[0].numpy() lowercase_ : Tuple = outputs_keywords[0].numpy() self.assertLess(np.sum(np.abs(output_dict - output_keywords ) ) ,1e-6 ) def _UpperCAmelCase ( self ) -> List[str]: '''simple docstring''' np.random.seed(2 ) lowercase_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() lowercase_ : Tuple = int((config.image_size // config.patch_size) ** 2 ) lowercase_ : List[Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) def prepare_numpy_arrays(__UpperCamelCase ): lowercase_ : Any = {} for k, v in inputs_dict.items(): if tf.is_tensor(__UpperCamelCase ): lowercase_ : Any = v.numpy() else: lowercase_ : Optional[int] = np.array(__UpperCamelCase ) return inputs_np_dict for model_class in self.all_model_classes: lowercase_ : Optional[int] = model_class(__UpperCamelCase ) lowercase_ : List[str] = self._prepare_for_class(__UpperCamelCase ,__UpperCamelCase ) lowercase_ : Any = prepare_numpy_arrays(__UpperCamelCase ) lowercase_ : Tuple = model(__UpperCamelCase ,noise=__UpperCamelCase ) lowercase_ : Tuple = model(**__UpperCamelCase ,noise=__UpperCamelCase ) self.assert_outputs_same(__UpperCamelCase ,__UpperCamelCase ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> Optional[Any]: '''simple docstring''' np.random.seed(2 ) lowercase_ : Optional[int] = int((tf_model.config.image_size // tf_model.config.patch_size) ** 2 ) lowercase_ : Tuple = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) lowercase_ : Tuple = tf.constant(__UpperCamelCase ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument lowercase_ : Optional[int] = tf_noise super().check_pt_tf_models(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) def _UpperCAmelCase ( self ) -> Any: '''simple docstring''' np.random.seed(2 ) lowercase_ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() lowercase_ : int = { module_member for model_class in self.all_model_classes for module in (import_module(model_class.__module__ ),) for module_member_name in dir(__UpperCamelCase ) if module_member_name.endswith('MainLayer' ) # This condition is required, since `modeling_tf_clip.py` has 3 classes whose names end with `MainLayer`. and module_member_name[: -len('MainLayer' )] == model_class.__name__[: -len('Model' )] for module_member in (getattr(__UpperCamelCase ,__UpperCamelCase ),) if isinstance(__UpperCamelCase ,__UpperCamelCase ) and tf.keras.layers.Layer in module_member.__bases__ and getattr(__UpperCamelCase ,'_keras_serializable' ,__UpperCamelCase ) } lowercase_ : List[str] = int((config.image_size // config.patch_size) ** 2 ) lowercase_ : Optional[Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) lowercase_ : List[str] = tf.convert_to_tensor(__UpperCamelCase ) inputs_dict.update({'noise': noise} ) for main_layer_class in tf_main_layer_classes: lowercase_ : Optional[int] = main_layer_class(__UpperCamelCase ) lowercase_ : str = { name: tf.keras.Input(tensor.shape[1:] ,dtype=tensor.dtype ) for name, tensor in inputs_dict.items() } lowercase_ : Optional[int] = tf.keras.Model(__UpperCamelCase ,outputs=main_layer(__UpperCamelCase ) ) lowercase_ : List[str] = model(__UpperCamelCase ) with tempfile.TemporaryDirectory() as tmpdirname: lowercase_ : List[Any] = os.path.join(__UpperCamelCase ,'keras_model.h5' ) model.save(__UpperCamelCase ) lowercase_ : Optional[int] = tf.keras.models.load_model( __UpperCamelCase ,custom_objects={main_layer_class.__name__: main_layer_class} ) assert isinstance(__UpperCamelCase ,tf.keras.Model ) lowercase_ : Any = model(__UpperCamelCase ) self.assert_outputs_same(__UpperCamelCase ,__UpperCamelCase ) @slow def _UpperCAmelCase ( self ) -> int: '''simple docstring''' np.random.seed(2 ) lowercase_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() lowercase_ : Dict = int((config.image_size // config.patch_size) ** 2 ) lowercase_ : Dict = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: lowercase_ : Any = model_class(__UpperCamelCase ) lowercase_ : Dict = self._prepare_for_class(__UpperCamelCase ,__UpperCamelCase ) lowercase_ : str = model(__UpperCamelCase ,noise=__UpperCamelCase ) if model_class.__name__ == "TFViTMAEModel": lowercase_ : Tuple = outputs.last_hidden_state.numpy() lowercase_ : Dict = 0 else: lowercase_ : List[Any] = outputs.logits.numpy() lowercase_ : Tuple = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(__UpperCamelCase ,saved_model=__UpperCamelCase ) lowercase_ : Union[str, Any] = model_class.from_pretrained(__UpperCamelCase ) lowercase_ : Dict = model(__UpperCamelCase ,noise=__UpperCamelCase ) if model_class.__name__ == "TFViTMAEModel": lowercase_ : Tuple = after_outputs["last_hidden_state"].numpy() lowercase_ : List[Any] = 0 else: lowercase_ : int = after_outputs["logits"].numpy() lowercase_ : str = 0 lowercase_ : List[str] = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(__UpperCamelCase ,1e-5 ) def _UpperCAmelCase ( self ) -> Tuple: '''simple docstring''' np.random.seed(2 ) lowercase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowercase_ : Union[str, Any] = int((config.image_size // config.patch_size) ** 2 ) lowercase_ : str = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: lowercase_ : Tuple = model_class(__UpperCamelCase ) lowercase_ : List[Any] = self._prepare_for_class(__UpperCamelCase ,__UpperCamelCase ) lowercase_ : Any = model(__UpperCamelCase ,noise=__UpperCamelCase ) lowercase_ : Union[str, Any] = model.get_config() # make sure that returned config is jsonifiable, which is required by keras json.dumps(__UpperCamelCase ) lowercase_ : Any = model_class.from_config(model.get_config() ) # make sure it also accepts a normal config lowercase_ : Optional[int] = model_class.from_config(model.config ) lowercase_ : Any = new_model(__UpperCamelCase ) # Build model new_model.set_weights(model.get_weights() ) lowercase_ : str = new_model(__UpperCamelCase ,noise=__UpperCamelCase ) self.assert_outputs_same(__UpperCamelCase ,__UpperCamelCase ) @unittest.skip( reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.' ) def _UpperCAmelCase ( self ) -> Optional[Any]: '''simple docstring''' pass @unittest.skip(reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load' ) def _UpperCAmelCase ( self ) -> List[Any]: '''simple docstring''' pass @slow def _UpperCAmelCase ( self ) -> Tuple: '''simple docstring''' lowercase_ : str = TFViTMAEModel.from_pretrained('google/vit-base-patch16-224' ) self.assertIsNotNone(__UpperCamelCase ) def lowercase__( ): lowercase_ : Dict = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class UpperCamelCase ( unittest.TestCase ): @cached_property def _UpperCAmelCase ( self ) -> str: '''simple docstring''' return ViTImageProcessor.from_pretrained('facebook/vit-mae-base' ) if is_vision_available() else None @slow def _UpperCAmelCase ( self ) -> Tuple: '''simple docstring''' np.random.seed(2 ) lowercase_ : Any = TFViTMAEForPreTraining.from_pretrained('facebook/vit-mae-base' ) lowercase_ : Tuple = self.default_image_processor lowercase_ : List[str] = prepare_img() lowercase_ : Optional[Any] = image_processor(images=__UpperCamelCase ,return_tensors='tf' ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) lowercase_ : List[Any] = ViTMAEConfig() lowercase_ : str = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) lowercase_ : Tuple = np.random.uniform(size=(1, num_patches) ) # forward pass lowercase_ : str = model(**__UpperCamelCase ,noise=__UpperCamelCase ) # verify the logits lowercase_ : Union[str, Any] = tf.convert_to_tensor([1, 196, 768] ) self.assertEqual(outputs.logits.shape ,__UpperCamelCase ) lowercase_ : List[str] = tf.convert_to_tensor( [[-0.0548, -1.7023, -0.9325], [0.3721, -0.5670, -0.2233], [0.8235, -1.3878, -0.3524]] ) tf.debugging.assert_near(outputs.logits[0, :3, :3] ,__UpperCamelCase ,atol=1e-4 )
425
_snake_case : Optional[int] = {"a": ["c", "b"], "b": ["d", "e"], "c": [], "d": [], "e": []} _snake_case : Dict = ["a", "b", "c", "d", "e"] def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : List[str] = start # add current to visited visited.append(__lowerCamelCase ) __snake_case : List[Any] = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: __snake_case : Tuple = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # if all neighbors visited add current to sort sort.append(__lowerCamelCase ) # if all vertices haven't been visited select a new one to visit if len(__lowerCamelCase ) != len(__lowerCamelCase ): for vertice in vertices: if vertice not in visited: __snake_case : int = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # return sort return sort if __name__ == "__main__": _snake_case : List[Any] = topological_sort("a", [], []) print(sort)
81
0
'''simple docstring''' from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) snake_case_ : List[str] = logging.get_logger(__name__) # pylint: disable=invalid-name snake_case_ : Optional[int] = "\n Examples:\n ```py\n >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline\n >>> import torch\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-prior\")\n >>> pipe_prior.to(\"cuda\")\n >>> prompt = \"red cat, 4k photo\"\n >>> out = pipe_prior(prompt)\n >>> image_emb = out.image_embeds\n >>> zero_image_emb = out.negative_image_embeds\n >>> pipe = KandinskyV22Pipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-decoder\")\n >>> pipe.to(\"cuda\")\n >>> image = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=50,\n ... ).images\n >>> image[0].save(\"cat.png\")\n ```\n" def __snake_case ( _UpperCAmelCase : Union[str, Any], _UpperCAmelCase : Tuple, _UpperCAmelCase : Union[str, Any]=8): UpperCamelCase = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 UpperCamelCase = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class lowercase__ ( _lowerCAmelCase ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): '''simple docstring''' super().__init__() self.register_modules( unet=lowerCamelCase__ , scheduler=lowerCamelCase__ , movq=lowerCamelCase__ , ) UpperCamelCase = 2 ** (len(self.movq.config.block_out_channels ) - 1) def UpperCAmelCase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' if latents is None: UpperCamelCase = randn_tensor(lowerCamelCase__ , generator=lowerCamelCase__ , device=lowerCamelCase__ , dtype=lowerCamelCase__ ) else: if latents.shape != shape: raise ValueError(f'Unexpected latents shape, got {latents.shape}, expected {shape}' ) UpperCamelCase = latents.to(lowerCamelCase__ ) UpperCamelCase = latents * scheduler.init_noise_sigma return latents def UpperCAmelCase ( self , lowerCamelCase__=0 ): '''simple docstring''' if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('''Please install accelerate via `pip install accelerate`''' ) UpperCamelCase = torch.device(f'cuda:{gpu_id}' ) UpperCamelCase = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(lowerCamelCase__ , lowerCamelCase__ ) def UpperCAmelCase ( self , lowerCamelCase__=0 ): '''simple docstring''' if is_accelerate_available() and is_accelerate_version('''>=''' , '''0.17.0.dev0''' ): from accelerate import cpu_offload_with_hook else: raise ImportError('''`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.''' ) UpperCamelCase = torch.device(f'cuda:{gpu_id}' ) if self.device.type != "cpu": self.to('''cpu''' , silence_dtype_warnings=lowerCamelCase__ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) UpperCamelCase = None for cpu_offloaded_model in [self.unet, self.movq]: UpperCamelCase = cpu_offload_with_hook(lowerCamelCase__ , lowerCamelCase__ , prev_module_hook=lowerCamelCase__ ) # We'll offload the last model manually. UpperCamelCase = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCAmelCase ( self ): '''simple docstring''' if not hasattr(self.unet , '''_hf_hook''' ): return self.device for module in self.unet.modules(): if ( hasattr(lowerCamelCase__ , '''_hf_hook''' ) and hasattr(module._hf_hook , '''execution_device''' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(lowerCamelCase__ ) def __call__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = 5_1_2 , lowerCamelCase__ = 5_1_2 , lowerCamelCase__ = 1_0_0 , lowerCamelCase__ = 4.0 , lowerCamelCase__ = 1 , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = "pil" , lowerCamelCase__ = True , ): '''simple docstring''' UpperCamelCase = self._execution_device UpperCamelCase = guidance_scale > 1.0 if isinstance(lowerCamelCase__ , lowerCamelCase__ ): UpperCamelCase = torch.cat(lowerCamelCase__ , dim=0 ) UpperCamelCase = image_embeds.shape[0] * num_images_per_prompt if isinstance(lowerCamelCase__ , lowerCamelCase__ ): UpperCamelCase = torch.cat(lowerCamelCase__ , dim=0 ) if do_classifier_free_guidance: UpperCamelCase = image_embeds.repeat_interleave(lowerCamelCase__ , dim=0 ) UpperCamelCase = negative_image_embeds.repeat_interleave(lowerCamelCase__ , dim=0 ) UpperCamelCase = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=lowerCamelCase__ ) self.scheduler.set_timesteps(lowerCamelCase__ , device=lowerCamelCase__ ) UpperCamelCase = self.scheduler.timesteps UpperCamelCase = self.unet.config.in_channels UpperCamelCase = downscale_height_and_width(lowerCamelCase__ , lowerCamelCase__ , self.movq_scale_factor ) # create initial latent UpperCamelCase = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , self.scheduler , ) for i, t in enumerate(self.progress_bar(lowerCamelCase__ ) ): # expand the latents if we are doing classifier free guidance UpperCamelCase = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents UpperCamelCase = {"image_embeds": image_embeds} UpperCamelCase = self.unet( sample=lowerCamelCase__ , timestep=lowerCamelCase__ , encoder_hidden_states=lowerCamelCase__ , added_cond_kwargs=lowerCamelCase__ , return_dict=lowerCamelCase__ , )[0] if do_classifier_free_guidance: UpperCamelCase = noise_pred.split(latents.shape[1] , dim=1 ) UpperCamelCase = noise_pred.chunk(2 ) UpperCamelCase = variance_pred.chunk(2 ) UpperCamelCase = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) UpperCamelCase = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , '''variance_type''' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): UpperCamelCase = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 UpperCamelCase = self.scheduler.step( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , generator=lowerCamelCase__ , )[0] # post-processing UpperCamelCase = self.movq.decode(lowerCamelCase__ , force_not_quantize=lowerCamelCase__ )["sample"] if output_type not in ["pt", "np", "pil"]: raise ValueError(f'Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}' ) if output_type in ["np", "pil"]: UpperCamelCase = image * 0.5 + 0.5 UpperCamelCase = image.clamp(0 , 1 ) UpperCamelCase = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": UpperCamelCase = self.numpy_to_pil(lowerCamelCase__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCamelCase__ )
212
def lowerCAmelCase_ ( ): return [ a * b * (1_0_0_0 - a - b) for a in range(1 , 9_9_9 ) for b in range(__lowerCamelCase , 9_9_9 ) if (a * a + b * b == (1_0_0_0 - a - b) ** 2) ][0] if __name__ == "__main__": print(f'''{solution() = }''')
81
0
from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_pt_objects import * # noqa F403 else: from .scheduling_consistency_models import CMStochasticIterativeScheduler from .scheduling_ddim import DDIMScheduler from .scheduling_ddim_inverse import DDIMInverseScheduler from .scheduling_ddim_parallel import DDIMParallelScheduler from .scheduling_ddpm import DDPMScheduler from .scheduling_ddpm_parallel import DDPMParallelScheduler from .scheduling_deis_multistep import DEISMultistepScheduler from .scheduling_dpmsolver_multistep import DPMSolverMultistepScheduler from .scheduling_dpmsolver_multistep_inverse import DPMSolverMultistepInverseScheduler from .scheduling_dpmsolver_singlestep import DPMSolverSinglestepScheduler from .scheduling_euler_ancestral_discrete import EulerAncestralDiscreteScheduler from .scheduling_euler_discrete import EulerDiscreteScheduler from .scheduling_heun_discrete import HeunDiscreteScheduler from .scheduling_ipndm import IPNDMScheduler from .scheduling_k_dpm_2_ancestral_discrete import KDPMaAncestralDiscreteScheduler from .scheduling_k_dpm_2_discrete import KDPMaDiscreteScheduler from .scheduling_karras_ve import KarrasVeScheduler from .scheduling_pndm import PNDMScheduler from .scheduling_repaint import RePaintScheduler from .scheduling_sde_ve import ScoreSdeVeScheduler from .scheduling_sde_vp import ScoreSdeVpScheduler from .scheduling_unclip import UnCLIPScheduler from .scheduling_unipc_multistep import UniPCMultistepScheduler from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin from .scheduling_vq_diffusion import VQDiffusionScheduler try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_flax_objects import * # noqa F403 else: from .scheduling_ddim_flax import FlaxDDIMScheduler from .scheduling_ddpm_flax import FlaxDDPMScheduler from .scheduling_dpmsolver_multistep_flax import FlaxDPMSolverMultistepScheduler from .scheduling_karras_ve_flax import FlaxKarrasVeScheduler from .scheduling_lms_discrete_flax import FlaxLMSDiscreteScheduler from .scheduling_pndm_flax import FlaxPNDMScheduler from .scheduling_sde_ve_flax import FlaxScoreSdeVeScheduler from .scheduling_utils_flax import ( FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, broadcast_to_shape_from_left, ) try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .scheduling_lms_discrete import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .scheduling_dpmsolver_sde import DPMSolverSDEScheduler
454
from math import sqrt def lowerCAmelCase_ ( __lowerCamelCase = 1_0_0_0_0_0_0 ): __snake_case : int = 0 __snake_case : int = 0 __snake_case : int while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(__lowerCamelCase , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(f'''{solution() = }''')
81
0
"""simple docstring""" from __future__ import annotations from collections import namedtuple def _SCREAMING_SNAKE_CASE ( UpperCamelCase : Union[str, Any] , UpperCamelCase : str , UpperCamelCase : Optional[int] ): A__ = namedtuple("""result""" , """name value""" ) if (voltage, current, power).count(0 ) != 1: raise ValueError("""Only one argument must be 0""" ) elif power < 0: raise ValueError( """Power cannot be negative in any electrical/electronics system""" ) elif voltage == 0: return result("""voltage""" , power / current ) elif current == 0: return result("""current""" , power / voltage ) elif power == 0: return result("""power""" , float(round(abs(voltage * current ) , 2 ) ) ) else: raise ValueError("""Exactly one argument must be 0""" ) if __name__ == "__main__": import doctest doctest.testmod()
574
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class a (_lowerCAmelCase ): """simple docstring""" def __snake_case ( self : str ) -> str: __snake_case : Tuple = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCamelCase , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "neck_hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "num_attention_heads" ) ) class a : """simple docstring""" def __init__( self : Optional[int] , lowerCamelCase : List[str] , lowerCamelCase : Tuple=13 , lowerCamelCase : str=32 , lowerCamelCase : Dict=2 , lowerCamelCase : List[str]=3 , lowerCamelCase : Any=640 , lowerCamelCase : Optional[Any]=4 , lowerCamelCase : Tuple="silu" , lowerCamelCase : int=3 , lowerCamelCase : Dict=32 , lowerCamelCase : str=0.1 , lowerCamelCase : Optional[int]=0.1 , lowerCamelCase : Optional[Any]=0.1 , lowerCamelCase : Dict=0.02 , lowerCamelCase : Union[str, Any]=True , lowerCamelCase : Optional[int]=True , lowerCamelCase : Union[str, Any]=10 , lowerCamelCase : int=None , ) -> str: __snake_case : Optional[Any] = parent __snake_case : Optional[Any] = batch_size __snake_case : Any = image_size __snake_case : List[Any] = patch_size __snake_case : Any = num_channels __snake_case : Union[str, Any] = last_hidden_size __snake_case : Any = num_attention_heads __snake_case : Any = hidden_act __snake_case : Tuple = conv_kernel_size __snake_case : Any = output_stride __snake_case : Any = hidden_dropout_prob __snake_case : List[Any] = attention_probs_dropout_prob __snake_case : Optional[Any] = classifier_dropout_prob __snake_case : Union[str, Any] = use_labels __snake_case : Optional[int] = is_training __snake_case : Dict = num_labels __snake_case : Any = initializer_range __snake_case : Optional[int] = scope def __snake_case ( self : str ) -> Union[str, Any]: __snake_case : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : List[Any] = None __snake_case : Optional[int] = None if self.use_labels: __snake_case : Optional[int] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : Any = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __snake_case : Optional[Any] = self.get_config() return config, pixel_values, labels, pixel_labels def __snake_case ( self : Any ) -> Union[str, Any]: return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __snake_case ( self : Any , lowerCamelCase : Tuple , lowerCamelCase : int , lowerCamelCase : Tuple , lowerCamelCase : Optional[int] ) -> Dict: __snake_case : List[Any] = MobileViTModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[str] = model(lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[Any] , lowerCamelCase : List[str] , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] , lowerCamelCase : Tuple ) -> List[str]: __snake_case : str = self.num_labels __snake_case : List[Any] = MobileViTForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : Dict , lowerCamelCase : Any , lowerCamelCase : Any , lowerCamelCase : Optional[Any] , lowerCamelCase : Dict ) -> Dict: __snake_case : Union[str, Any] = self.num_labels __snake_case : Optional[int] = MobileViTForSemanticSegmentation(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Tuple = model(lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[int] ) -> List[Any]: __snake_case : Optional[Any] = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case , __snake_case : Union[str, Any] = config_and_inputs __snake_case : Dict = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) __UpperCAmelCase : Optional[Any] = ( { "feature-extraction": MobileViTModel, "image-classification": MobileViTForImageClassification, "image-segmentation": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) __UpperCAmelCase : List[str] = False __UpperCAmelCase : int = False __UpperCAmelCase : Optional[int] = False __UpperCAmelCase : Optional[int] = False def __snake_case ( self : Optional[int] ) -> Dict: __snake_case : Tuple = MobileViTModelTester(self ) __snake_case : Any = MobileViTConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase ) def __snake_case ( self : Optional[int] ) -> Dict: self.config_tester.run_common_tests() @unittest.skip(reason="MobileViT does not use inputs_embeds" ) def __snake_case ( self : Dict ) -> Any: pass @unittest.skip(reason="MobileViT does not support input and output embeddings" ) def __snake_case ( self : Dict ) -> List[Any]: pass @unittest.skip(reason="MobileViT does not output attentions" ) def __snake_case ( self : int ) -> Dict: pass def __snake_case ( self : int ) -> Union[str, Any]: __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Optional[Any] = model_class(lowerCamelCase ) __snake_case : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : int = [*signature.parameters.keys()] __snake_case : List[str] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __snake_case ( self : int ) -> Tuple: pass def __snake_case ( self : Any ) -> Tuple: __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def __snake_case ( self : Any ) -> str: def check_hidden_states_output(lowerCamelCase : Union[str, Any] , lowerCamelCase : Dict , lowerCamelCase : Any ): __snake_case : int = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : int = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : Union[str, Any] = outputs.hidden_states __snake_case : int = 5 self.assertEqual(len(lowerCamelCase ) , lowerCamelCase ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. __snake_case : List[Any] = 2 for i in range(len(lowerCamelCase ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) __snake_case , __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : str = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : List[Any] = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Any ) -> Any: __snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) def __snake_case ( self : List[str] ) -> List[str]: __snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowerCamelCase ) @slow def __snake_case ( self : List[str] ) -> Any: for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : List[str] = MobileViTModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def lowerCAmelCase_ ( ): __snake_case : Optional[int] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a (unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self : str ) -> Dict: return MobileViTImageProcessor.from_pretrained("apple/mobilevit-xx-small" ) if is_vision_available() else None @slow def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : Tuple = MobileViTForImageClassification.from_pretrained("apple/mobilevit-xx-small" ).to(lowerCamelCase ) __snake_case : Optional[Any] = self.default_image_processor __snake_case : Union[str, Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Dict = model(**lowerCamelCase ) # verify the logits __snake_case : Union[str, Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([-1.93_64, -1.23_27, -0.46_53] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : str ) -> Optional[int]: __snake_case : Optional[int] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : str = model.to(lowerCamelCase ) __snake_case : int = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Optional[int] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : List[str] = model(**lowerCamelCase ) __snake_case : Union[str, Any] = outputs.logits # verify the logits __snake_case : Tuple = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , lowerCamelCase ) __snake_case : Union[str, Any] = torch.tensor( [ [[6.97_13, 6.97_86, 7.24_22], [7.28_93, 7.28_25, 7.44_46], [7.65_80, 7.87_97, 7.94_20]], [[-10.68_69, -10.32_50, -10.34_71], [-10.42_28, -9.98_68, -9.71_32], [-11.04_05, -11.02_21, -10.73_18]], [[-3.30_89, -2.85_39, -2.67_40], [-3.27_06, -2.56_21, -2.51_08], [-3.25_34, -2.66_15, -2.66_51]], ] , device=lowerCamelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : Union[str, Any] ) -> Optional[int]: __snake_case : Optional[Any] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Tuple = model.to(lowerCamelCase ) __snake_case : Dict = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : List[Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Any = model(**lowerCamelCase ) __snake_case : Dict = outputs.logits.detach().cpu() __snake_case : Any = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase , target_sizes=[(50, 60)] ) __snake_case : int = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , lowerCamelCase ) __snake_case : List[str] = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase ) __snake_case : Optional[int] = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , lowerCamelCase )
81
0
'''simple docstring''' import argparse import os import sys from unittest.mock import patch import pytorch_lightning as pl import timeout_decorator import torch from distillation import SummarizationDistiller, distill_main from finetune import SummarizationModule, main from transformers import MarianMTModel from transformers.file_utils import cached_path from transformers.testing_utils import TestCasePlus, require_torch_gpu, slow from utils import load_json __a: Dict = "sshleifer/mar_enro_6_3_student" class UpperCAmelCase ( _lowerCAmelCase ): '''simple docstring''' def _lowerCAmelCase( self ) -> Any: super().setUp() lowercase__ : str = cached_path( '''https://cdn-datasets.huggingface.co/translation/wmt_en_ro-tr40k-va0.5k-te0.5k.tar.gz''' , extract_compressed_file=__lowerCAmelCase , ) lowercase__ : Optional[int] = F"""{data_cached}/wmt_en_ro-tr40k-va0.5k-te0.5k""" @slow @require_torch_gpu def _lowerCAmelCase( self ) -> Optional[Any]: MarianMTModel.from_pretrained(__lowerCAmelCase ) @slow @require_torch_gpu def _lowerCAmelCase( self ) -> List[str]: lowercase__ : List[Any] = { "$MAX_LEN": 64, "$BS": 64, "$GAS": 1, "$ENRO_DIR": self.data_dir, "facebook/mbart-large-cc25": MARIAN_MODEL, # "val_check_interval=0.25": "val_check_interval=1.0", "--learning_rate=3e-5": "--learning_rate 3e-4", "--num_train_epochs 6": "--num_train_epochs 1", } # Clean up bash script lowercase__ : Optional[Any] = (self.test_file_dir / "train_mbart_cc25_enro.sh").open().read().split('''finetune.py''' )[1].strip() lowercase__ : Any = bash_script.replace('''\\\n''' , '''''' ).strip().replace('''\"$@\"''' , '''''' ) for k, v in env_vars_to_replace.items(): lowercase__ : List[str] = bash_script.replace(__lowerCAmelCase , str(__lowerCAmelCase ) ) lowercase__ : Any = self.get_auto_remove_tmp_dir() # bash_script = bash_script.replace("--fp16 ", "") lowercase__ : int = F"""\n --output_dir {output_dir}\n --tokenizer_name Helsinki-NLP/opus-mt-en-ro\n --sortish_sampler\n --do_predict\n --gpus 1\n --freeze_encoder\n --n_train 40000\n --n_val 500\n --n_test 500\n --fp16_opt_level O1\n --num_sanity_val_steps 0\n --eval_beams 2\n """.split() # XXX: args.gpus > 1 : handle multi_gpu in the future lowercase__ : Tuple = ["finetune.py"] + bash_script.split() + args with patch.object(__lowerCAmelCase , '''argv''' , __lowerCAmelCase ): lowercase__ : Optional[Any] = argparse.ArgumentParser() lowercase__ : Tuple = pl.Trainer.add_argparse_args(__lowerCAmelCase ) lowercase__ : List[str] = SummarizationModule.add_model_specific_args(__lowerCAmelCase , os.getcwd() ) lowercase__ : List[Any] = parser.parse_args() lowercase__ : Optional[int] = main(__lowerCAmelCase ) # Check metrics lowercase__ : Union[str, Any] = load_json(model.metrics_save_path ) lowercase__ : Any = metrics["val"][0] lowercase__ : Tuple = metrics["val"][-1] self.assertEqual(len(metrics['''val'''] ) , (args.max_epochs / args.val_check_interval) ) assert isinstance(last_step_stats[F"""val_avg_{model.val_metric}"""] , __lowerCAmelCase ) self.assertGreater(last_step_stats['''val_avg_gen_time'''] , 0.0_1 ) # model hanging on generate. Maybe bad config was saved. (XXX: old comment/assert?) self.assertLessEqual(last_step_stats['''val_avg_gen_time'''] , 1.0 ) # test learning requirements: # 1. BLEU improves over the course of training by more than 2 pts self.assertGreater(last_step_stats['''val_avg_bleu'''] - first_step_stats['''val_avg_bleu'''] , 2 ) # 2. BLEU finishes above 17 self.assertGreater(last_step_stats['''val_avg_bleu'''] , 17 ) # 3. test BLEU and val BLEU within ~1.1 pt. self.assertLess(abs(metrics['''val'''][-1]['''val_avg_bleu'''] - metrics['''test'''][-1]['''test_avg_bleu'''] ) , 1.1 ) # check lightning ckpt can be loaded and has a reasonable statedict lowercase__ : Tuple = os.listdir(__lowerCAmelCase ) lowercase__ : Optional[int] = [x for x in contents if x.endswith('''.ckpt''' )][0] lowercase__ : Dict = os.path.join(args.output_dir , __lowerCAmelCase ) lowercase__ : str = torch.load(__lowerCAmelCase , map_location='''cpu''' ) lowercase__ : Union[str, Any] = "model.model.decoder.layers.0.encoder_attn_layer_norm.weight" assert expected_key in ckpt["state_dict"] assert ckpt["state_dict"]["model.model.decoder.layers.0.encoder_attn_layer_norm.weight"].dtype == torch.floataa # TODO: turn on args.do_predict when PL bug fixed. if args.do_predict: lowercase__ : str = {os.path.basename(__lowerCAmelCase ) for p in contents} assert "test_generations.txt" in contents assert "test_results.txt" in contents # assert len(metrics["val"]) == desired_n_evals assert len(metrics['''test'''] ) == 1 class UpperCAmelCase ( _lowerCAmelCase ): '''simple docstring''' @timeout_decorator.timeout(600 ) @slow @require_torch_gpu def _lowerCAmelCase( self ) -> int: lowercase__ : int = F"""{self.test_file_dir_str}/test_data/wmt_en_ro""" lowercase__ : Optional[Any] = { "--fp16_opt_level=O1": "", "$MAX_LEN": 128, "$BS": 16, "$GAS": 1, "$ENRO_DIR": data_dir, "$m": "sshleifer/student_marian_en_ro_6_1", "val_check_interval=0.25": "val_check_interval=1.0", } # Clean up bash script lowercase__ : List[Any] = ( (self.test_file_dir / "distil_marian_no_teacher.sh").open().read().split('''distillation.py''' )[1].strip() ) lowercase__ : int = bash_script.replace('''\\\n''' , '''''' ).strip().replace('''\"$@\"''' , '''''' ) lowercase__ : List[str] = bash_script.replace('''--fp16 ''' , ''' ''' ) for k, v in env_vars_to_replace.items(): lowercase__ : int = bash_script.replace(__lowerCAmelCase , str(__lowerCAmelCase ) ) lowercase__ : Dict = self.get_auto_remove_tmp_dir() lowercase__ : int = bash_script.replace('''--fp16''' , '''''' ) lowercase__ : List[str] = 6 lowercase__ : int = ( ["distillation.py"] + bash_script.split() + [ F"""--output_dir={output_dir}""", "--gpus=1", "--learning_rate=1e-3", F"""--num_train_epochs={epochs}""", "--warmup_steps=10", "--val_check_interval=1.0", "--do_predict", ] ) with patch.object(__lowerCAmelCase , '''argv''' , __lowerCAmelCase ): lowercase__ : Optional[Any] = argparse.ArgumentParser() lowercase__ : Any = pl.Trainer.add_argparse_args(__lowerCAmelCase ) lowercase__ : int = SummarizationDistiller.add_model_specific_args(__lowerCAmelCase , os.getcwd() ) lowercase__ : Union[str, Any] = parser.parse_args() # assert args.gpus == gpus THIS BREAKS for multi_gpu lowercase__ : Union[str, Any] = distill_main(__lowerCAmelCase ) # Check metrics lowercase__ : Dict = load_json(model.metrics_save_path ) lowercase__ : List[Any] = metrics["val"][0] lowercase__ : List[str] = metrics["val"][-1] assert len(metrics['''val'''] ) >= (args.max_epochs / args.val_check_interval) # +1 accounts for val_sanity_check assert last_step_stats["val_avg_gen_time"] >= 0.0_1 assert first_step_stats["val_avg_bleu"] < last_step_stats["val_avg_bleu"] # model learned nothing assert 1.0 >= last_step_stats["val_avg_gen_time"] # model hanging on generate. Maybe bad config was saved. assert isinstance(last_step_stats[F"""val_avg_{model.val_metric}"""] , __lowerCAmelCase ) # check lightning ckpt can be loaded and has a reasonable statedict lowercase__ : List[str] = os.listdir(__lowerCAmelCase ) lowercase__ : int = [x for x in contents if x.endswith('''.ckpt''' )][0] lowercase__ : List[str] = os.path.join(args.output_dir , __lowerCAmelCase ) lowercase__ : Optional[Any] = torch.load(__lowerCAmelCase , map_location='''cpu''' ) lowercase__ : int = "model.model.decoder.layers.0.encoder_attn_layer_norm.weight" assert expected_key in ckpt["state_dict"] assert ckpt["state_dict"]["model.model.decoder.layers.0.encoder_attn_layer_norm.weight"].dtype == torch.floataa # TODO: turn on args.do_predict when PL bug fixed. if args.do_predict: lowercase__ : Optional[int] = {os.path.basename(__lowerCAmelCase ) for p in contents} assert "test_generations.txt" in contents assert "test_results.txt" in contents # assert len(metrics["val"]) == desired_n_evals assert len(metrics['''test'''] ) == 1
152
import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor _snake_case : int = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[int] , *lowerCamelCase : List[Any] , **lowerCamelCase : int ) -> None: warnings.warn( "The class ImageGPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use ImageGPTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
0
def A_ ( lowercase_ , lowercase_ , lowercase_ ) -> List[Any]: return not any( neighbour == 1 and colored_vertices[i] == color for i, neighbour in enumerate(__lowerCamelCase ) ) def A_ ( lowercase_ , lowercase_ , lowercase_ , lowercase_ ) -> Tuple: # Base Case if index == len(__lowerCamelCase ): return True # Recursive Step for i in range(__lowerCamelCase ): if valid_coloring(graph[index] , __lowerCamelCase , __lowerCamelCase ): # Color current vertex _snake_case : List[str] = i # Validate coloring if util_color(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , index + 1 ): return True # Backtrack _snake_case : str = -1 return False def A_ ( lowercase_ , lowercase_ ) -> Any: _snake_case : List[str] = [-1] * len(__lowerCamelCase ) if util_color(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , 0 ): return colored_vertices return []
326
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _snake_case : Optional[Any] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[str] = ["pixel_values"] def __init__( self : List[Any] , lowerCamelCase : bool = True , lowerCamelCase : Union[int, float] = 1 / 255 , lowerCamelCase : bool = True , lowerCamelCase : int = 8 , **lowerCamelCase : Tuple , ) -> None: super().__init__(**lowerCamelCase ) __snake_case : Dict = do_rescale __snake_case : Dict = rescale_factor __snake_case : Optional[Any] = do_pad __snake_case : Tuple = pad_size def __snake_case ( self : Dict , lowerCamelCase : np.ndarray , lowerCamelCase : float , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase : Optional[int] ) -> np.ndarray: return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def __snake_case ( self : Optional[Any] , lowerCamelCase : np.ndarray , lowerCamelCase : int , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None ) -> Tuple: __snake_case , __snake_case : List[str] = get_image_size(lowerCamelCase ) __snake_case : Optional[Any] = (old_height // size + 1) * size - old_height __snake_case : List[Any] = (old_width // size + 1) * size - old_width return pad(lowerCamelCase , ((0, pad_height), (0, pad_width)) , mode="symmetric" , data_format=lowerCamelCase ) def __snake_case ( self : Tuple , lowerCamelCase : ImageInput , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[float] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[int] = None , lowerCamelCase : Optional[Union[str, TensorType]] = None , lowerCamelCase : Union[str, ChannelDimension] = ChannelDimension.FIRST , **lowerCamelCase : Union[str, Any] , ) -> List[str]: __snake_case : int = do_rescale if do_rescale is not None else self.do_rescale __snake_case : Dict = rescale_factor if rescale_factor is not None else self.rescale_factor __snake_case : str = do_pad if do_pad is not None else self.do_pad __snake_case : Any = pad_size if pad_size is not None else self.pad_size __snake_case : int = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) # All transformations expect numpy arrays. __snake_case : str = [to_numpy_array(lowerCamelCase ) for image in images] if do_rescale: __snake_case : Optional[int] = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_pad: __snake_case : Optional[Any] = [self.pad(lowerCamelCase , size=lowerCamelCase ) for image in images] __snake_case : int = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __snake_case : Union[str, Any] = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
81
0
"""simple docstring""" def _lowerCAmelCase ( UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = abs(__lowerCamelCase ) __SCREAMING_SNAKE_CASE = 0 while n > 0: res += n % 10 n //= 10 return res def _lowerCAmelCase ( UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = abs(__lowerCamelCase ) return n if n < 10 else n % 10 + sum_of_digits(n // 10 ) def _lowerCAmelCase ( UpperCamelCase_ ): return sum(int(__lowerCamelCase ) for c in str(abs(__lowerCamelCase ) ) ) def _lowerCAmelCase ( ): from collections.abc import Callable from timeit import timeit def benchmark_a_function(UpperCamelCase_ , UpperCamelCase_ ) -> None: __SCREAMING_SNAKE_CASE = f"{func.__name__}({value})" __SCREAMING_SNAKE_CASE = timeit(f"__main__.{call}" , setup="""import __main__""" ) print(f"{call:56} = {func(__lowerCamelCase )} -- {timing:.4f} seconds" ) for value in (26_2144, 1125_8999_0684_2624, 126_7650_6002_2822_9401_4967_0320_5376): for func in (sum_of_digits, sum_of_digits_recursion, sum_of_digits_compact): benchmark_a_function(__lowerCamelCase , __lowerCamelCase ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
155
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case : Any = { "configuration_informer": [ "INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "InformerConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "InformerForPrediction", "InformerModel", "InformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys _snake_case : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
0
"""simple docstring""" from . import ( albert, align, altclip, audio_spectrogram_transformer, auto, autoformer, bark, bart, barthez, bartpho, beit, bert, bert_generation, bert_japanese, bertweet, big_bird, bigbird_pegasus, biogpt, bit, blenderbot, blenderbot_small, blip, blip_a, bloom, bridgetower, byta, camembert, canine, chinese_clip, clap, clip, clipseg, codegen, conditional_detr, convbert, convnext, convnextva, cpm, cpmant, ctrl, cvt, dataavec, deberta, deberta_va, decision_transformer, deformable_detr, deit, deprecated, deta, detr, dialogpt, dinat, distilbert, dit, donut, dpr, dpt, efficientformer, efficientnet, electra, encodec, encoder_decoder, ernie, ernie_m, esm, falcon, flaubert, flava, fnet, focalnet, fsmt, funnel, git, glpn, gpta, gpt_bigcode, gpt_neo, gpt_neox, gpt_neox_japanese, gpt_swa, gptj, gptsan_japanese, graphormer, groupvit, herbert, hubert, ibert, imagegpt, informer, instructblip, jukebox, layoutlm, layoutlmva, layoutlmva, layoutxlm, led, levit, lilt, llama, longformer, longta, luke, lxmert, mam_aaa, marian, markuplm, maskaformer, maskformer, mbart, mbartaa, mega, megatron_bert, megatron_gpta, mgp_str, mluke, mobilebert, mobilenet_va, mobilenet_va, mobilevit, mobilevitva, mpnet, mra, mta, musicgen, mvp, nat, nezha, nllb, nllb_moe, nystromformer, oneformer, open_llama, openai, opt, owlvit, pegasus, pegasus_x, perceiver, phobert, pixastruct, plbart, poolformer, prophetnet, qdqbert, rag, realm, reformer, regnet, rembert, resnet, roberta, roberta_prelayernorm, roc_bert, roformer, rwkv, sam, segformer, sew, sew_d, speech_encoder_decoder, speech_to_text, speech_to_text_a, speechta, splinter, squeezebert, swiftformer, swin, swinasr, swinva, switch_transformers, ta, table_transformer, tapas, time_series_transformer, timesformer, timm_backbone, transfo_xl, trocr, tvlt, umta, unispeech, unispeech_sat, upernet, videomae, vilt, vision_encoder_decoder, vision_text_dual_encoder, visual_bert, vit, vit_hybrid, vit_mae, vit_msn, vivit, wavaveca, wavaveca_conformer, wavaveca_phoneme, wavaveca_with_lm, wavlm, whisper, x_clip, xglm, xlm, xlm_prophetnet, xlm_roberta, xlm_roberta_xl, xlnet, xmod, yolos, yoso, )
82
"""simple docstring""" import math_equivalence # From: git+https://github.com/hendrycks/math.git import datasets lowerCamelCase = """\ @article{hendrycksmath2021, title={Measuring Mathematical Problem Solving With the MATH Dataset}, author={Dan Hendrycks and Collin Burns and Saurav Kadavath and Akul Arora and Steven Basart and Eric Tang and Dawn Song and Jacob Steinhardt}, journal={arXiv preprint arXiv:2103.03874}, year={2021} } """ lowerCamelCase = """\ This metric is used to assess performance on the Mathematics Aptitude Test of Heuristics (MATH) dataset. It first canonicalizes the inputs (e.g., converting \"1/2\" to \"\\frac{1}{2}\") and then computes accuracy. """ lowerCamelCase = r""" Calculates accuracy after canonicalizing inputs. Args: predictions: list of predictions to score. Each prediction is a string that contains natural language and LaTex. references: list of reference for each prediction. Each reference is a string that contains natural language and LaTex. Returns: accuracy: accuracy after canonicalizing inputs (e.g., converting \"1/2\" to \"\\frac{1}{2}\") Examples: >>> metric = datasets.load_metric(\"competition_math\") >>> results = metric.compute(references=[\"\\frac{1}{2}\"], predictions=[\"1/2\"]) >>> print(results) {'accuracy': 1.0} """ @datasets.utils.file_utils.add_end_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase__ ( datasets.Metric ): '''simple docstring''' def lowercase__ ( self : int ) -> Tuple: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" ), "references": datasets.Value("string" ), } ) , homepage="https://github.com/hendrycks/math" , codebase_urls=["https://github.com/hendrycks/math"] , ) def lowercase__ ( self : Optional[int] , _UpperCAmelCase : Dict , _UpperCAmelCase : Any ) -> Any: '''simple docstring''' UpperCAmelCase_ = 0.0 for i, j in zip(_UpperCAmelCase , _UpperCAmelCase ): n_correct += 1.0 if math_equivalence.is_equiv(_UpperCAmelCase , _UpperCAmelCase ) else 0.0 UpperCAmelCase_ = n_correct / len(_UpperCAmelCase ) return { "accuracy": accuracy, }
82
1
"""simple docstring""" from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": lowerCamelCase = input("""Enter image url: """).strip() print(F"Downloading image from {url} ...") lowerCamelCase = BeautifulSoup(requests.get(url).content, """html.parser""") # The image URL is in the content field of the first meta tag with property og:image lowerCamelCase = soup.find("""meta""", {"""property""": """og:image"""})["""content"""] lowerCamelCase = requests.get(image_url).content lowerCamelCase = F"{datetime.now():%Y-%m-%d_%H:%M:%S}.jpg" with open(file_name, """wb""") as fp: fp.write(image_data) print(F"Done. Image saved to disk as {file_name}.")
82
"""simple docstring""" lowerCamelCase = """Alexander Joslin""" import operator as op from .stack import Stack def a__ ( lowerCAmelCase__ ): UpperCAmelCase_ = {"*": op.mul, "/": op.truediv, "+": op.add, "-": op.sub} UpperCAmelCase_ = Stack() UpperCAmelCase_ = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(lowerCAmelCase__ ) ) elif i in operators: # RULE 2 operator_stack.push(lowerCAmelCase__ ) elif i == ")": # RULE 4 UpperCAmelCase_ = operator_stack.peek() operator_stack.pop() UpperCAmelCase_ = operand_stack.peek() operand_stack.pop() UpperCAmelCase_ = operand_stack.peek() operand_stack.pop() UpperCAmelCase_ = operators[opr](lowerCAmelCase__ , lowerCAmelCase__ ) operand_stack.push(lowerCAmelCase__ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": lowerCamelCase = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F"{equation} = {dijkstras_two_stack_algorithm(equation)}")
82
1
"""simple docstring""" import argparse import json import math import os import time import traceback import zipfile from collections import Counter import requests def a__ ( lowerCAmelCase__ , lowerCAmelCase__=None ): UpperCAmelCase_ = None if token is not None: UpperCAmelCase_ = {"Accept": "application/vnd.github+json", "Authorization": f"""Bearer {token}"""} UpperCAmelCase_ = f"""https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100""" UpperCAmelCase_ = requests.get(lowerCAmelCase__ , headers=lowerCAmelCase__ ).json() UpperCAmelCase_ = {} try: job_links.update({job["name"]: job["html_url"] for job in result["jobs"]} ) UpperCAmelCase_ = math.ceil((result["total_count"] - 100) / 100 ) for i in range(lowerCAmelCase__ ): UpperCAmelCase_ = requests.get(url + f"""&page={i + 2}""" , headers=lowerCAmelCase__ ).json() job_links.update({job["name"]: job["html_url"] for job in result["jobs"]} ) return job_links except Exception: print(f"""Unknown error, could not fetch links:\n{traceback.format_exc()}""" ) return {} def a__ ( lowerCAmelCase__ , lowerCAmelCase__=None ): UpperCAmelCase_ = None if token is not None: UpperCAmelCase_ = {"Accept": "application/vnd.github+json", "Authorization": f"""Bearer {token}"""} UpperCAmelCase_ = f"""https://api.github.com/repos/huggingface/transformers/actions/runs/{worflow_run_id}/artifacts?per_page=100""" UpperCAmelCase_ = requests.get(lowerCAmelCase__ , headers=lowerCAmelCase__ ).json() UpperCAmelCase_ = {} try: artifacts.update({artifact["name"]: artifact["archive_download_url"] for artifact in result["artifacts"]} ) UpperCAmelCase_ = math.ceil((result["total_count"] - 100) / 100 ) for i in range(lowerCAmelCase__ ): UpperCAmelCase_ = requests.get(url + f"""&page={i + 2}""" , headers=lowerCAmelCase__ ).json() artifacts.update({artifact["name"]: artifact["archive_download_url"] for artifact in result["artifacts"]} ) return artifacts except Exception: print(f"""Unknown error, could not fetch links:\n{traceback.format_exc()}""" ) return {} def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = None if token is not None: UpperCAmelCase_ = {"Accept": "application/vnd.github+json", "Authorization": f"""Bearer {token}"""} UpperCAmelCase_ = requests.get(lowerCAmelCase__ , headers=lowerCAmelCase__ , allow_redirects=lowerCAmelCase__ ) UpperCAmelCase_ = result.headers["Location"] UpperCAmelCase_ = requests.get(lowerCAmelCase__ , allow_redirects=lowerCAmelCase__ ) UpperCAmelCase_ = os.path.join(lowerCAmelCase__ , f"""{artifact_name}.zip""" ) with open(lowerCAmelCase__ , "wb" ) as fp: fp.write(response.content ) def a__ ( lowerCAmelCase__ , lowerCAmelCase__=None ): UpperCAmelCase_ = [] UpperCAmelCase_ = [] UpperCAmelCase_ = None with zipfile.ZipFile(lowerCAmelCase__ ) as z: for filename in z.namelist(): if not os.path.isdir(lowerCAmelCase__ ): # read the file if filename in ["failures_line.txt", "summary_short.txt", "job_name.txt"]: with z.open(lowerCAmelCase__ ) as f: for line in f: UpperCAmelCase_ = line.decode("UTF-8" ).strip() if filename == "failures_line.txt": try: # `error_line` is the place where `error` occurs UpperCAmelCase_ = line[: line.index(": " )] UpperCAmelCase_ = line[line.index(": " ) + len(": " ) :] errors.append([error_line, error] ) except Exception: # skip un-related lines pass elif filename == "summary_short.txt" and line.startswith("FAILED " ): # `test` is the test method that failed UpperCAmelCase_ = line[len("FAILED " ) :] failed_tests.append(lowerCAmelCase__ ) elif filename == "job_name.txt": UpperCAmelCase_ = line if len(lowerCAmelCase__ ) != len(lowerCAmelCase__ ): raise ValueError( f"""`errors` and `failed_tests` should have the same number of elements. Got {len(lowerCAmelCase__ )} for `errors` """ f"""and {len(lowerCAmelCase__ )} for `failed_tests` instead. The test reports in {artifact_zip_path} have some""" " problem." ) UpperCAmelCase_ = None if job_name and job_links: UpperCAmelCase_ = job_links.get(lowerCAmelCase__ , lowerCAmelCase__ ) # A list with elements of the form (line of error, error, failed test) UpperCAmelCase_ = [x + [y] + [job_link] for x, y in zip(lowerCAmelCase__ , lowerCAmelCase__ )] return result def a__ ( lowerCAmelCase__ , lowerCAmelCase__=None ): UpperCAmelCase_ = [] UpperCAmelCase_ = [os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) for p in os.listdir(lowerCAmelCase__ ) if p.endswith(".zip" )] for p in paths: errors.extend(get_errors_from_single_artifact(lowerCAmelCase__ , job_links=lowerCAmelCase__ ) ) return errors def a__ ( lowerCAmelCase__ , lowerCAmelCase__=None ): UpperCAmelCase_ = Counter() counter.update([x[1] for x in logs] ) UpperCAmelCase_ = counter.most_common() UpperCAmelCase_ = {} for error, count in counts: if error_filter is None or error not in error_filter: UpperCAmelCase_ = {"count": count, "failed_tests": [(x[2], x[0]) for x in logs if x[1] == error]} UpperCAmelCase_ = dict(sorted(r.items() , key=lambda lowerCAmelCase__ : item[1]["count"] , reverse=lowerCAmelCase__ ) ) return r def a__ ( lowerCAmelCase__ ): UpperCAmelCase_ = test.split("::" )[0] if test.startswith("tests/models/" ): UpperCAmelCase_ = test.split("/" )[2] else: UpperCAmelCase_ = None return test def a__ ( lowerCAmelCase__ , lowerCAmelCase__=None ): UpperCAmelCase_ = [(x[0], x[1], get_model(x[2] )) for x in logs] UpperCAmelCase_ = [x for x in logs if x[2] is not None] UpperCAmelCase_ = {x[2] for x in logs} UpperCAmelCase_ = {} for test in tests: UpperCAmelCase_ = Counter() # count by errors in `test` counter.update([x[1] for x in logs if x[2] == test] ) UpperCAmelCase_ = counter.most_common() UpperCAmelCase_ = {error: count for error, count in counts if (error_filter is None or error not in error_filter)} UpperCAmelCase_ = sum(error_counts.values() ) if n_errors > 0: UpperCAmelCase_ = {"count": n_errors, "errors": error_counts} UpperCAmelCase_ = dict(sorted(r.items() , key=lambda lowerCAmelCase__ : item[1]["count"] , reverse=lowerCAmelCase__ ) ) return r def a__ ( lowerCAmelCase__ ): UpperCAmelCase_ = "| no. | error | status |" UpperCAmelCase_ = "|-:|:-|:-|" UpperCAmelCase_ = [header, sep] for error in reduced_by_error: UpperCAmelCase_ = reduced_by_error[error]["count"] UpperCAmelCase_ = f"""| {count} | {error[:100]} | |""" lines.append(lowerCAmelCase__ ) return "\n".join(lowerCAmelCase__ ) def a__ ( lowerCAmelCase__ ): UpperCAmelCase_ = "| model | no. of errors | major error | count |" UpperCAmelCase_ = "|-:|-:|-:|-:|" UpperCAmelCase_ = [header, sep] for model in reduced_by_model: UpperCAmelCase_ = reduced_by_model[model]["count"] UpperCAmelCase_ , UpperCAmelCase_ = list(reduced_by_model[model]["errors"].items() )[0] UpperCAmelCase_ = f"""| {model} | {count} | {error[:60]} | {_count} |""" lines.append(lowerCAmelCase__ ) return "\n".join(lowerCAmelCase__ ) if __name__ == "__main__": lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument("""--workflow_run_id""", type=str, required=True, help="""A GitHub Actions workflow run id.""") parser.add_argument( """--output_dir""", type=str, required=True, help="""Where to store the downloaded artifacts and other result files.""", ) parser.add_argument("""--token""", default=None, type=str, help="""A token that has actions:read permission.""") lowerCamelCase = parser.parse_args() os.makedirs(args.output_dir, exist_ok=True) lowerCamelCase = get_job_links(args.workflow_run_id, token=args.token) lowerCamelCase = {} # To deal with `workflow_call` event, where a job name is the combination of the job names in the caller and callee. # For example, `PyTorch 1.11 / Model tests (models/albert, single-gpu)`. if _job_links: for k, v in _job_links.items(): # This is how GitHub actions combine job names. if " / " in k: lowerCamelCase = k.find(""" / """) lowerCamelCase = k[index + len(""" / """) :] lowerCamelCase = v with open(os.path.join(args.output_dir, """job_links.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(job_links, fp, ensure_ascii=False, indent=4) lowerCamelCase = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, """artifacts.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) for idx, (name, url) in enumerate(artifacts.items()): download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) lowerCamelCase = get_all_errors(args.output_dir, job_links=job_links) # `e[1]` is the error lowerCamelCase = Counter() counter.update([e[1] for e in errors]) # print the top 30 most common test errors lowerCamelCase = counter.most_common(30) for item in most_common: print(item) with open(os.path.join(args.output_dir, """errors.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(errors, fp, ensure_ascii=False, indent=4) lowerCamelCase = reduce_by_error(errors) lowerCamelCase = reduce_by_model(errors) lowerCamelCase = make_github_table(reduced_by_error) lowerCamelCase = make_github_table_per_model(reduced_by_model) with open(os.path.join(args.output_dir, """reduced_by_error.txt"""), """w""", encoding="""UTF-8""") as fp: fp.write(sa) with open(os.path.join(args.output_dir, """reduced_by_model.txt"""), """w""", encoding="""UTF-8""") as fp: fp.write(sa)
82
"""simple docstring""" from __future__ import annotations from fractions import Fraction from math import gcd, sqrt def a__ ( lowerCAmelCase__ ): UpperCAmelCase_ = int(number**0.5 ) return number == sq * sq def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = x_num * y_den * z_den + y_num * x_den * z_den + z_num * x_den * y_den UpperCAmelCase_ = x_den * y_den * z_den UpperCAmelCase_ = gcd(lowerCAmelCase__ , lowerCAmelCase__ ) top //= hcf bottom //= hcf return top, bottom def a__ ( lowerCAmelCase__ = 35 ): UpperCAmelCase_ = set() UpperCAmelCase_ = 42 UpperCAmelCase_ = Fraction(0 ) UpperCAmelCase_ = 42 for x_num in range(1 , order + 1 ): for x_den in range(x_num + 1 , order + 1 ): for y_num in range(1 , order + 1 ): for y_den in range(y_num + 1 , order + 1 ): # n=1 UpperCAmelCase_ = x_num * y_den + x_den * y_num UpperCAmelCase_ = x_den * y_den UpperCAmelCase_ = gcd(lowerCAmelCase__ , lowerCAmelCase__ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: UpperCAmelCase_ = add_three( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) unique_s.add(lowerCAmelCase__ ) # n=2 UpperCAmelCase_ = ( x_num * x_num * y_den * y_den + x_den * x_den * y_num * y_num ) UpperCAmelCase_ = x_den * x_den * y_den * y_den if is_sq(lowerCAmelCase__ ) and is_sq(lowerCAmelCase__ ): UpperCAmelCase_ = int(sqrt(lowerCAmelCase__ ) ) UpperCAmelCase_ = int(sqrt(lowerCAmelCase__ ) ) UpperCAmelCase_ = gcd(lowerCAmelCase__ , lowerCAmelCase__ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: UpperCAmelCase_ = add_three( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) unique_s.add(lowerCAmelCase__ ) # n=-1 UpperCAmelCase_ = x_num * y_num UpperCAmelCase_ = x_den * y_num + x_num * y_den UpperCAmelCase_ = gcd(lowerCAmelCase__ , lowerCAmelCase__ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: UpperCAmelCase_ = add_three( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) unique_s.add(lowerCAmelCase__ ) # n=2 UpperCAmelCase_ = x_num * x_num * y_num * y_num UpperCAmelCase_ = ( x_den * x_den * y_num * y_num + x_num * x_num * y_den * y_den ) if is_sq(lowerCAmelCase__ ) and is_sq(lowerCAmelCase__ ): UpperCAmelCase_ = int(sqrt(lowerCAmelCase__ ) ) UpperCAmelCase_ = int(sqrt(lowerCAmelCase__ ) ) UpperCAmelCase_ = gcd(lowerCAmelCase__ , lowerCAmelCase__ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: UpperCAmelCase_ = add_three( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) unique_s.add(lowerCAmelCase__ ) for num, den in unique_s: total += Fraction(lowerCAmelCase__ , lowerCAmelCase__ ) return total.denominator + total.numerator if __name__ == "__main__": print(F"{solution() = }")
82
1
"""simple docstring""" # Lint as: python3 # pylint: enable=line-too-long # pylint: disable=g-import-not-at-top,g-bad-import-order,wrong-import-position lowerCamelCase = """2.13.1""" import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse("""3.7"""): raise ImportWarning( """To use `datasets`, Python>=3.7 is required, and the current version of Python doesn't match this condition.""" ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( """To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn't match this condition.\n""" """If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.""" ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip lowerCamelCase = concatenate_datasets lowerCamelCase = DownloadConfig lowerCamelCase = DownloadManager lowerCamelCase = DownloadMode lowerCamelCase = DownloadConfig lowerCamelCase = DownloadMode lowerCamelCase = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
82
"""simple docstring""" from __future__ import annotations from math import pi, sqrt def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ): if inductance <= 0: raise ValueError("Inductance cannot be 0 or negative" ) elif capacitance <= 0: raise ValueError("Capacitance cannot be 0 or negative" ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
82
1
"""simple docstring""" import unittest import numpy as np import torch from diffusers import KarrasVePipeline, KarrasVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class lowercase__ ( unittest.TestCase ): '''simple docstring''' @property def lowercase__ ( self : Any ) -> Optional[int]: '''simple docstring''' torch.manual_seed(0 ) UpperCAmelCase_ = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=("DownBlock2D", "AttnDownBlock2D") , up_block_types=("AttnUpBlock2D", "UpBlock2D") , ) return model def lowercase__ ( self : Optional[Any] ) -> Optional[int]: '''simple docstring''' UpperCAmelCase_ = self.dummy_uncond_unet UpperCAmelCase_ = KarrasVeScheduler() UpperCAmelCase_ = KarrasVePipeline(unet=_UpperCAmelCase , scheduler=_UpperCAmelCase ) pipe.to(_UpperCAmelCase ) pipe.set_progress_bar_config(disable=_UpperCAmelCase ) UpperCAmelCase_ = torch.manual_seed(0 ) UpperCAmelCase_ = pipe(num_inference_steps=2 , generator=_UpperCAmelCase , output_type="numpy" ).images UpperCAmelCase_ = torch.manual_seed(0 ) UpperCAmelCase_ = pipe(num_inference_steps=2 , generator=_UpperCAmelCase , output_type="numpy" , return_dict=_UpperCAmelCase )[0] UpperCAmelCase_ = image[0, -3:, -3:, -1] UpperCAmelCase_ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) UpperCAmelCase_ = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class lowercase__ ( unittest.TestCase ): '''simple docstring''' def lowercase__ ( self : List[Any] ) -> int: '''simple docstring''' UpperCAmelCase_ = "google/ncsnpp-celebahq-256" UpperCAmelCase_ = UNetaDModel.from_pretrained(_UpperCAmelCase ) UpperCAmelCase_ = KarrasVeScheduler() UpperCAmelCase_ = KarrasVePipeline(unet=_UpperCAmelCase , scheduler=_UpperCAmelCase ) pipe.to(_UpperCAmelCase ) pipe.set_progress_bar_config(disable=_UpperCAmelCase ) UpperCAmelCase_ = torch.manual_seed(0 ) UpperCAmelCase_ = pipe(num_inference_steps=20 , generator=_UpperCAmelCase , output_type="numpy" ).images UpperCAmelCase_ = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) UpperCAmelCase_ = np.array([0.578, 0.5811, 0.5924, 0.5809, 0.587, 0.5886, 0.5861, 0.5802, 0.586] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
82
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase = logging.get_logger(__name__) lowerCamelCase = { """google/vit-base-patch16-224""": """https://huggingface.co/vit-base-patch16-224/resolve/main/config.json""", # See all ViT models at https://huggingface.co/models?filter=vit } class lowercase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' UpperCamelCase = '''vit''' def __init__( self : List[str] , _UpperCAmelCase : Optional[int]=768 , _UpperCAmelCase : Optional[Any]=12 , _UpperCAmelCase : Dict=12 , _UpperCAmelCase : int=3072 , _UpperCAmelCase : Optional[Any]="gelu" , _UpperCAmelCase : Dict=0.0 , _UpperCAmelCase : Dict=0.0 , _UpperCAmelCase : Union[str, Any]=0.02 , _UpperCAmelCase : int=1e-12 , _UpperCAmelCase : List[str]=224 , _UpperCAmelCase : Tuple=16 , _UpperCAmelCase : Optional[Any]=3 , _UpperCAmelCase : Union[str, Any]=True , _UpperCAmelCase : Optional[int]=16 , **_UpperCAmelCase : List[str] , ) -> List[str]: '''simple docstring''' super().__init__(**_UpperCAmelCase ) UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_act UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = initializer_range UpperCAmelCase_ = layer_norm_eps UpperCAmelCase_ = image_size UpperCAmelCase_ = patch_size UpperCAmelCase_ = num_channels UpperCAmelCase_ = qkv_bias UpperCAmelCase_ = encoder_stride class lowercase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' UpperCamelCase = version.parse('''1.11''' ) @property def lowercase__ ( self : Dict ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ] ) @property def lowercase__ ( self : Union[str, Any] ) -> float: '''simple docstring''' return 1e-4
82
1
"""simple docstring""" class lowercase__ : '''simple docstring''' def __init__( self : str ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase_ = "" UpperCAmelCase_ = "" UpperCAmelCase_ = [] def lowercase__ ( self : List[Any] , _UpperCAmelCase : int , _UpperCAmelCase : int ) -> int: '''simple docstring''' if m == -1: return n + 1 elif n == -1: return m + 1 elif self.dp[m][n] > -1: return self.dp[m][n] else: if self.worda[m] == self.worda[n]: UpperCAmelCase_ = self.__min_dist_top_down_dp(m - 1 , n - 1 ) else: UpperCAmelCase_ = self.__min_dist_top_down_dp(_UpperCAmelCase , n - 1 ) UpperCAmelCase_ = self.__min_dist_top_down_dp(m - 1 , _UpperCAmelCase ) UpperCAmelCase_ = self.__min_dist_top_down_dp(m - 1 , n - 1 ) UpperCAmelCase_ = 1 + min(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) return self.dp[m][n] def lowercase__ ( self : str , _UpperCAmelCase : str , _UpperCAmelCase : str ) -> int: '''simple docstring''' UpperCAmelCase_ = worda UpperCAmelCase_ = worda UpperCAmelCase_ = [[-1 for _ in range(len(_UpperCAmelCase ) )] for _ in range(len(_UpperCAmelCase ) )] return self.__min_dist_top_down_dp(len(_UpperCAmelCase ) - 1 , len(_UpperCAmelCase ) - 1 ) def lowercase__ ( self : int , _UpperCAmelCase : str , _UpperCAmelCase : str ) -> int: '''simple docstring''' UpperCAmelCase_ = worda UpperCAmelCase_ = worda UpperCAmelCase_ = len(_UpperCAmelCase ) UpperCAmelCase_ = len(_UpperCAmelCase ) UpperCAmelCase_ = [[0 for _ in range(n + 1 )] for _ in range(m + 1 )] for i in range(m + 1 ): for j in range(n + 1 ): if i == 0: # first string is empty UpperCAmelCase_ = j elif j == 0: # second string is empty UpperCAmelCase_ = i elif worda[i - 1] == worda[j - 1]: # last characters are equal UpperCAmelCase_ = self.dp[i - 1][j - 1] else: UpperCAmelCase_ = self.dp[i][j - 1] UpperCAmelCase_ = self.dp[i - 1][j] UpperCAmelCase_ = self.dp[i - 1][j - 1] UpperCAmelCase_ = 1 + min(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) return self.dp[m][n] if __name__ == "__main__": lowerCamelCase = EditDistance() print("""****************** Testing Edit Distance DP Algorithm ******************""") print() lowerCamelCase = input("""Enter the first string: """).strip() lowerCamelCase = input("""Enter the second string: """).strip() print() print(F"The minimum edit distance is: {solver.min_dist_top_down(Sa, Sa)}") print(F"The minimum edit distance is: {solver.min_dist_bottom_up(Sa, Sa)}") print() print("""*************** End of Testing Edit Distance DP Algorithm ***************""")
82
"""simple docstring""" import unittest import numpy as np import requests from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch from transformers.pytorch_utils import is_torch_greater_or_equal_than_1_11 else: lowerCamelCase = False if is_vision_available(): from PIL import Image from transformers import PixaStructImageProcessor class lowercase__ ( unittest.TestCase ): '''simple docstring''' def __init__( self : Tuple , _UpperCAmelCase : Dict , _UpperCAmelCase : Optional[Any]=7 , _UpperCAmelCase : str=3 , _UpperCAmelCase : Any=18 , _UpperCAmelCase : int=30 , _UpperCAmelCase : Tuple=400 , _UpperCAmelCase : List[Any]=None , _UpperCAmelCase : str=True , _UpperCAmelCase : List[Any]=True , _UpperCAmelCase : int=None , ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase_ = size if size is not None else {"height": 20, "width": 20} UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = num_channels UpperCAmelCase_ = image_size UpperCAmelCase_ = min_resolution UpperCAmelCase_ = max_resolution UpperCAmelCase_ = size UpperCAmelCase_ = do_normalize UpperCAmelCase_ = do_convert_rgb UpperCAmelCase_ = [512, 1024, 2048, 4096] UpperCAmelCase_ = patch_size if patch_size is not None else {"height": 16, "width": 16} def lowercase__ ( self : List[Any] ) -> List[Any]: '''simple docstring''' return {"do_normalize": self.do_normalize, "do_convert_rgb": self.do_convert_rgb} def lowercase__ ( self : List[Any] ) -> Dict: '''simple docstring''' UpperCAmelCase_ = "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/transformers/tasks/australia.jpg" UpperCAmelCase_ = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ).convert("RGB" ) return raw_image @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason='''`Pix2StructImageProcessor` requires `torch>=1.11.0`.''' , ) @require_torch @require_vision class lowercase__ ( SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' UpperCamelCase = PixaStructImageProcessor if is_vision_available() else None def lowercase__ ( self : Optional[int] ) -> int: '''simple docstring''' UpperCAmelCase_ = PixaStructImageProcessingTester(self ) @property def lowercase__ ( self : List[Any] ) -> Any: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def lowercase__ ( self : Dict ) -> Tuple: '''simple docstring''' UpperCAmelCase_ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_UpperCAmelCase , "do_normalize" ) ) self.assertTrue(hasattr(_UpperCAmelCase , "do_convert_rgb" ) ) def lowercase__ ( self : str ) -> List[str]: '''simple docstring''' UpperCAmelCase_ = self.image_processor_tester.prepare_dummy_image() UpperCAmelCase_ = self.image_processing_class(**self.image_processor_dict ) UpperCAmelCase_ = 2048 UpperCAmelCase_ = image_processor(_UpperCAmelCase , return_tensors="pt" , max_patches=_UpperCAmelCase ) self.assertTrue(torch.allclose(inputs.flattened_patches.mean() , torch.tensor(0.0606 ) , atol=1e-3 , rtol=1e-3 ) ) def lowercase__ ( self : List[Any] ) -> Optional[int]: '''simple docstring''' UpperCAmelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(_UpperCAmelCase , Image.Image ) # Test not batched input UpperCAmelCase_ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input UpperCAmelCase_ = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=_UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched UpperCAmelCase_ = image_processor( _UpperCAmelCase , return_tensors="pt" , max_patches=_UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def lowercase__ ( self : str ) -> int: '''simple docstring''' UpperCAmelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(_UpperCAmelCase , Image.Image ) # Test not batched input UpperCAmelCase_ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 UpperCAmelCase_ = True for max_patch in self.image_processor_tester.max_patches: # Test not batched input with self.assertRaises(_UpperCAmelCase ): UpperCAmelCase_ = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=_UpperCAmelCase ).flattened_patches UpperCAmelCase_ = "Hello" UpperCAmelCase_ = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=_UpperCAmelCase , header_text=_UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched UpperCAmelCase_ = image_processor( _UpperCAmelCase , return_tensors="pt" , max_patches=_UpperCAmelCase , header_text=_UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def lowercase__ ( self : str ) -> Tuple: '''simple docstring''' UpperCAmelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_UpperCAmelCase , numpify=_UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(_UpperCAmelCase , np.ndarray ) UpperCAmelCase_ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input UpperCAmelCase_ = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=_UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched UpperCAmelCase_ = image_processor( _UpperCAmelCase , return_tensors="pt" , max_patches=_UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def lowercase__ ( self : Optional[int] ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_UpperCAmelCase , torchify=_UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(_UpperCAmelCase , torch.Tensor ) # Test not batched input UpperCAmelCase_ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input UpperCAmelCase_ = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=_UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched UpperCAmelCase_ = image_processor( _UpperCAmelCase , return_tensors="pt" , max_patches=_UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason='''`Pix2StructImageProcessor` requires `torch>=1.11.0`.''' , ) @require_torch @require_vision class lowercase__ ( SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' UpperCamelCase = PixaStructImageProcessor if is_vision_available() else None def lowercase__ ( self : List[str] ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase_ = PixaStructImageProcessingTester(self , num_channels=4 ) UpperCAmelCase_ = 3 @property def lowercase__ ( self : str ) -> Optional[int]: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def lowercase__ ( self : str ) -> Optional[int]: '''simple docstring''' UpperCAmelCase_ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_UpperCAmelCase , "do_normalize" ) ) self.assertTrue(hasattr(_UpperCAmelCase , "do_convert_rgb" ) ) def lowercase__ ( self : List[str] ) -> Tuple: '''simple docstring''' UpperCAmelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(_UpperCAmelCase , Image.Image ) # Test not batched input UpperCAmelCase_ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * (self.image_processor_tester.num_channels - 1) ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input UpperCAmelCase_ = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=_UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched UpperCAmelCase_ = image_processor( _UpperCAmelCase , return_tensors="pt" , max_patches=_UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , )
82
1
"""simple docstring""" from typing import Dict from transformers import EvalPrediction, HfArgumentParser, TrainingArguments, is_torch_available from transformers.testing_utils import ( TestCasePlus, execute_subprocess_async, get_torch_dist_unique_port, require_torch_multi_gpu, require_torch_neuroncore, ) from transformers.training_args import ParallelMode from transformers.utils import logging lowerCamelCase = logging.get_logger(__name__) if is_torch_available(): import torch from torch import nn from torch.utils.data import Dataset from transformers import Trainer class lowercase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self : Dict , _UpperCAmelCase : int = 101 ) -> List[Any]: '''simple docstring''' UpperCAmelCase_ = length def __len__( self : int ) -> Union[str, Any]: '''simple docstring''' return self.length def __getitem__( self : int , _UpperCAmelCase : Optional[int] ) -> int: '''simple docstring''' return i class lowercase__ : '''simple docstring''' def __call__( self : int , _UpperCAmelCase : str ) -> Optional[int]: '''simple docstring''' return {"input_ids": torch.tensor(_UpperCAmelCase ), "labels": torch.tensor(_UpperCAmelCase )} class lowercase__ ( nn.Module ): '''simple docstring''' def __init__( self : Optional[int] ) -> int: '''simple docstring''' super().__init__() # Add some (unused) params otherwise DDP will complain. UpperCAmelCase_ = nn.Linear(120 , 80 ) def lowercase__ ( self : Dict , _UpperCAmelCase : List[Any] , _UpperCAmelCase : Optional[Any]=None ) -> Any: '''simple docstring''' if labels is not None: return torch.tensor(0.0 , device=input_ids.device ), input_ids else: return input_ids class lowercase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' @require_torch_neuroncore def lowercase__ ( self : Dict ) -> List[str]: '''simple docstring''' UpperCAmelCase_ = F"""--nproc_per_node=2 --master_port={get_torch_dist_unique_port()} {self.test_file_dir}/test_trainer_distributed.py """.split() UpperCAmelCase_ = self.get_auto_remove_tmp_dir() UpperCAmelCase_ = F"""--output_dir {output_dir}""".split() UpperCAmelCase_ = ["torchrun"] + distributed_args + args execute_subprocess_async(_UpperCAmelCase , env=self.get_env() ) # successful return here == success - any errors would have caused an error in the sub-call class lowercase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' @require_torch_multi_gpu def lowercase__ ( self : Optional[int] ) -> int: '''simple docstring''' UpperCAmelCase_ = F"""--nproc_per_node={torch.cuda.device_count()} --master_port={get_torch_dist_unique_port()} {self.test_file_dir}/test_trainer_distributed.py """.split() UpperCAmelCase_ = self.get_auto_remove_tmp_dir() UpperCAmelCase_ = F"""--output_dir {output_dir}""".split() UpperCAmelCase_ = ["torchrun"] + distributed_args + args execute_subprocess_async(_UpperCAmelCase , env=self.get_env() ) # successful return here == success - any errors would have caused an error in the sub-call if __name__ == "__main__": # The script below is meant to be run under torch.distributed, on a machine with multiple GPUs: # # PYTHONPATH="src" python -m torch.distributed.run --nproc_per_node 2 --output_dir output_dir ./tests/test_trainer_distributed.py lowerCamelCase = HfArgumentParser((TrainingArguments,)) lowerCamelCase = parser.parse_args_into_dataclasses()[0] logger.warning( F"Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}, " F"distributed training: {training_args.parallel_mode != ParallelMode.NOT_DISTRIBUTED}" ) # Essentially, what we want to verify in the distributed case is that we get all samples back, # in the right order. (this is crucial for prediction for instance) for dataset_length in [101, 40, 7]: lowerCamelCase = DummyDataset(dataset_length) def a__ ( lowerCAmelCase__ ): UpperCAmelCase_ = list(range(len(lowerCAmelCase__ ) ) ) UpperCAmelCase_ = p.predictions.tolist() == sequential and p.label_ids.tolist() == sequential if not success and training_args.local_rank == 0: logger.warning( "Predictions and/or labels do not match expected results:\n - predictions: " f"""{p.predictions.tolist()}\n - labels: {p.label_ids.tolist()}\n - expected: {sequential}""" ) return {"success": success} lowerCamelCase = Trainer( model=DummyModel(), args=training_args, data_collator=DummyDataCollator(), eval_dataset=dataset, compute_metrics=compute_metrics, ) lowerCamelCase = trainer.evaluate() logger.info(metrics) if metrics["eval_success"] is not True: logger.error(metrics) exit(1) lowerCamelCase = trainer.predict(dataset) logger.info(p.metrics) if p.metrics["test_success"] is not True: logger.error(p.metrics) exit(1) lowerCamelCase = 2 lowerCamelCase = trainer.evaluate() logger.info(metrics) if metrics["eval_success"] is not True: logger.error(metrics) exit(1) lowerCamelCase = trainer.predict(dataset) logger.info(p.metrics) if p.metrics["test_success"] is not True: logger.error(p.metrics) exit(1) lowerCamelCase = None
82
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from timm import create_model from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import BitConfig, BitForImageClassification, BitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase = logging.get_logger(__name__) def a__ ( lowerCAmelCase__ ): UpperCAmelCase_ = "huggingface/label-files" UpperCAmelCase_ = "imagenet-1k-id2label.json" UpperCAmelCase_ = json.load(open(hf_hub_download(lowerCAmelCase__ , lowerCAmelCase__ , repo_type="dataset" ) , "r" ) ) UpperCAmelCase_ = {int(lowerCAmelCase__ ): v for k, v in idalabel.items()} UpperCAmelCase_ = {v: k for k, v in idalabel.items()} UpperCAmelCase_ = "std_conv" if "bit" in model_name else False # note that when using BiT as backbone for ViT-hybrid checkpoints, # one needs to additionally set config.layer_type = "bottleneck", config.stem_type = "same", # config.conv_layer = "std_conv_same" UpperCAmelCase_ = BitConfig( conv_layer=lowerCAmelCase__ , num_labels=1000 , idalabel=lowerCAmelCase__ , labelaid=lowerCAmelCase__ , ) return config def a__ ( lowerCAmelCase__ ): if "stem.conv" in name: UpperCAmelCase_ = name.replace("stem.conv" , "bit.embedder.convolution" ) if "blocks" in name: UpperCAmelCase_ = name.replace("blocks" , "layers" ) if "head.fc" in name: UpperCAmelCase_ = name.replace("head.fc" , "classifier.1" ) if name.startswith("norm" ): UpperCAmelCase_ = "bit." + name if "bit" not in name and "classifier" not in name: UpperCAmelCase_ = "bit.encoder." + name return name def a__ ( ): UpperCAmelCase_ = "http://images.cocodataset.org/val2017/000000039769.jpg" UpperCAmelCase_ = Image.open(requests.get(lowerCAmelCase__ , stream=lowerCAmelCase__ ).raw ) return im @torch.no_grad() def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=False ): UpperCAmelCase_ = get_config(lowerCAmelCase__ ) # load original model from timm UpperCAmelCase_ = create_model(lowerCAmelCase__ , pretrained=lowerCAmelCase__ ) timm_model.eval() # load state_dict of original model UpperCAmelCase_ = timm_model.state_dict() for key in state_dict.copy().keys(): UpperCAmelCase_ = state_dict.pop(lowerCAmelCase__ ) UpperCAmelCase_ = val.squeeze() if "head" in key else val # load HuggingFace model UpperCAmelCase_ = BitForImageClassification(lowerCAmelCase__ ) model.eval() model.load_state_dict(lowerCAmelCase__ ) # create image processor UpperCAmelCase_ = create_transform(**resolve_data_config({} , model=lowerCAmelCase__ ) ) UpperCAmelCase_ = transform.transforms UpperCAmelCase_ = { "bilinear": PILImageResampling.BILINEAR, "bicubic": PILImageResampling.BICUBIC, "nearest": PILImageResampling.NEAREST, } UpperCAmelCase_ = BitImageProcessor( do_resize=lowerCAmelCase__ , size={"shortest_edge": timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=lowerCAmelCase__ , crop_size={"height": timm_transforms[1].size[0], "width": timm_transforms[1].size[1]} , do_normalize=lowerCAmelCase__ , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) UpperCAmelCase_ = prepare_img() UpperCAmelCase_ = transform(lowerCAmelCase__ ).unsqueeze(0 ) UpperCAmelCase_ = processor(lowerCAmelCase__ , return_tensors="pt" ).pixel_values # verify pixel values assert torch.allclose(lowerCAmelCase__ , lowerCAmelCase__ ) # verify logits with torch.no_grad(): UpperCAmelCase_ = model(lowerCAmelCase__ ) UpperCAmelCase_ = outputs.logits print("Logits:" , logits[0, :3] ) print("Predicted class:" , model.config.idalabel[logits.argmax(-1 ).item()] ) UpperCAmelCase_ = timm_model(lowerCAmelCase__ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(lowerCAmelCase__ , outputs.logits , atol=1e-3 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: Path(lowerCAmelCase__ ).mkdir(exist_ok=lowerCAmelCase__ ) print(f"""Saving model {model_name} and processor to {pytorch_dump_folder_path}""" ) model.save_pretrained(lowerCAmelCase__ ) processor.save_pretrained(lowerCAmelCase__ ) if push_to_hub: print(f"""Pushing model {model_name} and processor to the hub""" ) model.push_to_hub(f"""ybelkada/{model_name}""" ) processor.push_to_hub(f"""ybelkada/{model_name}""" ) if __name__ == "__main__": lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""resnetv2_50x1_bitm""", type=str, help="""Name of the BiT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to push the model to the hub.""", ) lowerCamelCase = parser.parse_args() convert_bit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
82
1
"""simple docstring""" from typing import Optional import torch import torch.utils.checkpoint from torch import Tensor, nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_outputs import ( BaseModelOutputWithNoAttention, BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention, ) from ...modeling_utils import PreTrainedModel from ...utils import logging from .configuration_regnet import RegNetConfig lowerCamelCase = logging.get_logger(__name__) # General docstring lowerCamelCase = """RegNetConfig""" # Base docstring lowerCamelCase = """facebook/regnet-y-040""" lowerCamelCase = [1, 1_088, 7, 7] # Image classification docstring lowerCamelCase = """facebook/regnet-y-040""" lowerCamelCase = """tabby, tabby cat""" lowerCamelCase = [ """facebook/regnet-y-040""", # See all regnet models at https://huggingface.co/models?filter=regnet ] class lowercase__ ( nn.Module ): '''simple docstring''' def __init__( self : Optional[Any] , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int = 3 , _UpperCAmelCase : int = 1 , _UpperCAmelCase : int = 1 , _UpperCAmelCase : Optional[str] = "relu" , ) -> Tuple: '''simple docstring''' super().__init__() UpperCAmelCase_ = nn.Convad( _UpperCAmelCase , _UpperCAmelCase , kernel_size=_UpperCAmelCase , stride=_UpperCAmelCase , padding=kernel_size // 2 , groups=_UpperCAmelCase , bias=_UpperCAmelCase , ) UpperCAmelCase_ = nn.BatchNormad(_UpperCAmelCase ) UpperCAmelCase_ = ACTaFN[activation] if activation is not None else nn.Identity() def lowercase__ ( self : List[str] , _UpperCAmelCase : Optional[Any] ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase_ = self.convolution(_UpperCAmelCase ) UpperCAmelCase_ = self.normalization(_UpperCAmelCase ) UpperCAmelCase_ = self.activation(_UpperCAmelCase ) return hidden_state class lowercase__ ( nn.Module ): '''simple docstring''' def __init__( self : Union[str, Any] , _UpperCAmelCase : RegNetConfig ) -> Any: '''simple docstring''' super().__init__() UpperCAmelCase_ = RegNetConvLayer( config.num_channels , config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act ) UpperCAmelCase_ = config.num_channels def lowercase__ ( self : int , _UpperCAmelCase : Optional[int] ) -> List[str]: '''simple docstring''' UpperCAmelCase_ = pixel_values.shape[1] if num_channels != self.num_channels: raise ValueError( "Make sure that the channel dimension of the pixel values match with the one set in the configuration." ) UpperCAmelCase_ = self.embedder(_UpperCAmelCase ) return hidden_state class lowercase__ ( nn.Module ): '''simple docstring''' def __init__( self : List[str] , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int = 2 ) -> List[str]: '''simple docstring''' super().__init__() UpperCAmelCase_ = nn.Convad(_UpperCAmelCase , _UpperCAmelCase , kernel_size=1 , stride=_UpperCAmelCase , bias=_UpperCAmelCase ) UpperCAmelCase_ = nn.BatchNormad(_UpperCAmelCase ) def lowercase__ ( self : int , _UpperCAmelCase : Tensor ) -> Tensor: '''simple docstring''' UpperCAmelCase_ = self.convolution(_UpperCAmelCase ) UpperCAmelCase_ = self.normalization(_UpperCAmelCase ) return hidden_state class lowercase__ ( nn.Module ): '''simple docstring''' def __init__( self : List[str] , _UpperCAmelCase : int , _UpperCAmelCase : int ) -> List[str]: '''simple docstring''' super().__init__() UpperCAmelCase_ = nn.AdaptiveAvgPoolad((1, 1) ) UpperCAmelCase_ = nn.Sequential( nn.Convad(_UpperCAmelCase , _UpperCAmelCase , kernel_size=1 ) , nn.ReLU() , nn.Convad(_UpperCAmelCase , _UpperCAmelCase , kernel_size=1 ) , nn.Sigmoid() , ) def lowercase__ ( self : str , _UpperCAmelCase : Optional[int] ) -> Optional[int]: '''simple docstring''' UpperCAmelCase_ = self.pooler(_UpperCAmelCase ) UpperCAmelCase_ = self.attention(_UpperCAmelCase ) UpperCAmelCase_ = hidden_state * attention return hidden_state class lowercase__ ( nn.Module ): '''simple docstring''' def __init__( self : Optional[int] , _UpperCAmelCase : RegNetConfig , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int = 1 ) -> str: '''simple docstring''' super().__init__() UpperCAmelCase_ = in_channels != out_channels or stride != 1 UpperCAmelCase_ = max(1 , out_channels // config.groups_width ) UpperCAmelCase_ = ( RegNetShortCut(_UpperCAmelCase , _UpperCAmelCase , stride=_UpperCAmelCase ) if should_apply_shortcut else nn.Identity() ) UpperCAmelCase_ = nn.Sequential( RegNetConvLayer(_UpperCAmelCase , _UpperCAmelCase , kernel_size=1 , activation=config.hidden_act ) , RegNetConvLayer(_UpperCAmelCase , _UpperCAmelCase , stride=_UpperCAmelCase , groups=_UpperCAmelCase , activation=config.hidden_act ) , RegNetConvLayer(_UpperCAmelCase , _UpperCAmelCase , kernel_size=1 , activation=_UpperCAmelCase ) , ) UpperCAmelCase_ = ACTaFN[config.hidden_act] def lowercase__ ( self : str , _UpperCAmelCase : List[Any] ) -> List[Any]: '''simple docstring''' UpperCAmelCase_ = hidden_state UpperCAmelCase_ = self.layer(_UpperCAmelCase ) UpperCAmelCase_ = self.shortcut(_UpperCAmelCase ) hidden_state += residual UpperCAmelCase_ = self.activation(_UpperCAmelCase ) return hidden_state class lowercase__ ( nn.Module ): '''simple docstring''' def __init__( self : str , _UpperCAmelCase : RegNetConfig , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int = 1 ) -> Optional[Any]: '''simple docstring''' super().__init__() UpperCAmelCase_ = in_channels != out_channels or stride != 1 UpperCAmelCase_ = max(1 , out_channels // config.groups_width ) UpperCAmelCase_ = ( RegNetShortCut(_UpperCAmelCase , _UpperCAmelCase , stride=_UpperCAmelCase ) if should_apply_shortcut else nn.Identity() ) UpperCAmelCase_ = nn.Sequential( RegNetConvLayer(_UpperCAmelCase , _UpperCAmelCase , kernel_size=1 , activation=config.hidden_act ) , RegNetConvLayer(_UpperCAmelCase , _UpperCAmelCase , stride=_UpperCAmelCase , groups=_UpperCAmelCase , activation=config.hidden_act ) , RegNetSELayer(_UpperCAmelCase , reduced_channels=int(round(in_channels / 4 ) ) ) , RegNetConvLayer(_UpperCAmelCase , _UpperCAmelCase , kernel_size=1 , activation=_UpperCAmelCase ) , ) UpperCAmelCase_ = ACTaFN[config.hidden_act] def lowercase__ ( self : Any , _UpperCAmelCase : int ) -> int: '''simple docstring''' UpperCAmelCase_ = hidden_state UpperCAmelCase_ = self.layer(_UpperCAmelCase ) UpperCAmelCase_ = self.shortcut(_UpperCAmelCase ) hidden_state += residual UpperCAmelCase_ = self.activation(_UpperCAmelCase ) return hidden_state class lowercase__ ( nn.Module ): '''simple docstring''' def __init__( self : str , _UpperCAmelCase : RegNetConfig , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int = 2 , _UpperCAmelCase : int = 2 , ) -> List[str]: '''simple docstring''' super().__init__() UpperCAmelCase_ = RegNetXLayer if config.layer_type == "x" else RegNetYLayer UpperCAmelCase_ = nn.Sequential( # downsampling is done in the first layer with stride of 2 layer( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , stride=_UpperCAmelCase , ) , *[layer(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) for _ in range(depth - 1 )] , ) def lowercase__ ( self : Tuple , _UpperCAmelCase : str ) -> List[str]: '''simple docstring''' UpperCAmelCase_ = self.layers(_UpperCAmelCase ) return hidden_state class lowercase__ ( nn.Module ): '''simple docstring''' def __init__( self : int , _UpperCAmelCase : RegNetConfig ) -> Tuple: '''simple docstring''' super().__init__() UpperCAmelCase_ = nn.ModuleList([] ) # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( RegNetStage( _UpperCAmelCase , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , ) ) UpperCAmelCase_ = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for (in_channels, out_channels), depth in zip(_UpperCAmelCase , config.depths[1:] ): self.stages.append(RegNetStage(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , depth=_UpperCAmelCase ) ) def lowercase__ ( self : Optional[Any] , _UpperCAmelCase : Tensor , _UpperCAmelCase : bool = False , _UpperCAmelCase : bool = True ) -> BaseModelOutputWithNoAttention: '''simple docstring''' UpperCAmelCase_ = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: UpperCAmelCase_ = hidden_states + (hidden_state,) UpperCAmelCase_ = stage_module(_UpperCAmelCase ) if output_hidden_states: UpperCAmelCase_ = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return BaseModelOutputWithNoAttention(last_hidden_state=_UpperCAmelCase , hidden_states=_UpperCAmelCase ) class lowercase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' UpperCamelCase = RegNetConfig UpperCamelCase = '''regnet''' UpperCamelCase = '''pixel_values''' UpperCamelCase = True def lowercase__ ( self : Any , _UpperCAmelCase : Union[str, Any] ) -> Tuple: '''simple docstring''' if isinstance(_UpperCAmelCase , nn.Convad ): nn.init.kaiming_normal_(module.weight , mode="fan_out" , nonlinearity="relu" ) elif isinstance(_UpperCAmelCase , (nn.BatchNormad, nn.GroupNorm) ): nn.init.constant_(module.weight , 1 ) nn.init.constant_(module.bias , 0 ) def lowercase__ ( self : List[str] , _UpperCAmelCase : List[str] , _UpperCAmelCase : List[Any]=False ) -> Optional[Any]: '''simple docstring''' if isinstance(_UpperCAmelCase , _UpperCAmelCase ): UpperCAmelCase_ = value lowerCamelCase = r""" This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`RegNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. """ lowerCamelCase = r""" Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConvNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~file_utils.ModelOutput`] instead of a plain tuple. """ @add_start_docstrings( '''The bare RegNet model outputting raw features without any specific head on top.''' , SCREAMING_SNAKE_CASE , ) # Copied from transformers.models.resnet.modeling_resnet.ResNetModel with RESNET->REGNET,ResNet->RegNet class lowercase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self : Tuple , _UpperCAmelCase : Optional[int] ) -> List[Any]: '''simple docstring''' super().__init__(_UpperCAmelCase ) UpperCAmelCase_ = config UpperCAmelCase_ = RegNetEmbeddings(_UpperCAmelCase ) UpperCAmelCase_ = RegNetEncoder(_UpperCAmelCase ) UpperCAmelCase_ = nn.AdaptiveAvgPoolad((1, 1) ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(_UpperCAmelCase ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=_UpperCAmelCase , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def lowercase__ ( self : List[str] , _UpperCAmelCase : Tensor , _UpperCAmelCase : Optional[bool] = None , _UpperCAmelCase : Optional[bool] = None ) -> BaseModelOutputWithPoolingAndNoAttention: '''simple docstring''' UpperCAmelCase_ = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) UpperCAmelCase_ = return_dict if return_dict is not None else self.config.use_return_dict UpperCAmelCase_ = self.embedder(_UpperCAmelCase ) UpperCAmelCase_ = self.encoder( _UpperCAmelCase , output_hidden_states=_UpperCAmelCase , return_dict=_UpperCAmelCase ) UpperCAmelCase_ = encoder_outputs[0] UpperCAmelCase_ = self.pooler(_UpperCAmelCase ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=_UpperCAmelCase , pooler_output=_UpperCAmelCase , hidden_states=encoder_outputs.hidden_states , ) @add_start_docstrings( ''' RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for ImageNet. ''' , SCREAMING_SNAKE_CASE , ) # Copied from transformers.models.resnet.modeling_resnet.ResNetForImageClassification with RESNET->REGNET,ResNet->RegNet,resnet->regnet class lowercase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self : Any , _UpperCAmelCase : int ) -> List[str]: '''simple docstring''' super().__init__(_UpperCAmelCase ) UpperCAmelCase_ = config.num_labels UpperCAmelCase_ = RegNetModel(_UpperCAmelCase ) # classification head UpperCAmelCase_ = nn.Sequential( nn.Flatten() , nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() , ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(_UpperCAmelCase ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=_UpperCAmelCase , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def lowercase__ ( self : int , _UpperCAmelCase : Optional[torch.FloatTensor] = None , _UpperCAmelCase : Optional[torch.LongTensor] = None , _UpperCAmelCase : Optional[bool] = None , _UpperCAmelCase : Optional[bool] = None , ) -> ImageClassifierOutputWithNoAttention: '''simple docstring''' UpperCAmelCase_ = return_dict if return_dict is not None else self.config.use_return_dict UpperCAmelCase_ = self.regnet(_UpperCAmelCase , output_hidden_states=_UpperCAmelCase , return_dict=_UpperCAmelCase ) UpperCAmelCase_ = outputs.pooler_output if return_dict else outputs[1] UpperCAmelCase_ = self.classifier(_UpperCAmelCase ) UpperCAmelCase_ = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: UpperCAmelCase_ = "regression" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): UpperCAmelCase_ = "single_label_classification" else: UpperCAmelCase_ = "multi_label_classification" if self.config.problem_type == "regression": UpperCAmelCase_ = MSELoss() if self.num_labels == 1: UpperCAmelCase_ = loss_fct(logits.squeeze() , labels.squeeze() ) else: UpperCAmelCase_ = loss_fct(_UpperCAmelCase , _UpperCAmelCase ) elif self.config.problem_type == "single_label_classification": UpperCAmelCase_ = CrossEntropyLoss() UpperCAmelCase_ = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": UpperCAmelCase_ = BCEWithLogitsLoss() UpperCAmelCase_ = loss_fct(_UpperCAmelCase , _UpperCAmelCase ) if not return_dict: UpperCAmelCase_ = (logits,) + outputs[2:] return (loss,) + output if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=_UpperCAmelCase , logits=_UpperCAmelCase , hidden_states=outputs.hidden_states )
82
"""simple docstring""" from bisect import bisect from itertools import accumulate def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = sorted(zip(lowerCAmelCase__ , lowerCAmelCase__ ) , key=lambda lowerCAmelCase__ : x[0] / x[1] , reverse=lowerCAmelCase__ ) UpperCAmelCase_ , UpperCAmelCase_ = [i[0] for i in r], [i[1] for i in r] UpperCAmelCase_ = list(accumulate(lowerCAmelCase__ ) ) UpperCAmelCase_ = bisect(lowerCAmelCase__ , lowerCAmelCase__ ) return ( 0 if k == 0 else sum(vl[:k] ) + (w - acc[k - 1]) * (vl[k]) / (wt[k]) if k != n else sum(vl[:k] ) ) if __name__ == "__main__": import doctest doctest.testmod()
82
1
"""simple docstring""" def a__ ( lowerCAmelCase__ ): UpperCAmelCase_ = len(lowerCAmelCase__ ) UpperCAmelCase_ = sum(lowerCAmelCase__ ) UpperCAmelCase_ = [[False for x in range(s + 1 )] for y in range(n + 1 )] for i in range(1 , n + 1 ): UpperCAmelCase_ = True for i in range(1 , s + 1 ): UpperCAmelCase_ = False for i in range(1 , n + 1 ): for j in range(1 , s + 1 ): UpperCAmelCase_ = dp[i][j - 1] if arr[i - 1] <= j: UpperCAmelCase_ = dp[i][j] or dp[i - 1][j - arr[i - 1]] for j in range(int(s / 2 ) , -1 , -1 ): if dp[n][j] is True: UpperCAmelCase_ = s - 2 * j break return diff
82
"""simple docstring""" import argparse import collections import json import os import re import string import sys import numpy as np lowerCamelCase = re.compile(r"""\b(a|an|the)\b""", re.UNICODE) lowerCamelCase = None def a__ ( ): UpperCAmelCase_ = argparse.ArgumentParser("Official evaluation script for SQuAD version 2.0." ) parser.add_argument("data_file" , metavar="data.json" , help="Input data JSON file." ) parser.add_argument("pred_file" , metavar="pred.json" , help="Model predictions." ) parser.add_argument( "--out-file" , "-o" , metavar="eval.json" , help="Write accuracy metrics to file (default is stdout)." ) parser.add_argument( "--na-prob-file" , "-n" , metavar="na_prob.json" , help="Model estimates of probability of no answer." ) parser.add_argument( "--na-prob-thresh" , "-t" , type=lowerCAmelCase__ , default=1.0 , help="Predict \"\" if no-answer probability exceeds this (default = 1.0)." , ) parser.add_argument( "--out-image-dir" , "-p" , metavar="out_images" , default=lowerCAmelCase__ , help="Save precision-recall curves to directory." ) parser.add_argument("--verbose" , "-v" , action="store_true" ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def a__ ( lowerCAmelCase__ ): UpperCAmelCase_ = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: UpperCAmelCase_ = bool(qa["answers"]["text"] ) return qid_to_has_ans def a__ ( lowerCAmelCase__ ): def remove_articles(lowerCAmelCase__ ): return ARTICLES_REGEX.sub(" " , lowerCAmelCase__ ) def white_space_fix(lowerCAmelCase__ ): return " ".join(text.split() ) def remove_punc(lowerCAmelCase__ ): UpperCAmelCase_ = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(lowerCAmelCase__ ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(lowerCAmelCase__ ) ) ) ) def a__ ( lowerCAmelCase__ ): if not s: return [] return normalize_answer(lowerCAmelCase__ ).split() def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ): return int(normalize_answer(lowerCAmelCase__ ) == normalize_answer(lowerCAmelCase__ ) ) def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = get_tokens(lowerCAmelCase__ ) UpperCAmelCase_ = get_tokens(lowerCAmelCase__ ) UpperCAmelCase_ = collections.Counter(lowerCAmelCase__ ) & collections.Counter(lowerCAmelCase__ ) UpperCAmelCase_ = sum(common.values() ) if len(lowerCAmelCase__ ) == 0 or len(lowerCAmelCase__ ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 UpperCAmelCase_ = 1.0 * num_same / len(lowerCAmelCase__ ) UpperCAmelCase_ = 1.0 * num_same / len(lowerCAmelCase__ ) UpperCAmelCase_ = (2 * precision * recall) / (precision + recall) return fa def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = {} UpperCAmelCase_ = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: UpperCAmelCase_ = qa["id"] UpperCAmelCase_ = [t for t in qa["answers"]["text"] if normalize_answer(lowerCAmelCase__ )] if not gold_answers: # For unanswerable questions, only correct answer is empty string UpperCAmelCase_ = [""] if qid not in preds: print(f"""Missing prediction for {qid}""" ) continue UpperCAmelCase_ = preds[qid] # Take max over all gold answers UpperCAmelCase_ = max(compute_exact(lowerCAmelCase__ , lowerCAmelCase__ ) for a in gold_answers ) UpperCAmelCase_ = max(compute_fa(lowerCAmelCase__ , lowerCAmelCase__ ) for a in gold_answers ) return exact_scores, fa_scores def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = {} for qid, s in scores.items(): UpperCAmelCase_ = na_probs[qid] > na_prob_thresh if pred_na: UpperCAmelCase_ = float(not qid_to_has_ans[qid] ) else: UpperCAmelCase_ = s return new_scores def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None ): if not qid_list: UpperCAmelCase_ = len(lowerCAmelCase__ ) return collections.OrderedDict( [ ("exact", 100.0 * sum(exact_scores.values() ) / total), ("f1", 100.0 * sum(fa_scores.values() ) / total), ("total", total), ] ) else: UpperCAmelCase_ = len(lowerCAmelCase__ ) return collections.OrderedDict( [ ("exact", 100.0 * sum(exact_scores[k] for k in qid_list ) / total), ("f1", 100.0 * sum(fa_scores[k] for k in qid_list ) / total), ("total", total), ] ) def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): for k in new_eval: UpperCAmelCase_ = new_eval[k] def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): plt.step(lowerCAmelCase__ , lowerCAmelCase__ , color="b" , alpha=0.2 , where="post" ) plt.fill_between(lowerCAmelCase__ , lowerCAmelCase__ , step="post" , alpha=0.2 , color="b" ) plt.xlabel("Recall" ) plt.ylabel("Precision" ) plt.xlim([0.0, 1.05] ) plt.ylim([0.0, 1.05] ) plt.title(lowerCAmelCase__ ) plt.savefig(lowerCAmelCase__ ) plt.clf() def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None ): UpperCAmelCase_ = sorted(lowerCAmelCase__ , key=lambda lowerCAmelCase__ : na_probs[k] ) UpperCAmelCase_ = 0.0 UpperCAmelCase_ = 1.0 UpperCAmelCase_ = 0.0 UpperCAmelCase_ = [1.0] UpperCAmelCase_ = [0.0] UpperCAmelCase_ = 0.0 for i, qid in enumerate(lowerCAmelCase__ ): if qid_to_has_ans[qid]: true_pos += scores[qid] UpperCAmelCase_ = true_pos / float(i + 1 ) UpperCAmelCase_ = true_pos / float(lowerCAmelCase__ ) if i == len(lowerCAmelCase__ ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(lowerCAmelCase__ ) recalls.append(lowerCAmelCase__ ) if out_image: plot_pr_curve(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) return {"ap": 100.0 * avg_prec} def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): if out_image_dir and not os.path.exists(lowerCAmelCase__ ): os.makedirs(lowerCAmelCase__ ) UpperCAmelCase_ = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return UpperCAmelCase_ = make_precision_recall_eval( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , out_image=os.path.join(lowerCAmelCase__ , "pr_exact.png" ) , title="Precision-Recall curve for Exact Match score" , ) UpperCAmelCase_ = make_precision_recall_eval( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , out_image=os.path.join(lowerCAmelCase__ , "pr_f1.png" ) , title="Precision-Recall curve for F1 score" , ) UpperCAmelCase_ = {k: float(lowerCAmelCase__ ) for k, v in qid_to_has_ans.items()} UpperCAmelCase_ = make_precision_recall_eval( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , out_image=os.path.join(lowerCAmelCase__ , "pr_oracle.png" ) , title="Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)" , ) merge_eval(lowerCAmelCase__ , lowerCAmelCase__ , "pr_exact" ) merge_eval(lowerCAmelCase__ , lowerCAmelCase__ , "pr_f1" ) merge_eval(lowerCAmelCase__ , lowerCAmelCase__ , "pr_oracle" ) def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): if not qid_list: return UpperCAmelCase_ = [na_probs[k] for k in qid_list] UpperCAmelCase_ = np.ones_like(lowerCAmelCase__ ) / float(len(lowerCAmelCase__ ) ) plt.hist(lowerCAmelCase__ , weights=lowerCAmelCase__ , bins=20 , range=(0.0, 1.0) ) plt.xlabel("Model probability of no-answer" ) plt.ylabel("Proportion of dataset" ) plt.title(f"""Histogram of no-answer probability: {name}""" ) plt.savefig(os.path.join(lowerCAmelCase__ , f"""na_prob_hist_{name}.png""" ) ) plt.clf() def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) UpperCAmelCase_ = num_no_ans UpperCAmelCase_ = cur_score UpperCAmelCase_ = 0.0 UpperCAmelCase_ = sorted(lowerCAmelCase__ , key=lambda lowerCAmelCase__ : na_probs[k] ) for i, qid in enumerate(lowerCAmelCase__ ): if qid not in scores: continue if qid_to_has_ans[qid]: UpperCAmelCase_ = scores[qid] else: if preds[qid]: UpperCAmelCase_ = -1 else: UpperCAmelCase_ = 0 cur_score += diff if cur_score > best_score: UpperCAmelCase_ = cur_score UpperCAmelCase_ = na_probs[qid] return 100.0 * best_score / len(lowerCAmelCase__ ), best_thresh def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ , UpperCAmelCase_ = find_best_thresh(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) UpperCAmelCase_ , UpperCAmelCase_ = find_best_thresh(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) UpperCAmelCase_ = best_exact UpperCAmelCase_ = exact_thresh UpperCAmelCase_ = best_fa UpperCAmelCase_ = fa_thresh def a__ ( ): with open(OPTS.data_file ) as f: UpperCAmelCase_ = json.load(lowerCAmelCase__ ) UpperCAmelCase_ = dataset_json["data"] with open(OPTS.pred_file ) as f: UpperCAmelCase_ = json.load(lowerCAmelCase__ ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: UpperCAmelCase_ = json.load(lowerCAmelCase__ ) else: UpperCAmelCase_ = {k: 0.0 for k in preds} UpperCAmelCase_ = make_qid_to_has_ans(lowerCAmelCase__ ) # maps qid to True/False UpperCAmelCase_ = [k for k, v in qid_to_has_ans.items() if v] UpperCAmelCase_ = [k for k, v in qid_to_has_ans.items() if not v] UpperCAmelCase_ , UpperCAmelCase_ = get_raw_scores(lowerCAmelCase__ , lowerCAmelCase__ ) UpperCAmelCase_ = apply_no_ans_threshold(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , OPTS.na_prob_thresh ) UpperCAmelCase_ = apply_no_ans_threshold(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , OPTS.na_prob_thresh ) UpperCAmelCase_ = make_eval_dict(lowerCAmelCase__ , lowerCAmelCase__ ) if has_ans_qids: UpperCAmelCase_ = make_eval_dict(lowerCAmelCase__ , lowerCAmelCase__ , qid_list=lowerCAmelCase__ ) merge_eval(lowerCAmelCase__ , lowerCAmelCase__ , "HasAns" ) if no_ans_qids: UpperCAmelCase_ = make_eval_dict(lowerCAmelCase__ , lowerCAmelCase__ , qid_list=lowerCAmelCase__ ) merge_eval(lowerCAmelCase__ , lowerCAmelCase__ , "NoAns" ) if OPTS.na_prob_file: find_all_best_thresh(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , OPTS.out_image_dir ) histogram_na_prob(lowerCAmelCase__ , lowerCAmelCase__ , OPTS.out_image_dir , "hasAns" ) histogram_na_prob(lowerCAmelCase__ , lowerCAmelCase__ , OPTS.out_image_dir , "noAns" ) if OPTS.out_file: with open(OPTS.out_file , "w" ) as f: json.dump(lowerCAmelCase__ , lowerCAmelCase__ ) else: print(json.dumps(lowerCAmelCase__ , indent=2 ) ) if __name__ == "__main__": lowerCamelCase = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use("""Agg""") import matplotlib.pyplot as plt main()
82
1
"""simple docstring""" import sacrebleu as scb from packaging import version from sacrebleu import CHRF import datasets lowerCamelCase = """\ @inproceedings{popovic-2015-chrf, title = \"chr{F}: character n-gram {F}-score for automatic {MT} evaluation\", author = \"Popovi{\'c}, Maja\", booktitle = \"Proceedings of the Tenth Workshop on Statistical Machine Translation\", month = sep, year = \"2015\", address = \"Lisbon, Portugal\", publisher = \"Association for Computational Linguistics\", url = \"https://aclanthology.org/W15-3049\", doi = \"10.18653/v1/W15-3049\", pages = \"392--395\", } @inproceedings{popovic-2017-chrf, title = \"chr{F}++: words helping character n-grams\", author = \"Popovi{\'c}, Maja\", booktitle = \"Proceedings of the Second Conference on Machine Translation\", month = sep, year = \"2017\", address = \"Copenhagen, Denmark\", publisher = \"Association for Computational Linguistics\", url = \"https://aclanthology.org/W17-4770\", doi = \"10.18653/v1/W17-4770\", pages = \"612--618\", } @inproceedings{post-2018-call, title = \"A Call for Clarity in Reporting {BLEU} Scores\", author = \"Post, Matt\", booktitle = \"Proceedings of the Third Conference on Machine Translation: Research Papers\", month = oct, year = \"2018\", address = \"Belgium, Brussels\", publisher = \"Association for Computational Linguistics\", url = \"https://www.aclweb.org/anthology/W18-6319\", pages = \"186--191\", } """ lowerCamelCase = """\ ChrF and ChrF++ are two MT evaluation metrics. They both use the F-score statistic for character n-gram matches, and ChrF++ adds word n-grams as well which correlates more strongly with direct assessment. We use the implementation that is already present in sacrebleu. The implementation here is slightly different from sacrebleu in terms of the required input format. The length of the references and hypotheses lists need to be the same, so you may need to transpose your references compared to sacrebleu's required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534 See the README.md file at https://github.com/mjpost/sacreBLEU#chrf--chrf for more information. """ lowerCamelCase = """ Produces ChrF(++) scores for hypotheses given reference translations. Args: predictions (list of str): The predicted sentences. references (list of list of str): The references. There should be one reference sub-list for each prediction sentence. char_order (int): Character n-gram order. Defaults to `6`. word_order (int): Word n-gram order. If equals to `2`, the metric is referred to as chrF++. Defaults to `0`. beta (int): Determine the importance of recall w.r.t precision. Defaults to `2`. lowercase (bool): if `True`, enables case-insensitivity. Defaults to `False`. whitespace (bool): If `True`, include whitespaces when extracting character n-grams. eps_smoothing (bool): If `True`, applies epsilon smoothing similar to reference chrF++.py, NLTK and Moses implementations. If `False`, it takes into account effective match order similar to sacreBLEU < 2.0.0. Defaults to `False`. Returns: 'score' (float): The chrF (chrF++) score, 'char_order' (int): The character n-gram order, 'word_order' (int): The word n-gram order. If equals to 2, the metric is referred to as chrF++, 'beta' (int): Determine the importance of recall w.r.t precision Examples: Example 1--a simple example of calculating chrF: >>> prediction = [\"The relationship between cats and dogs is not exactly friendly.\", \"a good bookshop is just a genteel black hole that knows how to read.\"] >>> reference = [[\"The relationship between dogs and cats is not exactly friendly.\"], [\"A good bookshop is just a genteel Black Hole that knows how to read.\"]] >>> chrf = datasets.load_metric(\"chrf\") >>> results = chrf.compute(predictions=prediction, references=reference) >>> print(results) {'score': 84.64214891738334, 'char_order': 6, 'word_order': 0, 'beta': 2} Example 2--the same example, but with the argument word_order=2, to calculate chrF++ instead of chrF: >>> prediction = [\"The relationship between cats and dogs is not exactly friendly.\", \"a good bookshop is just a genteel black hole that knows how to read.\"] >>> reference = [[\"The relationship between dogs and cats is not exactly friendly.\"], [\"A good bookshop is just a genteel Black Hole that knows how to read.\"]] >>> chrf = datasets.load_metric(\"chrf\") >>> results = chrf.compute(predictions=prediction, ... references=reference, ... word_order=2) >>> print(results) {'score': 82.87263732906315, 'char_order': 6, 'word_order': 2, 'beta': 2} Example 3--the same chrF++ example as above, but with `lowercase=True` to normalize all case: >>> prediction = [\"The relationship between cats and dogs is not exactly friendly.\", \"a good bookshop is just a genteel black hole that knows how to read.\"] >>> reference = [[\"The relationship between dogs and cats is not exactly friendly.\"], [\"A good bookshop is just a genteel Black Hole that knows how to read.\"]] >>> chrf = datasets.load_metric(\"chrf\") >>> results = chrf.compute(predictions=prediction, ... references=reference, ... word_order=2, ... lowercase=True) >>> print(results) {'score': 92.12853119829202, 'char_order': 6, 'word_order': 2, 'beta': 2} """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase__ ( datasets.Metric ): '''simple docstring''' def lowercase__ ( self : Dict ) -> Optional[Any]: '''simple docstring''' if version.parse(scb.__version__ ) < version.parse("1.4.12" ): raise ImportWarning( "To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn't match this condition.\n" "You can install it with `pip install \"sacrebleu>=1.4.12\"`." ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="https://github.com/mjpost/sacreBLEU#chrf--chrf" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Sequence(datasets.Value("string" , id="sequence" ) , id="references" ), } ) , codebase_urls=["https://github.com/mjpost/sacreBLEU#chrf--chrf"] , reference_urls=[ "https://github.com/m-popovic/chrF", ] , ) def lowercase__ ( self : Optional[int] , _UpperCAmelCase : List[str] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : int = CHRF.CHAR_ORDER , _UpperCAmelCase : int = CHRF.WORD_ORDER , _UpperCAmelCase : int = CHRF.BETA , _UpperCAmelCase : bool = False , _UpperCAmelCase : bool = False , _UpperCAmelCase : bool = False , ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase_ = len(references[0] ) if any(len(_UpperCAmelCase ) != references_per_prediction for refs in references ): raise ValueError("Sacrebleu requires the same number of references for each prediction" ) UpperCAmelCase_ = [[refs[i] for refs in references] for i in range(_UpperCAmelCase )] UpperCAmelCase_ = CHRF(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) UpperCAmelCase_ = sb_chrf.corpus_score(_UpperCAmelCase , _UpperCAmelCase ) return { "score": output.score, "char_order": output.char_order, "word_order": output.word_order, "beta": output.beta, }
82
"""simple docstring""" def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): return round(float(moles / volume ) * nfactor ) def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): return round(float((moles * 0.0821 * temperature) / (volume) ) ) def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): return round(float((moles * 0.0821 * temperature) / (pressure) ) ) def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): return round(float((pressure * volume) / (0.0821 * moles) ) ) if __name__ == "__main__": import doctest doctest.testmod()
82
1
"""simple docstring""" def a__ ( lowerCAmelCase__ ): return " ".join(input_str.split()[::-1] ) if __name__ == "__main__": import doctest doctest.testmod()
82
"""simple docstring""" from math import atan, cos, radians, sin, tan from .haversine_distance import haversine_distance lowerCamelCase = 6_378_137.0 lowerCamelCase = 6_356_752.314_245 lowerCamelCase = 6_378_137 def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = (AXIS_A - AXIS_B) / AXIS_A # Parametric latitudes # https://en.wikipedia.org/wiki/Latitude#Parametric_(or_reduced)_latitude UpperCAmelCase_ = atan((1 - flattening) * tan(radians(lowerCAmelCase__ ) ) ) UpperCAmelCase_ = atan((1 - flattening) * tan(radians(lowerCAmelCase__ ) ) ) # Compute central angle between two points # using haversine theta. sigma = haversine_distance / equatorial radius UpperCAmelCase_ = haversine_distance(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) / EQUATORIAL_RADIUS # Intermediate P and Q values UpperCAmelCase_ = (b_lata + b_lata) / 2 UpperCAmelCase_ = (b_lata - b_lata) / 2 # Intermediate X value # X = (sigma - sin(sigma)) * sin^2Pcos^2Q / cos^2(sigma/2) UpperCAmelCase_ = (sin(lowerCAmelCase__ ) ** 2) * (cos(lowerCAmelCase__ ) ** 2) UpperCAmelCase_ = cos(sigma / 2 ) ** 2 UpperCAmelCase_ = (sigma - sin(lowerCAmelCase__ )) * (x_numerator / x_demonimator) # Intermediate Y value # Y = (sigma + sin(sigma)) * cos^2Psin^2Q / sin^2(sigma/2) UpperCAmelCase_ = (cos(lowerCAmelCase__ ) ** 2) * (sin(lowerCAmelCase__ ) ** 2) UpperCAmelCase_ = sin(sigma / 2 ) ** 2 UpperCAmelCase_ = (sigma + sin(lowerCAmelCase__ )) * (y_numerator / y_denominator) return EQUATORIAL_RADIUS * (sigma - ((flattening / 2) * (x_value + y_value))) if __name__ == "__main__": import doctest doctest.testmod()
82
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase = logging.get_logger(__name__) lowerCamelCase = { """kssteven/ibert-roberta-base""": """https://huggingface.co/kssteven/ibert-roberta-base/resolve/main/config.json""", """kssteven/ibert-roberta-large""": """https://huggingface.co/kssteven/ibert-roberta-large/resolve/main/config.json""", """kssteven/ibert-roberta-large-mnli""": ( """https://huggingface.co/kssteven/ibert-roberta-large-mnli/resolve/main/config.json""" ), } class lowercase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' UpperCamelCase = '''ibert''' def __init__( self : int , _UpperCAmelCase : List[Any]=30522 , _UpperCAmelCase : Union[str, Any]=768 , _UpperCAmelCase : Optional[int]=12 , _UpperCAmelCase : List[Any]=12 , _UpperCAmelCase : List[str]=3072 , _UpperCAmelCase : Dict="gelu" , _UpperCAmelCase : Union[str, Any]=0.1 , _UpperCAmelCase : Dict=0.1 , _UpperCAmelCase : Optional[Any]=512 , _UpperCAmelCase : Optional[int]=2 , _UpperCAmelCase : str=0.02 , _UpperCAmelCase : Any=1e-12 , _UpperCAmelCase : int=1 , _UpperCAmelCase : Union[str, Any]=0 , _UpperCAmelCase : Tuple=2 , _UpperCAmelCase : Optional[int]="absolute" , _UpperCAmelCase : int=False , _UpperCAmelCase : Any="none" , **_UpperCAmelCase : Any , ) -> Union[str, Any]: '''simple docstring''' super().__init__(pad_token_id=_UpperCAmelCase , bos_token_id=_UpperCAmelCase , eos_token_id=_UpperCAmelCase , **_UpperCAmelCase ) UpperCAmelCase_ = vocab_size UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = hidden_act UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = type_vocab_size UpperCAmelCase_ = initializer_range UpperCAmelCase_ = layer_norm_eps UpperCAmelCase_ = position_embedding_type UpperCAmelCase_ = quant_mode UpperCAmelCase_ = force_dequant class lowercase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' @property def lowercase__ ( self : Any ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": UpperCAmelCase_ = {0: "batch", 1: "choice", 2: "sequence"} else: UpperCAmelCase_ = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
82
"""simple docstring""" import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class lowercase__ : '''simple docstring''' def __init__( self : List[Any] , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Tuple=2 , _UpperCAmelCase : Union[str, Any]=8 , _UpperCAmelCase : Union[str, Any]=True , _UpperCAmelCase : Optional[int]=True , _UpperCAmelCase : Tuple=True , _UpperCAmelCase : Any=True , _UpperCAmelCase : Optional[int]=99 , _UpperCAmelCase : Tuple=16 , _UpperCAmelCase : Union[str, Any]=5 , _UpperCAmelCase : Optional[Any]=2 , _UpperCAmelCase : Dict=36 , _UpperCAmelCase : List[str]="gelu" , _UpperCAmelCase : Optional[int]=0.0 , _UpperCAmelCase : Dict=0.0 , _UpperCAmelCase : Tuple=512 , _UpperCAmelCase : Optional[int]=16 , _UpperCAmelCase : List[Any]=2 , _UpperCAmelCase : str=0.02 , _UpperCAmelCase : Dict=3 , _UpperCAmelCase : List[str]=4 , _UpperCAmelCase : Optional[Any]=None , ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = seq_length UpperCAmelCase_ = is_training UpperCAmelCase_ = use_input_mask UpperCAmelCase_ = use_token_type_ids UpperCAmelCase_ = use_labels UpperCAmelCase_ = vocab_size UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_act UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = type_vocab_size UpperCAmelCase_ = type_sequence_label_size UpperCAmelCase_ = initializer_range UpperCAmelCase_ = num_labels UpperCAmelCase_ = num_choices UpperCAmelCase_ = scope def lowercase__ ( self : Union[str, Any] ) -> List[str]: '''simple docstring''' UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase_ = None if self.use_input_mask: UpperCAmelCase_ = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase_ = None if self.use_token_type_ids: UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase_ = None UpperCAmelCase_ = None UpperCAmelCase_ = None if self.use_labels: UpperCAmelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase_ = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase_ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowercase__ ( self : int ) -> Dict: '''simple docstring''' return MraConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_UpperCAmelCase , initializer_range=self.initializer_range , ) def lowercase__ ( self : Dict ) -> str: '''simple docstring''' UpperCAmelCase_ = self.get_config() UpperCAmelCase_ = 300 return config def lowercase__ ( self : int ) -> List[Any]: '''simple docstring''' ( ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ) = self.prepare_config_and_inputs() UpperCAmelCase_ = True UpperCAmelCase_ = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def lowercase__ ( self : int , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Tuple , _UpperCAmelCase : Tuple , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Tuple ) -> Any: '''simple docstring''' UpperCAmelCase_ = MraModel(config=_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() UpperCAmelCase_ = model(_UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase ) UpperCAmelCase_ = model(_UpperCAmelCase , token_type_ids=_UpperCAmelCase ) UpperCAmelCase_ = model(_UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase__ ( self : int , _UpperCAmelCase : Tuple , _UpperCAmelCase : Dict , _UpperCAmelCase : Dict , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : List[str] , _UpperCAmelCase : Dict , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Tuple , _UpperCAmelCase : Optional[int] , ) -> int: '''simple docstring''' UpperCAmelCase_ = True UpperCAmelCase_ = MraModel(_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() UpperCAmelCase_ = model( _UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase , encoder_hidden_states=_UpperCAmelCase , encoder_attention_mask=_UpperCAmelCase , ) UpperCAmelCase_ = model( _UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase , encoder_hidden_states=_UpperCAmelCase , ) UpperCAmelCase_ = model(_UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase__ ( self : List[Any] , _UpperCAmelCase : Tuple , _UpperCAmelCase : List[str] , _UpperCAmelCase : str , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : str , _UpperCAmelCase : Optional[int] ) -> Tuple: '''simple docstring''' UpperCAmelCase_ = MraForMaskedLM(config=_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() UpperCAmelCase_ = model(_UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase , labels=_UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowercase__ ( self : str , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Any , _UpperCAmelCase : str , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Tuple ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase_ = MraForQuestionAnswering(config=_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() UpperCAmelCase_ = model( _UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase , start_positions=_UpperCAmelCase , end_positions=_UpperCAmelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowercase__ ( self : Any , _UpperCAmelCase : Tuple , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : str , _UpperCAmelCase : int , _UpperCAmelCase : List[str] , _UpperCAmelCase : Any , _UpperCAmelCase : Dict ) -> Any: '''simple docstring''' UpperCAmelCase_ = self.num_labels UpperCAmelCase_ = MraForSequenceClassification(_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() UpperCAmelCase_ = model(_UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase , labels=_UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowercase__ ( self : Dict , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Any , _UpperCAmelCase : Any , _UpperCAmelCase : List[str] , _UpperCAmelCase : Tuple , _UpperCAmelCase : List[Any] ) -> int: '''simple docstring''' UpperCAmelCase_ = self.num_labels UpperCAmelCase_ = MraForTokenClassification(config=_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() UpperCAmelCase_ = model(_UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase , labels=_UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowercase__ ( self : Optional[int] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Tuple , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : int , _UpperCAmelCase : List[Any] , _UpperCAmelCase : Any , _UpperCAmelCase : List[str] ) -> str: '''simple docstring''' UpperCAmelCase_ = self.num_choices UpperCAmelCase_ = MraForMultipleChoice(config=_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() UpperCAmelCase_ = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase_ = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase_ = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase_ = model( _UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase , labels=_UpperCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowercase__ ( self : Optional[Any] ) -> Optional[int]: '''simple docstring''' UpperCAmelCase_ = self.prepare_config_and_inputs() ( ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ) = config_and_inputs UpperCAmelCase_ = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class lowercase__ ( SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' UpperCamelCase = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = () def lowercase__ ( self : Optional[int] ) -> Tuple: '''simple docstring''' UpperCAmelCase_ = MraModelTester(self ) UpperCAmelCase_ = ConfigTester(self , config_class=_UpperCAmelCase , hidden_size=37 ) def lowercase__ ( self : Optional[int] ) -> Dict: '''simple docstring''' self.config_tester.run_common_tests() def lowercase__ ( self : List[Any] ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_UpperCAmelCase ) def lowercase__ ( self : Tuple ) -> Tuple: '''simple docstring''' UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCAmelCase_ = type self.model_tester.create_and_check_model(*_UpperCAmelCase ) def lowercase__ ( self : List[str] ) -> str: '''simple docstring''' UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_UpperCAmelCase ) def lowercase__ ( self : Optional[int] ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*_UpperCAmelCase ) def lowercase__ ( self : Any ) -> List[Any]: '''simple docstring''' UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_UpperCAmelCase ) def lowercase__ ( self : Optional[int] ) -> str: '''simple docstring''' UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_UpperCAmelCase ) def lowercase__ ( self : str ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_UpperCAmelCase ) @slow def lowercase__ ( self : Union[str, Any] ) -> Tuple: '''simple docstring''' for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ = MraModel.from_pretrained(_UpperCAmelCase ) self.assertIsNotNone(_UpperCAmelCase ) @unittest.skip(reason="MRA does not output attentions" ) def lowercase__ ( self : Optional[Any] ) -> Any: '''simple docstring''' return @require_torch class lowercase__ ( unittest.TestCase ): '''simple docstring''' @slow def lowercase__ ( self : Any ) -> List[str]: '''simple docstring''' UpperCAmelCase_ = MraModel.from_pretrained("uw-madison/mra-base-512-4" ) UpperCAmelCase_ = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): UpperCAmelCase_ = model(_UpperCAmelCase )[0] UpperCAmelCase_ = torch.Size((1, 256, 768) ) self.assertEqual(output.shape , _UpperCAmelCase ) UpperCAmelCase_ = torch.tensor( [[[-0.0140, 0.0830, -0.0381], [0.1546, 0.1402, 0.0220], [0.1162, 0.0851, 0.0165]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , _UpperCAmelCase , atol=1e-4 ) ) @slow def lowercase__ ( self : Union[str, Any] ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase_ = MraForMaskedLM.from_pretrained("uw-madison/mra-base-512-4" ) UpperCAmelCase_ = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): UpperCAmelCase_ = model(_UpperCAmelCase )[0] UpperCAmelCase_ = 50265 UpperCAmelCase_ = torch.Size((1, 256, vocab_size) ) self.assertEqual(output.shape , _UpperCAmelCase ) UpperCAmelCase_ = torch.tensor( [[[9.2595, -3.6038, 11.8819], [9.3869, -3.2693, 11.0956], [11.8524, -3.4938, 13.1210]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , _UpperCAmelCase , atol=1e-4 ) ) @slow def lowercase__ ( self : Tuple ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase_ = MraForMaskedLM.from_pretrained("uw-madison/mra-base-4096-8-d3" ) UpperCAmelCase_ = torch.arange(4096 ).unsqueeze(0 ) with torch.no_grad(): UpperCAmelCase_ = model(_UpperCAmelCase )[0] UpperCAmelCase_ = 50265 UpperCAmelCase_ = torch.Size((1, 4096, vocab_size) ) self.assertEqual(output.shape , _UpperCAmelCase ) UpperCAmelCase_ = torch.tensor( [[[5.4789, -2.3564, 7.5064], [7.9067, -1.3369, 9.9668], [9.0712, -1.8106, 7.0380]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , _UpperCAmelCase , atol=1e-4 ) )
82
1
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision.transforms import functional as F from transformers import DetrImageProcessor, TableTransformerConfig, TableTransformerForObjectDetection from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) lowerCamelCase = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F"transformer.encoder.layers.{i}.self_attn.out_proj.weight", F"encoder.layers.{i}.self_attn.out_proj.weight") ) rename_keys.append( (F"transformer.encoder.layers.{i}.self_attn.out_proj.bias", F"encoder.layers.{i}.self_attn.out_proj.bias") ) rename_keys.append((F"transformer.encoder.layers.{i}.linear1.weight", F"encoder.layers.{i}.fc1.weight")) rename_keys.append((F"transformer.encoder.layers.{i}.linear1.bias", F"encoder.layers.{i}.fc1.bias")) rename_keys.append((F"transformer.encoder.layers.{i}.linear2.weight", F"encoder.layers.{i}.fc2.weight")) rename_keys.append((F"transformer.encoder.layers.{i}.linear2.bias", F"encoder.layers.{i}.fc2.bias")) rename_keys.append( (F"transformer.encoder.layers.{i}.norm1.weight", F"encoder.layers.{i}.self_attn_layer_norm.weight") ) rename_keys.append((F"transformer.encoder.layers.{i}.norm1.bias", F"encoder.layers.{i}.self_attn_layer_norm.bias")) rename_keys.append((F"transformer.encoder.layers.{i}.norm2.weight", F"encoder.layers.{i}.final_layer_norm.weight")) rename_keys.append((F"transformer.encoder.layers.{i}.norm2.bias", F"encoder.layers.{i}.final_layer_norm.bias")) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (F"transformer.decoder.layers.{i}.self_attn.out_proj.weight", F"decoder.layers.{i}.self_attn.out_proj.weight") ) rename_keys.append( (F"transformer.decoder.layers.{i}.self_attn.out_proj.bias", F"decoder.layers.{i}.self_attn.out_proj.bias") ) rename_keys.append( ( F"transformer.decoder.layers.{i}.multihead_attn.out_proj.weight", F"decoder.layers.{i}.encoder_attn.out_proj.weight", ) ) rename_keys.append( ( F"transformer.decoder.layers.{i}.multihead_attn.out_proj.bias", F"decoder.layers.{i}.encoder_attn.out_proj.bias", ) ) rename_keys.append((F"transformer.decoder.layers.{i}.linear1.weight", F"decoder.layers.{i}.fc1.weight")) rename_keys.append((F"transformer.decoder.layers.{i}.linear1.bias", F"decoder.layers.{i}.fc1.bias")) rename_keys.append((F"transformer.decoder.layers.{i}.linear2.weight", F"decoder.layers.{i}.fc2.weight")) rename_keys.append((F"transformer.decoder.layers.{i}.linear2.bias", F"decoder.layers.{i}.fc2.bias")) rename_keys.append( (F"transformer.decoder.layers.{i}.norm1.weight", F"decoder.layers.{i}.self_attn_layer_norm.weight") ) rename_keys.append((F"transformer.decoder.layers.{i}.norm1.bias", F"decoder.layers.{i}.self_attn_layer_norm.bias")) rename_keys.append( (F"transformer.decoder.layers.{i}.norm2.weight", F"decoder.layers.{i}.encoder_attn_layer_norm.weight") ) rename_keys.append( (F"transformer.decoder.layers.{i}.norm2.bias", F"decoder.layers.{i}.encoder_attn_layer_norm.bias") ) rename_keys.append((F"transformer.decoder.layers.{i}.norm3.weight", F"decoder.layers.{i}.final_layer_norm.weight")) rename_keys.append((F"transformer.decoder.layers.{i}.norm3.bias", F"decoder.layers.{i}.final_layer_norm.bias")) # convolutional projection + query embeddings + layernorm of encoder + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("""input_proj.weight""", """input_projection.weight"""), ("""input_proj.bias""", """input_projection.bias"""), ("""query_embed.weight""", """query_position_embeddings.weight"""), ("""transformer.encoder.norm.weight""", """encoder.layernorm.weight"""), ("""transformer.encoder.norm.bias""", """encoder.layernorm.bias"""), ("""transformer.decoder.norm.weight""", """decoder.layernorm.weight"""), ("""transformer.decoder.norm.bias""", """decoder.layernorm.bias"""), ("""class_embed.weight""", """class_labels_classifier.weight"""), ("""class_embed.bias""", """class_labels_classifier.bias"""), ("""bbox_embed.layers.0.weight""", """bbox_predictor.layers.0.weight"""), ("""bbox_embed.layers.0.bias""", """bbox_predictor.layers.0.bias"""), ("""bbox_embed.layers.1.weight""", """bbox_predictor.layers.1.weight"""), ("""bbox_embed.layers.1.bias""", """bbox_predictor.layers.1.bias"""), ("""bbox_embed.layers.2.weight""", """bbox_predictor.layers.2.weight"""), ("""bbox_embed.layers.2.bias""", """bbox_predictor.layers.2.bias"""), ] ) def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = state_dict.pop(lowerCAmelCase__ ) UpperCAmelCase_ = val def a__ ( lowerCAmelCase__ ): UpperCAmelCase_ = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: UpperCAmelCase_ = key.replace("backbone.0.body" , "backbone.conv_encoder.model" ) UpperCAmelCase_ = value else: UpperCAmelCase_ = value return new_state_dict def a__ ( lowerCAmelCase__ ): UpperCAmelCase_ = "" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) UpperCAmelCase_ = state_dict.pop(f"""{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight""" ) UpperCAmelCase_ = state_dict.pop(f"""{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase_ = in_proj_weight[:256, :] UpperCAmelCase_ = in_proj_bias[:256] UpperCAmelCase_ = in_proj_weight[256:512, :] UpperCAmelCase_ = in_proj_bias[256:512] UpperCAmelCase_ = in_proj_weight[-256:, :] UpperCAmelCase_ = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention UpperCAmelCase_ = state_dict.pop(f"""{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight""" ) UpperCAmelCase_ = state_dict.pop(f"""{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase_ = in_proj_weight[:256, :] UpperCAmelCase_ = in_proj_bias[:256] UpperCAmelCase_ = in_proj_weight[256:512, :] UpperCAmelCase_ = in_proj_bias[256:512] UpperCAmelCase_ = in_proj_weight[-256:, :] UpperCAmelCase_ = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention UpperCAmelCase_ = state_dict.pop( f"""{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight""" ) UpperCAmelCase_ = state_dict.pop(f"""{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) of cross-attention to the state dict UpperCAmelCase_ = in_proj_weight_cross_attn[:256, :] UpperCAmelCase_ = in_proj_bias_cross_attn[:256] UpperCAmelCase_ = in_proj_weight_cross_attn[256:512, :] UpperCAmelCase_ = in_proj_bias_cross_attn[256:512] UpperCAmelCase_ = in_proj_weight_cross_attn[-256:, :] UpperCAmelCase_ = in_proj_bias_cross_attn[-256:] def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ , UpperCAmelCase_ = image.size UpperCAmelCase_ = max(lowerCAmelCase__ , lowerCAmelCase__ ) UpperCAmelCase_ = 800 if "detection" in checkpoint_url else 1000 UpperCAmelCase_ = target_max_size / current_max_size UpperCAmelCase_ = image.resize((int(round(scale * width ) ), int(round(scale * height ) )) ) return resized_image def a__ ( lowerCAmelCase__ ): UpperCAmelCase_ = F.to_tensor(lowerCAmelCase__ ) UpperCAmelCase_ = F.normalize(lowerCAmelCase__ , mean=[0.485, 0.456, 0.406] , std=[0.229, 0.224, 0.225] ) return image @torch.no_grad() def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): logger.info("Converting model..." ) # load original state dict UpperCAmelCase_ = torch.hub.load_state_dict_from_url(lowerCAmelCase__ , map_location="cpu" ) # rename keys for src, dest in rename_keys: rename_key(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) UpperCAmelCase_ = rename_backbone_keys(lowerCAmelCase__ ) # query, key and value matrices need special treatment read_in_q_k_v(lowerCAmelCase__ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them UpperCAmelCase_ = "model." for key in state_dict.copy().keys(): if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): UpperCAmelCase_ = state_dict.pop(lowerCAmelCase__ ) UpperCAmelCase_ = val # create HuggingFace model and load state dict UpperCAmelCase_ = TableTransformerConfig( backbone="resnet18" , mask_loss_coefficient=1 , dice_loss_coefficient=1 , ce_loss_coefficient=1 , bbox_loss_coefficient=5 , giou_loss_coefficient=2 , eos_coefficient=0.4 , class_cost=1 , bbox_cost=5 , giou_cost=2 , ) if "detection" in checkpoint_url: UpperCAmelCase_ = 15 UpperCAmelCase_ = 2 UpperCAmelCase_ = {0: "table", 1: "table rotated"} UpperCAmelCase_ = idalabel UpperCAmelCase_ = {v: k for k, v in idalabel.items()} else: UpperCAmelCase_ = 125 UpperCAmelCase_ = 6 UpperCAmelCase_ = { 0: "table", 1: "table column", 2: "table row", 3: "table column header", 4: "table projected row header", 5: "table spanning cell", } UpperCAmelCase_ = idalabel UpperCAmelCase_ = {v: k for k, v in idalabel.items()} UpperCAmelCase_ = DetrImageProcessor( format="coco_detection" , max_size=800 if "detection" in checkpoint_url else 1000 ) UpperCAmelCase_ = TableTransformerForObjectDetection(lowerCAmelCase__ ) model.load_state_dict(lowerCAmelCase__ ) model.eval() # verify our conversion UpperCAmelCase_ = "example_pdf.png" if "detection" in checkpoint_url else "example_table.png" UpperCAmelCase_ = hf_hub_download(repo_id="nielsr/example-pdf" , repo_type="dataset" , filename=lowerCAmelCase__ ) UpperCAmelCase_ = Image.open(lowerCAmelCase__ ).convert("RGB" ) UpperCAmelCase_ = normalize(resize(lowerCAmelCase__ , lowerCAmelCase__ ) ).unsqueeze(0 ) UpperCAmelCase_ = model(lowerCAmelCase__ ) if "detection" in checkpoint_url: UpperCAmelCase_ = (1, 15, 3) UpperCAmelCase_ = torch.tensor( [[-6.7897, -16.9985, 6.7937], [-8.0186, -22.2192, 6.9677], [-7.3117, -21.0708, 7.4055]] ) UpperCAmelCase_ = torch.tensor([[0.4867, 0.1767, 0.6732], [0.6718, 0.4479, 0.3830], [0.4716, 0.1760, 0.6364]] ) else: UpperCAmelCase_ = (1, 125, 7) UpperCAmelCase_ = torch.tensor( [[-18.1430, -8.3214, 4.8274], [-18.4685, -7.1361, -4.2667], [-26.3693, -9.3429, -4.9962]] ) UpperCAmelCase_ = torch.tensor([[0.4983, 0.5595, 0.9440], [0.4916, 0.6315, 0.5954], [0.6108, 0.8637, 0.1135]] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, :3, :3] , lowerCAmelCase__ , atol=1e-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , lowerCAmelCase__ , atol=1e-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(f"""Saving PyTorch model and image processor to {pytorch_dump_folder_path}...""" ) Path(lowerCAmelCase__ ).mkdir(exist_ok=lowerCAmelCase__ ) model.save_pretrained(lowerCAmelCase__ ) image_processor.save_pretrained(lowerCAmelCase__ ) if push_to_hub: # Push model to HF hub logger.info("Pushing model to the hub..." ) UpperCAmelCase_ = ( "microsoft/table-transformer-detection" if "detection" in checkpoint_url else "microsoft/table-transformer-structure-recognition" ) model.push_to_hub(lowerCAmelCase__ ) image_processor.push_to_hub(lowerCAmelCase__ ) if __name__ == "__main__": lowerCamelCase = argparse.ArgumentParser() parser.add_argument( """--checkpoint_url""", default="""https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", type=str, choices=[ """https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", """https://pubtables1m.blob.core.windows.net/model/pubtables1m_structure_detr_r18.pth""", ], help="""URL of the Table Transformer checkpoint you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) lowerCamelCase = parser.parse_args() convert_table_transformer_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
82
"""simple docstring""" import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration lowerCamelCase = 50_000 lowerCamelCase = 5_000 lowerCamelCase , lowerCamelCase = os.path.split(__file__) lowerCamelCase = os.path.join(RESULTS_BASEPATH, """results""", RESULTS_FILENAME.replace(""".py""", """.json""")) @get_duration def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ): for i in range(lowerCAmelCase__ ): UpperCAmelCase_ = dataset[i] @get_duration def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): for i in range(0 , len(lowerCAmelCase__ ) , lowerCAmelCase__ ): UpperCAmelCase_ = dataset[i : i + batch_size] @get_duration def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): with dataset.formatted_as(type=lowerCAmelCase__ ): for i in range(lowerCAmelCase__ ): UpperCAmelCase_ = dataset[i] @get_duration def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): with dataset.formatted_as(type=lowerCAmelCase__ ): for i in range(0 , lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = dataset[i : i + batch_size] def a__ ( ): UpperCAmelCase_ = {"num examples": SPEED_TEST_N_EXAMPLES} UpperCAmelCase_ = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted, {"type": "pandas", "length": SMALL_TEST}), (read_formatted, {"type": "torch", "length": SMALL_TEST}), (read_formatted, {"type": "tensorflow", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1000}), ] UpperCAmelCase_ = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print("generating dataset" ) UpperCAmelCase_ = datasets.Features( {"list": datasets.Sequence(datasets.Value("float32" ) ), "numbers": datasets.Value("float32" )} ) UpperCAmelCase_ = generate_example_dataset( os.path.join(lowerCAmelCase__ , "dataset.arrow" ) , lowerCAmelCase__ , num_examples=lowerCAmelCase__ , seq_shapes={"list": (100,)} , ) print("first set of iterations" ) for func, kwargs in functions: print(func.__name__ , str(lowerCAmelCase__ ) ) UpperCAmelCase_ = func(lowerCAmelCase__ , **lowerCAmelCase__ ) print("shuffling dataset" ) UpperCAmelCase_ = dataset.shuffle() print("Second set of iterations (after shuffling" ) for func, kwargs in functions_shuffled: print("shuffled " , func.__name__ , str(lowerCAmelCase__ ) ) UpperCAmelCase_ = func( lowerCAmelCase__ , **lowerCAmelCase__ ) with open(lowerCAmelCase__ , "wb" ) as f: f.write(json.dumps(lowerCAmelCase__ ).encode("utf-8" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
82
1
"""simple docstring""" import logging import os from typing import List, TextIO, Union from conllu import parse_incr from utils_ner import InputExample, Split, TokenClassificationTask lowerCamelCase = logging.getLogger(__name__) class lowercase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self : List[Any] , _UpperCAmelCase : str=-1 ) -> List[Any]: '''simple docstring''' UpperCAmelCase_ = label_idx def lowercase__ ( self : Optional[int] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Union[Split, str] ) -> List[InputExample]: '''simple docstring''' if isinstance(_UpperCAmelCase , _UpperCAmelCase ): UpperCAmelCase_ = mode.value UpperCAmelCase_ = os.path.join(_UpperCAmelCase , F"""{mode}.txt""" ) UpperCAmelCase_ = 1 UpperCAmelCase_ = [] with open(_UpperCAmelCase , encoding="utf-8" ) as f: UpperCAmelCase_ = [] UpperCAmelCase_ = [] for line in f: if line.startswith("-DOCSTART-" ) or line == "" or line == "\n": if words: examples.append(InputExample(guid=F"""{mode}-{guid_index}""" , words=_UpperCAmelCase , labels=_UpperCAmelCase ) ) guid_index += 1 UpperCAmelCase_ = [] UpperCAmelCase_ = [] else: UpperCAmelCase_ = line.split(" " ) words.append(splits[0] ) if len(_UpperCAmelCase ) > 1: labels.append(splits[self.label_idx].replace("\n" , "" ) ) else: # Examples could have no label for mode = "test" labels.append("O" ) if words: examples.append(InputExample(guid=F"""{mode}-{guid_index}""" , words=_UpperCAmelCase , labels=_UpperCAmelCase ) ) return examples def lowercase__ ( self : Dict , _UpperCAmelCase : TextIO , _UpperCAmelCase : TextIO , _UpperCAmelCase : List ) -> Dict: '''simple docstring''' UpperCAmelCase_ = 0 for line in test_input_reader: if line.startswith("-DOCSTART-" ) or line == "" or line == "\n": writer.write(_UpperCAmelCase ) if not preds_list[example_id]: example_id += 1 elif preds_list[example_id]: UpperCAmelCase_ = line.split()[0] + " " + preds_list[example_id].pop(0 ) + "\n" writer.write(_UpperCAmelCase ) else: logger.warning("Maximum sequence length exceeded: No prediction for '%s'." , line.split()[0] ) def lowercase__ ( self : str , _UpperCAmelCase : str ) -> List[str]: '''simple docstring''' if path: with open(_UpperCAmelCase , "r" ) as f: UpperCAmelCase_ = f.read().splitlines() if "O" not in labels: UpperCAmelCase_ = ["O"] + labels return labels else: return ["O", "B-MISC", "I-MISC", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC"] class lowercase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self : Optional[int] ) -> Optional[int]: '''simple docstring''' super().__init__(label_idx=-2 ) def lowercase__ ( self : Any , _UpperCAmelCase : str ) -> List[str]: '''simple docstring''' if path: with open(_UpperCAmelCase , "r" ) as f: UpperCAmelCase_ = f.read().splitlines() if "O" not in labels: UpperCAmelCase_ = ["O"] + labels return labels else: return [ "O", "B-ADVP", "B-INTJ", "B-LST", "B-PRT", "B-NP", "B-SBAR", "B-VP", "B-ADJP", "B-CONJP", "B-PP", "I-ADVP", "I-INTJ", "I-LST", "I-PRT", "I-NP", "I-SBAR", "I-VP", "I-ADJP", "I-CONJP", "I-PP", ] class lowercase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' def lowercase__ ( self : Optional[Any] , _UpperCAmelCase : Dict , _UpperCAmelCase : Union[Split, str] ) -> List[InputExample]: '''simple docstring''' if isinstance(_UpperCAmelCase , _UpperCAmelCase ): UpperCAmelCase_ = mode.value UpperCAmelCase_ = os.path.join(_UpperCAmelCase , F"""{mode}.txt""" ) UpperCAmelCase_ = 1 UpperCAmelCase_ = [] with open(_UpperCAmelCase , encoding="utf-8" ) as f: for sentence in parse_incr(_UpperCAmelCase ): UpperCAmelCase_ = [] UpperCAmelCase_ = [] for token in sentence: words.append(token["form"] ) labels.append(token["upos"] ) assert len(_UpperCAmelCase ) == len(_UpperCAmelCase ) if words: examples.append(InputExample(guid=F"""{mode}-{guid_index}""" , words=_UpperCAmelCase , labels=_UpperCAmelCase ) ) guid_index += 1 return examples def lowercase__ ( self : Optional[Any] , _UpperCAmelCase : TextIO , _UpperCAmelCase : TextIO , _UpperCAmelCase : List ) -> int: '''simple docstring''' UpperCAmelCase_ = 0 for sentence in parse_incr(_UpperCAmelCase ): UpperCAmelCase_ = preds_list[example_id] UpperCAmelCase_ = "" for token in sentence: out += F"""{token['form']} ({token['upos']}|{s_p.pop(0 )}) """ out += "\n" writer.write(_UpperCAmelCase ) example_id += 1 def lowercase__ ( self : List[str] , _UpperCAmelCase : str ) -> List[str]: '''simple docstring''' if path: with open(_UpperCAmelCase , "r" ) as f: return f.read().splitlines() else: return [ "ADJ", "ADP", "ADV", "AUX", "CCONJ", "DET", "INTJ", "NOUN", "NUM", "PART", "PRON", "PROPN", "PUNCT", "SCONJ", "SYM", "VERB", "X", ]
82
"""simple docstring""" import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Image from .base import TaskTemplate @dataclass(frozen=SCREAMING_SNAKE_CASE ) class lowercase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' UpperCamelCase = field(default='''image-classification''' , metadata={'''include_in_asdict_even_if_is_default''': True} ) UpperCamelCase = Features({'''image''': Image()} ) UpperCamelCase = Features({'''labels''': ClassLabel} ) UpperCamelCase = "image" UpperCamelCase = "labels" def lowercase__ ( self : str , _UpperCAmelCase : str ) -> Dict: '''simple docstring''' if self.label_column not in features: raise ValueError(F"""Column {self.label_column} is not present in features.""" ) if not isinstance(features[self.label_column] , _UpperCAmelCase ): raise ValueError(F"""Column {self.label_column} is not a ClassLabel.""" ) UpperCAmelCase_ = copy.deepcopy(self ) UpperCAmelCase_ = self.label_schema.copy() UpperCAmelCase_ = features[self.label_column] UpperCAmelCase_ = label_schema return task_template @property def lowercase__ ( self : List[str] ) -> Dict[str, str]: '''simple docstring''' return { self.image_column: "image", self.label_column: "labels", }
82
1
"""simple docstring""" def a__ ( lowerCAmelCase__ ): if edge <= 0 or not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): raise ValueError("Length must be a positive." ) return 3 * ((25 + 10 * (5 ** (1 / 2))) ** (1 / 2)) * (edge**2) def a__ ( lowerCAmelCase__ ): if edge <= 0 or not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): raise ValueError("Length must be a positive." ) return ((15 + (7 * (5 ** (1 / 2)))) / 4) * (edge**3) if __name__ == "__main__": import doctest doctest.testmod()
82
"""simple docstring""" import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device lowerCamelCase = False class lowercase__ ( unittest.TestCase ): '''simple docstring''' pass @nightly @require_torch_gpu class lowercase__ ( unittest.TestCase ): '''simple docstring''' def lowercase__ ( self : Dict ) -> Tuple: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase__ ( self : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase_ = VersatileDiffusionTextToImagePipeline.from_pretrained("shi-labs/versatile-diffusion" ) # remove text_unet pipe.remove_unused_weights() pipe.to(_UpperCAmelCase ) pipe.set_progress_bar_config(disable=_UpperCAmelCase ) UpperCAmelCase_ = "A painting of a squirrel eating a burger " UpperCAmelCase_ = torch.manual_seed(0 ) UpperCAmelCase_ = pipe( prompt=_UpperCAmelCase , generator=_UpperCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="numpy" ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(_UpperCAmelCase ) UpperCAmelCase_ = VersatileDiffusionTextToImagePipeline.from_pretrained(_UpperCAmelCase ) pipe.to(_UpperCAmelCase ) pipe.set_progress_bar_config(disable=_UpperCAmelCase ) UpperCAmelCase_ = generator.manual_seed(0 ) UpperCAmelCase_ = pipe( prompt=_UpperCAmelCase , generator=_UpperCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="numpy" ).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def lowercase__ ( self : Dict ) -> Optional[int]: '''simple docstring''' UpperCAmelCase_ = VersatileDiffusionTextToImagePipeline.from_pretrained( "shi-labs/versatile-diffusion" , torch_dtype=torch.floataa ) pipe.to(_UpperCAmelCase ) pipe.set_progress_bar_config(disable=_UpperCAmelCase ) UpperCAmelCase_ = "A painting of a squirrel eating a burger " UpperCAmelCase_ = torch.manual_seed(0 ) UpperCAmelCase_ = pipe( prompt=_UpperCAmelCase , generator=_UpperCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="numpy" ).images UpperCAmelCase_ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase_ = np.array([0.3367, 0.3169, 0.2656, 0.3870, 0.4790, 0.3796, 0.4009, 0.4878, 0.4778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
82
1
"""simple docstring""" import math_equivalence # From: git+https://github.com/hendrycks/math.git import datasets lowerCamelCase = """\ @article{hendrycksmath2021, title={Measuring Mathematical Problem Solving With the MATH Dataset}, author={Dan Hendrycks and Collin Burns and Saurav Kadavath and Akul Arora and Steven Basart and Eric Tang and Dawn Song and Jacob Steinhardt}, journal={arXiv preprint arXiv:2103.03874}, year={2021} } """ lowerCamelCase = """\ This metric is used to assess performance on the Mathematics Aptitude Test of Heuristics (MATH) dataset. It first canonicalizes the inputs (e.g., converting \"1/2\" to \"\\frac{1}{2}\") and then computes accuracy. """ lowerCamelCase = r""" Calculates accuracy after canonicalizing inputs. Args: predictions: list of predictions to score. Each prediction is a string that contains natural language and LaTex. references: list of reference for each prediction. Each reference is a string that contains natural language and LaTex. Returns: accuracy: accuracy after canonicalizing inputs (e.g., converting \"1/2\" to \"\\frac{1}{2}\") Examples: >>> metric = datasets.load_metric(\"competition_math\") >>> results = metric.compute(references=[\"\\frac{1}{2}\"], predictions=[\"1/2\"]) >>> print(results) {'accuracy': 1.0} """ @datasets.utils.file_utils.add_end_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase__ ( datasets.Metric ): '''simple docstring''' def lowercase__ ( self : int ) -> Tuple: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" ), "references": datasets.Value("string" ), } ) , homepage="https://github.com/hendrycks/math" , codebase_urls=["https://github.com/hendrycks/math"] , ) def lowercase__ ( self : Optional[int] , _UpperCAmelCase : Dict , _UpperCAmelCase : Any ) -> Any: '''simple docstring''' UpperCAmelCase_ = 0.0 for i, j in zip(_UpperCAmelCase , _UpperCAmelCase ): n_correct += 1.0 if math_equivalence.is_equiv(_UpperCAmelCase , _UpperCAmelCase ) else 0.0 UpperCAmelCase_ = n_correct / len(_UpperCAmelCase ) return { "accuracy": accuracy, }
82
"""simple docstring""" def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ): return x if y == 0 else greatest_common_divisor(lowerCAmelCase__ , x % y ) def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ): return (x * y) // greatest_common_divisor(lowerCAmelCase__ , lowerCAmelCase__ ) def a__ ( lowerCAmelCase__ = 20 ): UpperCAmelCase_ = 1 for i in range(1 , n + 1 ): UpperCAmelCase_ = lcm(lowerCAmelCase__ , lowerCAmelCase__ ) return g if __name__ == "__main__": print(F"{solution() = }")
82
1
"""simple docstring""" from __future__ import annotations class lowercase__ : '''simple docstring''' def __init__( self : List[str] , _UpperCAmelCase : int ) -> None: '''simple docstring''' UpperCAmelCase_ = data UpperCAmelCase_ = None UpperCAmelCase_ = None def a__ ( lowerCAmelCase__ ): # In Order traversal of the tree if tree: display(tree.left ) print(tree.data ) display(tree.right ) def a__ ( lowerCAmelCase__ ): return 1 + max(depth_of_tree(tree.left ) , depth_of_tree(tree.right ) ) if tree else 0 def a__ ( lowerCAmelCase__ ): if not tree: return True if tree.left and tree.right: return is_full_binary_tree(tree.left ) and is_full_binary_tree(tree.right ) else: return not tree.left and not tree.right def a__ ( ): # Main function for testing. UpperCAmelCase_ = Node(1 ) UpperCAmelCase_ = Node(2 ) UpperCAmelCase_ = Node(3 ) UpperCAmelCase_ = Node(4 ) UpperCAmelCase_ = Node(5 ) UpperCAmelCase_ = Node(6 ) UpperCAmelCase_ = Node(7 ) UpperCAmelCase_ = Node(8 ) UpperCAmelCase_ = Node(9 ) print(is_full_binary_tree(lowerCAmelCase__ ) ) print(depth_of_tree(lowerCAmelCase__ ) ) print("Tree is: " ) display(lowerCAmelCase__ ) if __name__ == "__main__": main()
82
"""simple docstring""" import argparse from torch import nn # transformers_old should correspond to branch `save_old_prophetnet_model_structure` here # original prophetnet_checkpoints are saved under `patrickvonplaten/..._old` respectively from transformers_old.modeling_prophetnet import ( ProphetNetForConditionalGeneration as ProphetNetForConditionalGenerationOld, ) from transformers_old.modeling_xlm_prophetnet import ( XLMProphetNetForConditionalGeneration as XLMProphetNetForConditionalGenerationOld, ) from transformers import ProphetNetForConditionalGeneration, XLMProphetNetForConditionalGeneration, logging lowerCamelCase = logging.get_logger(__name__) logging.set_verbosity_info() def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ): if "xprophetnet" in prophetnet_checkpoint_path: UpperCAmelCase_ = XLMProphetNetForConditionalGenerationOld.from_pretrained(lowerCAmelCase__ ) UpperCAmelCase_ , UpperCAmelCase_ = XLMProphetNetForConditionalGeneration.from_pretrained( lowerCAmelCase__ , output_loading_info=lowerCAmelCase__ ) else: UpperCAmelCase_ = ProphetNetForConditionalGenerationOld.from_pretrained(lowerCAmelCase__ ) UpperCAmelCase_ , UpperCAmelCase_ = ProphetNetForConditionalGeneration.from_pretrained( lowerCAmelCase__ , output_loading_info=lowerCAmelCase__ ) UpperCAmelCase_ = ["key_proj", "value_proj", "query_proj"] UpperCAmelCase_ = { "self_attn": "ngram_self_attn", "cross_attn": "encoder_attn", "cross_attn_layer_norm": "encoder_attn_layer_norm", "feed_forward_layer_norm": "final_layer_norm", "feed_forward": "", "intermediate": "fc1", "output": "fc2", "key_proj": "k_proj", "query_proj": "q_proj", "value_proj": "v_proj", "word_embeddings": "embed_tokens", "embeddings_layer_norm": "emb_layer_norm", "relative_pos_embeddings": "relative_linear", "ngram_embeddings": "ngram_input_embed", "position_embeddings": "embed_positions", } for key in loading_info["missing_keys"]: UpperCAmelCase_ = key.split("." ) if attributes[0] == "lm_head": UpperCAmelCase_ = prophet UpperCAmelCase_ = prophet_old else: UpperCAmelCase_ = prophet.prophetnet UpperCAmelCase_ = prophet_old.model UpperCAmelCase_ = False for attribute in attributes: if attribute in mapping: UpperCAmelCase_ = mapping[attribute] if not hasattr(lowerCAmelCase__ , lowerCAmelCase__ ) and len(lowerCAmelCase__ ) > 0: UpperCAmelCase_ = attribute elif hasattr(lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = attribute if attribute == "weight": assert old_model.weight.shape == model.weight.shape, "Shapes have to match!" UpperCAmelCase_ = old_model.weight logger.info(f"""{attribute} is initialized.""" ) UpperCAmelCase_ = True break elif attribute == "bias": assert old_model.bias.shape == model.bias.shape, "Shapes have to match!" UpperCAmelCase_ = old_model.bias logger.info(f"""{attribute} is initialized""" ) UpperCAmelCase_ = True break elif attribute in special_keys and hasattr(lowerCAmelCase__ , "in_proj_weight" ): UpperCAmelCase_ = old_model.in_proj_weight.shape[0] // 3 UpperCAmelCase_ = getattr(lowerCAmelCase__ , lowerCAmelCase__ ) param.weight.shape == old_model.in_proj_weight[:embed_dim, :].shape, "Shapes have to match" param.bias.shape == old_model.in_proj_bias[:embed_dim].shape, "Shapes have to match" if attribute == "query_proj": UpperCAmelCase_ = nn.Parameter(old_model.in_proj_weight[:embed_dim, :] ) UpperCAmelCase_ = nn.Parameter(old_model.in_proj_bias[:embed_dim] ) elif attribute == "key_proj": UpperCAmelCase_ = nn.Parameter(old_model.in_proj_weight[embed_dim : 2 * embed_dim, :] ) UpperCAmelCase_ = nn.Parameter(old_model.in_proj_bias[embed_dim : 2 * embed_dim] ) elif attribute == "value_proj": UpperCAmelCase_ = nn.Parameter(old_model.in_proj_weight[2 * embed_dim :, :] ) UpperCAmelCase_ = nn.Parameter(old_model.in_proj_bias[2 * embed_dim :] ) UpperCAmelCase_ = True break elif attribute == "position_embeddings": assert ( model.position_embeddings.weight.shape[-1] == old_model.embed_positions.weight.shape[-1] ), "Hidden size has to match" assert model.position_embeddings.weight.shape[0] == 512, "We want 512 position_embeddings." UpperCAmelCase_ = nn.Parameter(old_model.embed_positions.weight[:512, :] ) UpperCAmelCase_ = True break if attribute.isdigit(): UpperCAmelCase_ = model[int(lowerCAmelCase__ )] UpperCAmelCase_ = old_model[int(lowerCAmelCase__ )] else: UpperCAmelCase_ = getattr(lowerCAmelCase__ , lowerCAmelCase__ ) if old_attribute == "": UpperCAmelCase_ = old_model else: if not hasattr(lowerCAmelCase__ , lowerCAmelCase__ ): raise ValueError(f"""{old_model} does not have {old_attribute}""" ) UpperCAmelCase_ = getattr(lowerCAmelCase__ , lowerCAmelCase__ ) if not is_key_init: raise ValueError(f"""{key} was not correctly initialized!""" ) print(f"""Saving model to {pytorch_dump_folder_path}""" ) prophet.save_pretrained(lowerCAmelCase__ ) if __name__ == "__main__": lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--prophetnet_checkpoint_path""", default=None, type=str, required=True, help="""Path the official PyTorch dump.""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) lowerCamelCase = parser.parse_args() convert_prophetnet_checkpoint_to_pytorch(args.prophetnet_checkpoint_path, args.pytorch_dump_folder_path)
82
1
"""simple docstring""" import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": lowerCamelCase = pd.read_csv("""sample_data.csv""", header=None) lowerCamelCase = df.shape[:1][0] # If you're using some other dataset input the target column lowerCamelCase = df.iloc[:, 1:2] lowerCamelCase = actual_data.values.reshape(len_data, 1) lowerCamelCase = MinMaxScaler().fit_transform(actual_data) lowerCamelCase = 10 lowerCamelCase = 5 lowerCamelCase = 20 lowerCamelCase = len_data - periods * look_back lowerCamelCase = actual_data[:division] lowerCamelCase = actual_data[division - look_back :] lowerCamelCase , lowerCamelCase = [], [] lowerCamelCase , lowerCamelCase = [], [] for i in range(0, len(train_data) - forward_days - look_back + 1): train_x.append(train_data[i : i + look_back]) train_y.append(train_data[i + look_back : i + look_back + forward_days]) for i in range(0, len(test_data) - forward_days - look_back + 1): test_x.append(test_data[i : i + look_back]) test_y.append(test_data[i + look_back : i + look_back + forward_days]) lowerCamelCase = np.array(train_x) lowerCamelCase = np.array(test_x) lowerCamelCase = np.array([list(i.ravel()) for i in train_y]) lowerCamelCase = np.array([list(i.ravel()) for i in test_y]) lowerCamelCase = Sequential() model.add(LSTM(128, input_shape=(look_back, 1), return_sequences=True)) model.add(LSTM(64, input_shape=(128, 1))) model.add(Dense(forward_days)) model.compile(loss="""mean_squared_error""", optimizer="""adam""") lowerCamelCase = model.fit( x_train, y_train, epochs=150, verbose=1, shuffle=True, batch_size=4 ) lowerCamelCase = model.predict(x_test)
82
"""simple docstring""" def a__ ( lowerCAmelCase__ ): UpperCAmelCase_ = 0 UpperCAmelCase_ = len(lowerCAmelCase__ ) for i in range(n - 1 ): for j in range(i + 1 , lowerCAmelCase__ ): if arr[i] > arr[j]: num_inversions += 1 return num_inversions def a__ ( lowerCAmelCase__ ): if len(lowerCAmelCase__ ) <= 1: return arr, 0 UpperCAmelCase_ = len(lowerCAmelCase__ ) // 2 UpperCAmelCase_ = arr[0:mid] UpperCAmelCase_ = arr[mid:] UpperCAmelCase_ , UpperCAmelCase_ = count_inversions_recursive(lowerCAmelCase__ ) UpperCAmelCase_ , UpperCAmelCase_ = count_inversions_recursive(lowerCAmelCase__ ) UpperCAmelCase_ , UpperCAmelCase_ = _count_cross_inversions(lowerCAmelCase__ , lowerCAmelCase__ ) UpperCAmelCase_ = inversion_p + inversions_q + cross_inversions return c, num_inversions def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = [] UpperCAmelCase_ = UpperCAmelCase_ = UpperCAmelCase_ = 0 while i < len(lowerCAmelCase__ ) and j < len(lowerCAmelCase__ ): if p[i] > q[j]: # if P[1] > Q[j], then P[k] > Q[k] for all i < k <= len(P) # These are all inversions. The claim emerges from the # property that P is sorted. num_inversion += len(lowerCAmelCase__ ) - i r.append(q[j] ) j += 1 else: r.append(p[i] ) i += 1 if i < len(lowerCAmelCase__ ): r.extend(p[i:] ) else: r.extend(q[j:] ) return r, num_inversion def a__ ( ): UpperCAmelCase_ = [10, 2, 1, 5, 5, 2, 11] # this arr has 8 inversions: # (10, 2), (10, 1), (10, 5), (10, 5), (10, 2), (2, 1), (5, 2), (5, 2) UpperCAmelCase_ = count_inversions_bf(lowerCAmelCase__ ) UpperCAmelCase_ , UpperCAmelCase_ = count_inversions_recursive(lowerCAmelCase__ ) assert num_inversions_bf == num_inversions_recursive == 8 print("number of inversions = " , lowerCAmelCase__ ) # testing an array with zero inversion (a sorted arr_1) arr_a.sort() UpperCAmelCase_ = count_inversions_bf(lowerCAmelCase__ ) UpperCAmelCase_ , UpperCAmelCase_ = count_inversions_recursive(lowerCAmelCase__ ) assert num_inversions_bf == num_inversions_recursive == 0 print("number of inversions = " , lowerCAmelCase__ ) # an empty list should also have zero inversions UpperCAmelCase_ = [] UpperCAmelCase_ = count_inversions_bf(lowerCAmelCase__ ) UpperCAmelCase_ , UpperCAmelCase_ = count_inversions_recursive(lowerCAmelCase__ ) assert num_inversions_bf == num_inversions_recursive == 0 print("number of inversions = " , lowerCAmelCase__ ) if __name__ == "__main__": main()
82
1
"""simple docstring""" import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def a__ ( lowerCAmelCase__ ): UpperCAmelCase_ = args.pruning_method UpperCAmelCase_ = args.threshold UpperCAmelCase_ = args.model_name_or_path.rstrip("/" ) UpperCAmelCase_ = args.target_model_path print(f"""Load fine-pruned model from {model_name_or_path}""" ) UpperCAmelCase_ = torch.load(os.path.join(lowerCAmelCase__ , "pytorch_model.bin" ) ) UpperCAmelCase_ = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: UpperCAmelCase_ = tensor print(f"""Copied layer {name}""" ) elif "classifier" in name or "qa_output" in name: UpperCAmelCase_ = tensor print(f"""Copied layer {name}""" ) elif "bias" in name: UpperCAmelCase_ = tensor print(f"""Copied layer {name}""" ) else: if pruning_method == "magnitude": UpperCAmelCase_ = MagnitudeBinarizer.apply(inputs=lowerCAmelCase__ , threshold=lowerCAmelCase__ ) UpperCAmelCase_ = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "topK": if "mask_scores" in name: continue UpperCAmelCase_ = name[:-6] UpperCAmelCase_ = model[f"""{prefix_}mask_scores"""] UpperCAmelCase_ = TopKBinarizer.apply(lowerCAmelCase__ , lowerCAmelCase__ ) UpperCAmelCase_ = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue UpperCAmelCase_ = name[:-6] UpperCAmelCase_ = model[f"""{prefix_}mask_scores"""] UpperCAmelCase_ = ThresholdBinarizer.apply(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) UpperCAmelCase_ = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "l0": if "mask_scores" in name: continue UpperCAmelCase_ = name[:-6] UpperCAmelCase_ = model[f"""{prefix_}mask_scores"""] UpperCAmelCase_ , UpperCAmelCase_ = -0.1, 1.1 UpperCAmelCase_ = torch.sigmoid(lowerCAmelCase__ ) UpperCAmelCase_ = s * (r - l) + l UpperCAmelCase_ = s_bar.clamp(min=0.0 , max=1.0 ) UpperCAmelCase_ = tensor * mask print(f"""Pruned layer {name}""" ) else: raise ValueError("Unknown pruning method" ) if target_model_path is None: UpperCAmelCase_ = os.path.join( os.path.dirname(lowerCAmelCase__ ) , f"""bertarized_{os.path.basename(lowerCAmelCase__ )}""" ) if not os.path.isdir(lowerCAmelCase__ ): shutil.copytree(lowerCAmelCase__ , lowerCAmelCase__ ) print(f"""\nCreated folder {target_model_path}""" ) torch.save(lowerCAmelCase__ , os.path.join(lowerCAmelCase__ , "pytorch_model.bin" ) ) print("\nPruned model saved! See you later!" ) if __name__ == "__main__": lowerCamelCase = argparse.ArgumentParser() parser.add_argument( """--pruning_method""", choices=["""l0""", """magnitude""", """topK""", """sigmoied_threshold"""], type=str, required=True, help=( """Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning,""" """ sigmoied_threshold = Soft movement pruning)""" ), ) parser.add_argument( """--threshold""", type=float, required=False, help=( """For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model.""" """For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared.""" """Not needed for `l0`""" ), ) parser.add_argument( """--model_name_or_path""", type=str, required=True, help="""Folder containing the model that was previously fine-pruned""", ) parser.add_argument( """--target_model_path""", default=None, type=str, required=False, help="""Folder containing the model that was previously fine-pruned""", ) lowerCamelCase = parser.parse_args() main(args)
82
"""simple docstring""" def a__ ( lowerCAmelCase__ ): if number > 0: raise ValueError("input must be a negative integer" ) UpperCAmelCase_ = len(bin(lowerCAmelCase__ )[3:] ) UpperCAmelCase_ = bin(abs(lowerCAmelCase__ ) - (1 << binary_number_length) )[3:] UpperCAmelCase_ = ( ( "1" + "0" * (binary_number_length - len(lowerCAmelCase__ )) + twos_complement_number ) if number < 0 else "0" ) return "0b" + twos_complement_number if __name__ == "__main__": import doctest doctest.testmod()
82
1
"""simple docstring""" import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import is_accelerate_available, is_torch_available, is_transformers_available, is_xformers_available from . import BaseDiffusersCLICommand def a__ ( lowerCAmelCase__ ): return EnvironmentCommand() class lowercase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' @staticmethod def lowercase__ ( _UpperCAmelCase : ArgumentParser ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase_ = parser.add_parser("env" ) download_parser.set_defaults(func=_UpperCAmelCase ) def lowercase__ ( self : List[str] ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase_ = huggingface_hub.__version__ UpperCAmelCase_ = "not installed" UpperCAmelCase_ = "NA" if is_torch_available(): import torch UpperCAmelCase_ = torch.__version__ UpperCAmelCase_ = torch.cuda.is_available() UpperCAmelCase_ = "not installed" if is_transformers_available(): import transformers UpperCAmelCase_ = transformers.__version__ UpperCAmelCase_ = "not installed" if is_accelerate_available(): import accelerate UpperCAmelCase_ = accelerate.__version__ UpperCAmelCase_ = "not installed" if is_xformers_available(): import xformers UpperCAmelCase_ = xformers.__version__ UpperCAmelCase_ = { "`diffusers` version": version, "Platform": platform.platform(), "Python version": platform.python_version(), "PyTorch version (GPU?)": F"""{pt_version} ({pt_cuda_available})""", "Huggingface_hub version": hub_version, "Transformers version": transformers_version, "Accelerate version": accelerate_version, "xFormers version": xformers_version, "Using GPU in script?": "<fill in>", "Using distributed or parallel set-up in script?": "<fill in>", } print("\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n" ) print(self.format_dict(_UpperCAmelCase ) ) return info @staticmethod def lowercase__ ( _UpperCAmelCase : List[str] ) -> str: '''simple docstring''' return "\n".join([F"""- {prop}: {val}""" for prop, val in d.items()] ) + "\n"
82
"""simple docstring""" from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING lowerCamelCase = logging.get_logger(__name__) @add_end_docstrings(SCREAMING_SNAKE_CASE ) class lowercase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self : Union[str, Any] , **_UpperCAmelCase : Optional[int] ) -> str: '''simple docstring''' super().__init__(**_UpperCAmelCase ) if self.framework == "tf": raise ValueError(F"""The {self.__class__} is only available in PyTorch.""" ) requires_backends(self , "vision" ) self.check_model_type(_UpperCAmelCase ) def __call__( self : int , _UpperCAmelCase : Union[str, "Image.Image", List[Dict[str, Any]]] , _UpperCAmelCase : Union[str, List[str]] = None , **_UpperCAmelCase : Optional[int] , ) -> List[Any]: '''simple docstring''' if "text_queries" in kwargs: UpperCAmelCase_ = kwargs.pop("text_queries" ) if isinstance(_UpperCAmelCase , (str, Image.Image) ): UpperCAmelCase_ = {"image": image, "candidate_labels": candidate_labels} else: UpperCAmelCase_ = image UpperCAmelCase_ = super().__call__(_UpperCAmelCase , **_UpperCAmelCase ) return results def lowercase__ ( self : str , **_UpperCAmelCase : List[Any] ) -> List[str]: '''simple docstring''' UpperCAmelCase_ = {} if "threshold" in kwargs: UpperCAmelCase_ = kwargs["threshold"] if "top_k" in kwargs: UpperCAmelCase_ = kwargs["top_k"] return {}, {}, postprocess_params def lowercase__ ( self : int , _UpperCAmelCase : int ) -> Any: '''simple docstring''' UpperCAmelCase_ = load_image(inputs["image"] ) UpperCAmelCase_ = inputs["candidate_labels"] if isinstance(_UpperCAmelCase , _UpperCAmelCase ): UpperCAmelCase_ = candidate_labels.split("," ) UpperCAmelCase_ = torch.tensor([[image.height, image.width]] , dtype=torch.intaa ) for i, candidate_label in enumerate(_UpperCAmelCase ): UpperCAmelCase_ = self.tokenizer(_UpperCAmelCase , return_tensors=self.framework ) UpperCAmelCase_ = self.image_processor(_UpperCAmelCase , return_tensors=self.framework ) yield { "is_last": i == len(_UpperCAmelCase ) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def lowercase__ ( self : int , _UpperCAmelCase : List[Any] ) -> int: '''simple docstring''' UpperCAmelCase_ = model_inputs.pop("target_size" ) UpperCAmelCase_ = model_inputs.pop("candidate_label" ) UpperCAmelCase_ = model_inputs.pop("is_last" ) UpperCAmelCase_ = self.model(**_UpperCAmelCase ) UpperCAmelCase_ = {"target_size": target_size, "candidate_label": candidate_label, "is_last": is_last, **outputs} return model_outputs def lowercase__ ( self : str , _UpperCAmelCase : List[str] , _UpperCAmelCase : Any=0.1 , _UpperCAmelCase : List[str]=None ) -> int: '''simple docstring''' UpperCAmelCase_ = [] for model_output in model_outputs: UpperCAmelCase_ = model_output["candidate_label"] UpperCAmelCase_ = BaseModelOutput(_UpperCAmelCase ) UpperCAmelCase_ = self.image_processor.post_process_object_detection( outputs=_UpperCAmelCase , threshold=_UpperCAmelCase , target_sizes=model_output["target_size"] )[0] for index in outputs["scores"].nonzero(): UpperCAmelCase_ = outputs["scores"][index].item() UpperCAmelCase_ = self._get_bounding_box(outputs["boxes"][index][0] ) UpperCAmelCase_ = {"score": score, "label": label, "box": box} results.append(_UpperCAmelCase ) UpperCAmelCase_ = sorted(_UpperCAmelCase , key=lambda _UpperCAmelCase : x["score"] , reverse=_UpperCAmelCase ) if top_k: UpperCAmelCase_ = results[:top_k] return results def lowercase__ ( self : str , _UpperCAmelCase : "torch.Tensor" ) -> Dict[str, int]: '''simple docstring''' if self.framework != "pt": raise ValueError("The ZeroShotObjectDetectionPipeline is only available in PyTorch." ) UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = box.int().tolist() UpperCAmelCase_ = { "xmin": xmin, "ymin": ymin, "xmax": xmax, "ymax": ymax, } return bbox
82
1