code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
"""simple docstring""" import numpy as np from sklearn.datasets import fetch_california_housing from sklearn.metrics import mean_absolute_error, mean_squared_error from sklearn.model_selection import train_test_split from xgboost import XGBRegressor def lowerCamelCase__ ( __snake_case ) -> tuple: """simple docstring""" return (data["data"], data["target"]) def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> np.ndarray: """simple docstring""" _UpperCamelCase = XGBRegressor(verbosity=0, random_state=42 ) xgb.fit(__snake_case, __snake_case ) # Predict target for test data _UpperCamelCase = xgb.predict(__snake_case ) _UpperCamelCase = predictions.reshape(len(__snake_case ), 1 ) return predictions def lowerCamelCase__ ( ) -> None: """simple docstring""" _UpperCamelCase = fetch_california_housing() _UpperCamelCase , _UpperCamelCase = data_handling(__snake_case ) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = train_test_split( __snake_case, __snake_case, test_size=0.25, random_state=1 ) _UpperCamelCase = xgboost(__snake_case, __snake_case, __snake_case ) # Error printing print(F'''Mean Absolute Error : {mean_absolute_error(__snake_case, __snake_case )}''' ) print(F'''Mean Square Error : {mean_squared_error(__snake_case, __snake_case )}''' ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) main()
78
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class _UpperCAmelCase( unittest.TestCase ): def __init__( self , __a , __a=7 , __a=3 , __a=18 , __a=30 , __a=4_00 , __a=True , __a=None , __a=True , __a=None , __a=True , ) -> int: '''simple docstring''' _UpperCamelCase = size if size is not None else {'''shortest_edge''': 20} _UpperCamelCase = crop_size if crop_size is not None else {'''height''': 18, '''width''': 18} _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = num_channels _UpperCamelCase = image_size _UpperCamelCase = min_resolution _UpperCamelCase = max_resolution _UpperCamelCase = do_resize _UpperCamelCase = size _UpperCamelCase = do_center_crop _UpperCamelCase = crop_size _UpperCamelCase = do_flip_channel_order def UpperCAmelCase ( self) -> str: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_flip_channel_order": self.do_flip_channel_order, } @require_torch @require_vision class _UpperCAmelCase( lowerCamelCase , unittest.TestCase ): lowercase__ = MobileViTImageProcessor if is_vision_available() else None def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = MobileViTImageProcessingTester(self) @property def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(__a , '''do_resize''')) self.assertTrue(hasattr(__a , '''size''')) self.assertTrue(hasattr(__a , '''do_center_crop''')) self.assertTrue(hasattr(__a , '''center_crop''')) self.assertTrue(hasattr(__a , '''do_flip_channel_order''')) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size , {'''shortest_edge''': 20}) self.assertEqual(image_processor.crop_size , {'''height''': 18, '''width''': 18}) _UpperCamelCase = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84) self.assertEqual(image_processor.size , {'''shortest_edge''': 42}) self.assertEqual(image_processor.crop_size , {'''height''': 84, '''width''': 84}) def UpperCAmelCase ( self) -> Dict: '''simple docstring''' pass def UpperCAmelCase ( self) -> str: '''simple docstring''' # Initialize image_processing _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random PIL images _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__a) for image in image_inputs: self.assertIsInstance(__a , Image.Image) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(__a , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' # Initialize image_processing _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__a , numpify=__a) for image in image_inputs: self.assertIsInstance(__a , np.ndarray) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(__a , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def UpperCAmelCase ( self) -> int: '''simple docstring''' # Initialize image_processing _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__a , torchify=__a) for image in image_inputs: self.assertIsInstance(__a , torch.Tensor) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(__a , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
78
1
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class _UpperCAmelCase( unittest.TestCase ): def __init__( self , __a , __a=7 , __a=3 , __a=18 , __a=30 , __a=4_00 , __a=True , __a=None , __a=True , __a=None , __a=True , ) -> int: '''simple docstring''' _UpperCamelCase = size if size is not None else {'''shortest_edge''': 20} _UpperCamelCase = crop_size if crop_size is not None else {'''height''': 18, '''width''': 18} _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = num_channels _UpperCamelCase = image_size _UpperCamelCase = min_resolution _UpperCamelCase = max_resolution _UpperCamelCase = do_resize _UpperCamelCase = size _UpperCamelCase = do_center_crop _UpperCamelCase = crop_size _UpperCamelCase = do_flip_channel_order def UpperCAmelCase ( self) -> str: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_flip_channel_order": self.do_flip_channel_order, } @require_torch @require_vision class _UpperCAmelCase( lowerCamelCase , unittest.TestCase ): lowercase__ = MobileViTImageProcessor if is_vision_available() else None def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = MobileViTImageProcessingTester(self) @property def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(__a , '''do_resize''')) self.assertTrue(hasattr(__a , '''size''')) self.assertTrue(hasattr(__a , '''do_center_crop''')) self.assertTrue(hasattr(__a , '''center_crop''')) self.assertTrue(hasattr(__a , '''do_flip_channel_order''')) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size , {'''shortest_edge''': 20}) self.assertEqual(image_processor.crop_size , {'''height''': 18, '''width''': 18}) _UpperCamelCase = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84) self.assertEqual(image_processor.size , {'''shortest_edge''': 42}) self.assertEqual(image_processor.crop_size , {'''height''': 84, '''width''': 84}) def UpperCAmelCase ( self) -> Dict: '''simple docstring''' pass def UpperCAmelCase ( self) -> str: '''simple docstring''' # Initialize image_processing _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random PIL images _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__a) for image in image_inputs: self.assertIsInstance(__a , Image.Image) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(__a , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' # Initialize image_processing _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__a , numpify=__a) for image in image_inputs: self.assertIsInstance(__a , np.ndarray) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(__a , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def UpperCAmelCase ( self) -> int: '''simple docstring''' # Initialize image_processing _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__a , torchify=__a) for image in image_inputs: self.assertIsInstance(__a , torch.Tensor) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(__a , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
78
"""simple docstring""" import warnings from typing import List import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import is_flax_available, is_tf_available, is_torch_available class _UpperCAmelCase( lowerCamelCase ): lowercase__ = ['image_processor', 'tokenizer'] lowercase__ = 'OwlViTImageProcessor' lowercase__ = ('CLIPTokenizer', 'CLIPTokenizerFast') def __init__( self , __a=None , __a=None , **__a) -> List[Any]: '''simple docstring''' _UpperCamelCase = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , __a , ) _UpperCamelCase = kwargs.pop('''feature_extractor''') _UpperCamelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''') if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''') super().__init__(__a , __a) def __call__( self , __a=None , __a=None , __a=None , __a="max_length" , __a="np" , **__a) -> List[str]: '''simple docstring''' if text is None and query_images is None and images is None: raise ValueError( '''You have to specify at least one text or query image or image. All three cannot be none.''') if text is not None: if isinstance(__a , __a) or (isinstance(__a , __a) and not isinstance(text[0] , __a)): _UpperCamelCase = [self.tokenizer(__a , padding=__a , return_tensors=__a , **__a)] elif isinstance(__a , __a) and isinstance(text[0] , __a): _UpperCamelCase = [] # Maximum number of queries across batch _UpperCamelCase = max([len(__a) for t in text]) # Pad all batch samples to max number of text queries for t in text: if len(__a) != max_num_queries: _UpperCamelCase = t + [''' '''] * (max_num_queries - len(__a)) _UpperCamelCase = self.tokenizer(__a , padding=__a , return_tensors=__a , **__a) encodings.append(__a) else: raise TypeError('''Input text should be a string, a list of strings or a nested list of strings''') if return_tensors == "np": _UpperCamelCase = np.concatenate([encoding['''input_ids'''] for encoding in encodings] , axis=0) _UpperCamelCase = np.concatenate([encoding['''attention_mask'''] for encoding in encodings] , axis=0) elif return_tensors == "jax" and is_flax_available(): import jax.numpy as jnp _UpperCamelCase = jnp.concatenate([encoding['''input_ids'''] for encoding in encodings] , axis=0) _UpperCamelCase = jnp.concatenate([encoding['''attention_mask'''] for encoding in encodings] , axis=0) elif return_tensors == "pt" and is_torch_available(): import torch _UpperCamelCase = torch.cat([encoding['''input_ids'''] for encoding in encodings] , dim=0) _UpperCamelCase = torch.cat([encoding['''attention_mask'''] for encoding in encodings] , dim=0) elif return_tensors == "tf" and is_tf_available(): import tensorflow as tf _UpperCamelCase = tf.stack([encoding['''input_ids'''] for encoding in encodings] , axis=0) _UpperCamelCase = tf.stack([encoding['''attention_mask'''] for encoding in encodings] , axis=0) else: raise ValueError('''Target return tensor type could not be returned''') _UpperCamelCase = BatchEncoding() _UpperCamelCase = input_ids _UpperCamelCase = attention_mask if query_images is not None: _UpperCamelCase = BatchEncoding() _UpperCamelCase = self.image_processor( __a , return_tensors=__a , **__a).pixel_values _UpperCamelCase = query_pixel_values if images is not None: _UpperCamelCase = self.image_processor(__a , return_tensors=__a , **__a) if text is not None and images is not None: _UpperCamelCase = image_features.pixel_values return encoding elif query_images is not None and images is not None: _UpperCamelCase = image_features.pixel_values return encoding elif text is not None or query_images is not None: return encoding else: return BatchEncoding(data=dict(**__a) , tensor_type=__a) def UpperCAmelCase ( self , *__a , **__a) -> str: '''simple docstring''' return self.image_processor.post_process(*__a , **__a) def UpperCAmelCase ( self , *__a , **__a) -> Dict: '''simple docstring''' return self.image_processor.post_process_object_detection(*__a , **__a) def UpperCAmelCase ( self , *__a , **__a) -> Optional[Any]: '''simple docstring''' return self.image_processor.post_process_image_guided_detection(*__a , **__a) def UpperCAmelCase ( self , *__a , **__a) -> Optional[int]: '''simple docstring''' return self.tokenizer.batch_decode(*__a , **__a) def UpperCAmelCase ( self , *__a , **__a) -> Optional[Any]: '''simple docstring''' return self.tokenizer.decode(*__a , **__a) @property def UpperCAmelCase ( self) -> str: '''simple docstring''' warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , __a , ) return self.image_processor_class @property def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , __a , ) return self.image_processor
78
1
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mvp import MvpTokenizer _a = logging.get_logger(__name__) _a = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} # See all MVP models at https://huggingface.co/models?filter=mvp _a = { """vocab_file""": { """RUCAIBox/mvp""": """https://huggingface.co/RUCAIBox/mvp/resolve/main/vocab.json""", }, """added_tokens.json""": { """RUCAIBox/mvp""": """https://huggingface.co/RUCAIBox/mvp/resolve/main/added_tokens.json""", }, """merges_file""": { """RUCAIBox/mvp""": """https://huggingface.co/RUCAIBox/mvp/resolve/main/merges.txt""", }, """tokenizer_file""": { """RUCAIBox/mvp""": """https://huggingface.co/RUCAIBox/mvp/resolve/main/tokenizer.json""", }, } _a = { """RUCAIBox/mvp""": 1024, } class _UpperCAmelCase( lowerCamelCase ): lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = ['input_ids', 'attention_mask'] lowercase__ = MvpTokenizer def __init__( self , __a=None , __a=None , __a=None , __a="replace" , __a="<s>" , __a="</s>" , __a="</s>" , __a="<s>" , __a="<unk>" , __a="<pad>" , __a="<mask>" , __a=False , __a=True , **__a , ) -> List[str]: '''simple docstring''' super().__init__( __a , __a , tokenizer_file=__a , errors=__a , bos_token=__a , eos_token=__a , sep_token=__a , cls_token=__a , unk_token=__a , pad_token=__a , mask_token=__a , add_prefix_space=__a , trim_offsets=__a , **__a , ) _UpperCamelCase = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__()) if pre_tok_state.get('''add_prefix_space''' , __a) != add_prefix_space: _UpperCamelCase = getattr(__a , pre_tok_state.pop('''type''')) _UpperCamelCase = add_prefix_space _UpperCamelCase = pre_tok_class(**__a) _UpperCamelCase = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` _UpperCamelCase = '''post_processor''' _UpperCamelCase = getattr(self.backend_tokenizer , __a , __a) if tokenizer_component_instance: _UpperCamelCase = json.loads(tokenizer_component_instance.__getstate__()) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: _UpperCamelCase = tuple(state['''sep''']) if "cls" in state: _UpperCamelCase = tuple(state['''cls''']) _UpperCamelCase = False if state.get('''add_prefix_space''' , __a) != add_prefix_space: _UpperCamelCase = add_prefix_space _UpperCamelCase = True if state.get('''trim_offsets''' , __a) != trim_offsets: _UpperCamelCase = trim_offsets _UpperCamelCase = True if changes_to_apply: _UpperCamelCase = getattr(__a , state.pop('''type''')) _UpperCamelCase = component_class(**__a) setattr(self.backend_tokenizer , __a , __a) @property def UpperCAmelCase ( self) -> str: '''simple docstring''' if self._mask_token is None: if self.verbose: logger.error('''Using mask_token, but it is not set yet.''') return None return str(self._mask_token) @mask_token.setter def UpperCAmelCase ( self , __a) -> Any: '''simple docstring''' _UpperCamelCase = AddedToken(__a , lstrip=__a , rstrip=__a) if isinstance(__a , __a) else value _UpperCamelCase = value def UpperCAmelCase ( self , *__a , **__a) -> BatchEncoding: '''simple docstring''' _UpperCamelCase = kwargs.get('''is_split_into_words''' , __a) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' '''to use it with pretokenized inputs.''') return super()._batch_encode_plus(*__a , **__a) def UpperCAmelCase ( self , *__a , **__a) -> BatchEncoding: '''simple docstring''' _UpperCamelCase = kwargs.get('''is_split_into_words''' , __a) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' '''to use it with pretokenized inputs.''') return super()._encode_plus(*__a , **__a) def UpperCAmelCase ( self , __a , __a = None) -> Tuple[str]: '''simple docstring''' _UpperCamelCase = self._tokenizer.model.save(__a , name=__a) return tuple(__a) def UpperCAmelCase ( self , __a , __a=None) -> Optional[int]: '''simple docstring''' _UpperCamelCase = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def UpperCAmelCase ( self , __a , __a = None) -> List[int]: '''simple docstring''' _UpperCamelCase = [self.sep_token_id] _UpperCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0]
78
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) _a = { """configuration_perceiver""": ["""PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """PerceiverConfig""", """PerceiverOnnxConfig"""], """tokenization_perceiver""": ["""PerceiverTokenizer"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ["""PerceiverFeatureExtractor"""] _a = ["""PerceiverImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ """PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST""", """PerceiverForImageClassificationConvProcessing""", """PerceiverForImageClassificationFourier""", """PerceiverForImageClassificationLearned""", """PerceiverForMaskedLM""", """PerceiverForMultimodalAutoencoding""", """PerceiverForOpticalFlow""", """PerceiverForSequenceClassification""", """PerceiverLayer""", """PerceiverModel""", """PerceiverPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys _a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
78
1
"""simple docstring""" from random import shuffle import tensorflow as tf from numpy import array def lowerCamelCase__ ( __snake_case, __snake_case ) -> Dict: """simple docstring""" _UpperCamelCase = int(__snake_case ) assert noofclusters < len(__snake_case ) # Find out the dimensionality _UpperCamelCase = len(vectors[0] ) # Will help select random centroids from among the available vectors _UpperCamelCase = list(range(len(__snake_case ) ) ) shuffle(__snake_case ) # GRAPH OF COMPUTATION # We initialize a new graph and set it as the default during each run # of this algorithm. This ensures that as this function is called # multiple times, the default graph doesn't keep getting crowded with # unused ops and Variables from previous function calls. _UpperCamelCase = tf.Graph() with graph.as_default(): # SESSION OF COMPUTATION _UpperCamelCase = tf.Session() ##CONSTRUCTING THE ELEMENTS OF COMPUTATION ##First lets ensure we have a Variable vector for each centroid, ##initialized to one of the vectors from the available data points _UpperCamelCase = [ tf.Variable(vectors[vector_indices[i]] ) for i in range(__snake_case ) ] ##These nodes will assign the centroid Variables the appropriate ##values _UpperCamelCase = tf.placeholder('''float64''', [dim] ) _UpperCamelCase = [] for centroid in centroids: cent_assigns.append(tf.assign(__snake_case, __snake_case ) ) ##Variables for cluster assignments of individual vectors(initialized ##to 0 at first) _UpperCamelCase = [tf.Variable(0 ) for i in range(len(__snake_case ) )] ##These nodes will assign an assignment Variable the appropriate ##value _UpperCamelCase = tf.placeholder('''int32''' ) _UpperCamelCase = [] for assignment in assignments: cluster_assigns.append(tf.assign(__snake_case, __snake_case ) ) ##Now lets construct the node that will compute the mean # The placeholder for the input _UpperCamelCase = tf.placeholder('''float''', [None, dim] ) # The Node/op takes the input and computes a mean along the 0th # dimension, i.e. the list of input vectors _UpperCamelCase = tf.reduce_mean(__snake_case, 0 ) ##Node for computing Euclidean distances # Placeholders for input _UpperCamelCase = tf.placeholder('''float''', [dim] ) _UpperCamelCase = tf.placeholder('''float''', [dim] ) _UpperCamelCase = tf.sqrt(tf.reduce_sum(tf.pow(tf.sub(__snake_case, __snake_case ), 2 ) ) ) ##This node will figure out which cluster to assign a vector to, ##based on Euclidean distances of the vector from the centroids. # Placeholder for input _UpperCamelCase = tf.placeholder('''float''', [noofclusters] ) _UpperCamelCase = tf.argmin(__snake_case, 0 ) ##INITIALIZING STATE VARIABLES ##This will help initialization of all Variables defined with respect ##to the graph. The Variable-initializer should be defined after ##all the Variables have been constructed, so that each of them ##will be included in the initialization. _UpperCamelCase = tf.initialize_all_variables() # Initialize all variables sess.run(__snake_case ) ##CLUSTERING ITERATIONS # Now perform the Expectation-Maximization steps of K-Means clustering # iterations. To keep things simple, we will only do a set number of # iterations, instead of using a Stopping Criterion. _UpperCamelCase = 1_00 for _ in range(__snake_case ): ##EXPECTATION STEP ##Based on the centroid locations till last iteration, compute ##the _expected_ centroid assignments. # Iterate over each vector for vector_n in range(len(__snake_case ) ): _UpperCamelCase = vectors[vector_n] # Compute Euclidean distance between this vector and each # centroid. Remember that this list cannot be named #'centroid_distances', since that is the input to the # cluster assignment node. _UpperCamelCase = [ sess.run(__snake_case, feed_dict={va: vect, va: sess.run(__snake_case )} ) for centroid in centroids ] # Now use the cluster assignment node, with the distances # as the input _UpperCamelCase = sess.run( __snake_case, feed_dict={centroid_distances: distances} ) # Now assign the value to the appropriate state variable sess.run( cluster_assigns[vector_n], feed_dict={assignment_value: assignment} ) ##MAXIMIZATION STEP # Based on the expected state computed from the Expectation Step, # compute the locations of the centroids so as to maximize the # overall objective of minimizing within-cluster Sum-of-Squares for cluster_n in range(__snake_case ): # Collect all the vectors assigned to this cluster _UpperCamelCase = [ vectors[i] for i in range(len(__snake_case ) ) if sess.run(assignments[i] ) == cluster_n ] # Compute new centroid location _UpperCamelCase = sess.run( __snake_case, feed_dict={mean_input: array(__snake_case )} ) # Assign value to appropriate variable sess.run( cent_assigns[cluster_n], feed_dict={centroid_value: new_location} ) # Return centroids and assignments _UpperCamelCase = sess.run(__snake_case ) _UpperCamelCase = sess.run(__snake_case ) return centroids, assignments
78
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class _UpperCAmelCase: def __init__( self , __a , __a=13 , __a=2 , __a=24 , __a=16 , __a=True , __a=True , __a=32 , __a=5 , __a=4 , __a=37 , __a="gelu" , __a=0.1 , __a=0.1 , __a=10 , __a=0.02 , __a=None , __a=2 , __a=2 , ) -> List[str]: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = patch_size _UpperCamelCase = max_length _UpperCamelCase = num_mel_bins _UpperCamelCase = is_training _UpperCamelCase = use_labels _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = type_sequence_label_size _UpperCamelCase = initializer_range _UpperCamelCase = scope _UpperCamelCase = frequency_stride _UpperCamelCase = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) _UpperCamelCase = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 _UpperCamelCase = (self.max_length - self.patch_size) // self.time_stride + 1 _UpperCamelCase = frequency_out_dimension * time_out_dimension _UpperCamelCase = num_patches + 2 def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins]) _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size) _UpperCamelCase = self.get_config() return config, input_values, labels def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__a , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def UpperCAmelCase ( self , __a , __a , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = ASTModel(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.prepare_config_and_inputs() ( ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ) = config_and_inputs _UpperCamelCase = {'''input_values''': input_values} return config, inputs_dict @require_torch class _UpperCAmelCase( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase__ = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) lowercase__ = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) lowercase__ = False lowercase__ = False lowercase__ = False lowercase__ = False def UpperCAmelCase ( self , __a , __a , __a , __a , __a) -> Optional[Any]: '''simple docstring''' if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = ASTModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=__a , has_text_modality=__a , hidden_size=37) def UpperCAmelCase ( self) -> int: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='''AST does not use inputs_embeds''') def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' pass def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = model_class(__a) self.assertIsInstance(model.get_input_embeddings() , (nn.Module)) _UpperCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__a , nn.Linear)) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = model_class(__a) _UpperCamelCase = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCamelCase = [*signature.parameters.keys()] _UpperCamelCase = ['''input_values'''] self.assertListEqual(arg_names[:1] , __a) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a) @slow def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase = ASTModel.from_pretrained(__a) self.assertIsNotNone(__a) def lowerCamelCase__ ( ) -> List[str]: """simple docstring""" _UpperCamelCase = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''', filename='''sample_audio.flac''', repo_type='''dataset''' ) _UpperCamelCase , _UpperCamelCase = torchaudio.load(__snake_case ) return audio, sampling_rate @require_torch @require_torchaudio class _UpperCAmelCase( unittest.TestCase ): @cached_property def UpperCAmelCase ( self) -> Dict: '''simple docstring''' return ( ASTFeatureExtractor.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''') if is_torchaudio_available() else None ) @slow def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.default_feature_extractor _UpperCamelCase = ASTForAudioClassification.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''').to(__a) _UpperCamelCase = self.default_feature_extractor _UpperCamelCase , _UpperCamelCase = prepare_audio() _UpperCamelCase = audio.squeeze().numpy() _UpperCamelCase = feature_extractor(__a , sampling_rate=__a , return_tensors='''pt''').to(__a) # forward pass with torch.no_grad(): _UpperCamelCase = model(**__a) # verify the logits _UpperCamelCase = torch.Size((1, 5_27)) self.assertEqual(outputs.logits.shape , __a) _UpperCamelCase = torch.tensor([-0.8760, -7.0042, -8.6602]).to(__a) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __a , atol=1e-4))
78
1
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _a = logging.get_logger(__name__) _a = {"""vocab_file""": """sentencepiece.model"""} _a = { """vocab_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/sentencepiece.model""", }, } _a = { """google/rembert""": 256, } class _UpperCAmelCase( lowerCamelCase ): lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , __a , __a=False , __a=True , __a=True , __a="[CLS]" , __a="[SEP]" , __a="[UNK]" , __a="[SEP]" , __a="[PAD]" , __a="[CLS]" , __a="[MASK]" , **__a , ) -> Optional[int]: '''simple docstring''' super().__init__( do_lower_case=__a , remove_space=__a , keep_accents=__a , bos_token=__a , eos_token=__a , unk_token=__a , sep_token=__a , pad_token=__a , cls_token=__a , mask_token=__a , **__a , ) _UpperCamelCase = do_lower_case _UpperCamelCase = remove_space _UpperCamelCase = keep_accents _UpperCamelCase = vocab_file _UpperCamelCase = spm.SentencePieceProcessor() self.sp_model.Load(__a) @property def UpperCAmelCase ( self) -> Dict: '''simple docstring''' return len(self.sp_model) def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = {self.convert_ids_to_tokens(__a): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __getstate__( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = self.__dict__.copy() _UpperCamelCase = None return state def __setstate__( self , __a) -> str: '''simple docstring''' _UpperCamelCase = d _UpperCamelCase = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file) def UpperCAmelCase ( self , __a , __a=False) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = self.sp_model.EncodeAsPieces(__a) return pieces def UpperCAmelCase ( self , __a) -> Union[str, Any]: '''simple docstring''' return self.sp_model.PieceToId(__a) def UpperCAmelCase ( self , __a) -> Dict: '''simple docstring''' return self.sp_model.IdToPiece(__a) def UpperCAmelCase ( self , __a) -> Tuple: '''simple docstring''' _UpperCamelCase = self.sp_model.decode_pieces(__a) return out_string def UpperCAmelCase ( self , __a , __a = None) -> List[int]: '''simple docstring''' _UpperCamelCase = [self.sep_token_id] _UpperCamelCase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def UpperCAmelCase ( self , __a , __a = None , __a = False) -> List[int]: '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''') return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(__a)) + [1] + ([0] * len(__a)) + [1] return [1] + ([0] * len(__a)) + [1] def UpperCAmelCase ( self , __a , __a = None) -> List[int]: '''simple docstring''' _UpperCamelCase = [self.sep_token_id] _UpperCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def UpperCAmelCase ( self , __a , __a = None) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(__a): logger.error('''Vocabulary path ({}) should be a directory'''.format(__a)) return _UpperCamelCase = os.path.join( __a , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file''']) if os.path.abspath(self.vocab_file) != os.path.abspath(__a): copyfile(self.vocab_file , __a) return (out_vocab_file,)
78
"""simple docstring""" def lowerCamelCase__ ( ) -> list[list[int]]: """simple docstring""" return [list(range(10_00 - i, -10_00 - i, -1 ) ) for i in range(10_00 )] _a = generate_large_matrix() _a = ( [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]], [[3, 2], [1, 0]], [[7, 7, 6]], [[7, 7, 6], [-1, -2, -3]], grid, ) def lowerCamelCase__ ( __snake_case ) -> None: """simple docstring""" assert all(row == sorted(__snake_case, reverse=__snake_case ) for row in grid ) assert all(list(__snake_case ) == sorted(__snake_case, reverse=__snake_case ) for col in zip(*__snake_case ) ) def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" _UpperCamelCase = 0 _UpperCamelCase = len(__snake_case ) - 1 # Edge cases such as no values or all numbers are negative. if not array or array[0] < 0: return 0 while right + 1 > left: _UpperCamelCase = (left + right) // 2 _UpperCamelCase = array[mid] # Num must be negative and the index must be greater than or equal to 0. if num < 0 and array[mid - 1] >= 0: return mid if num >= 0: _UpperCamelCase = mid + 1 else: _UpperCamelCase = mid - 1 # No negative numbers so return the last index of the array + 1 which is the length. return len(__snake_case ) def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" _UpperCamelCase = 0 _UpperCamelCase = len(grid[0] ) for i in range(len(__snake_case ) ): _UpperCamelCase = find_negative_index(grid[i][:bound] ) total += bound return (len(__snake_case ) * len(grid[0] )) - total def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" return len([number for row in grid for number in row if number < 0] ) def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" _UpperCamelCase = 0 for row in grid: for i, number in enumerate(__snake_case ): if number < 0: total += len(__snake_case ) - i break return total def lowerCamelCase__ ( ) -> None: """simple docstring""" from timeit import timeit print('''Running benchmarks''' ) _UpperCamelCase = ( '''from __main__ import count_negatives_binary_search, ''' '''count_negatives_brute_force, count_negatives_brute_force_with_break, grid''' ) for func in ( "count_negatives_binary_search", # took 0.7727 seconds "count_negatives_brute_force_with_break", # took 4.6505 seconds "count_negatives_brute_force", # took 12.8160 seconds ): _UpperCamelCase = timeit(F'''{func}(grid=grid)''', setup=__snake_case, number=5_00 ) print(F'''{func}() took {time:0.4f} seconds''' ) if __name__ == "__main__": import doctest doctest.testmod() benchmark()
78
1
"""simple docstring""" from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging _a = logging.get_logger(__name__) # pylint: disable=invalid-name class _UpperCAmelCase( lowerCamelCase ): def __init__( self , __a , __a , __a , __a , __a , __a , __a , __a , __a , ) -> Any: '''simple docstring''' super().__init__() if hasattr(scheduler.config , '''steps_offset''') and scheduler.config.steps_offset != 1: _UpperCamelCase = ( F'''The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`''' F''' should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure ''' '''to update the config accordingly as leaving `steps_offset` might led to incorrect results''' ''' in future versions. If you have downloaded this checkpoint from the Hugging Face Hub,''' ''' it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`''' ''' file''' ) deprecate('''steps_offset!=1''' , '''1.0.0''' , __a , standard_warn=__a) _UpperCamelCase = dict(scheduler.config) _UpperCamelCase = 1 _UpperCamelCase = FrozenDict(__a) if hasattr(scheduler.config , '''skip_prk_steps''') and scheduler.config.skip_prk_steps is False: _UpperCamelCase = ( F'''The configuration file of this scheduler: {scheduler} has not set the configuration''' ''' `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make''' ''' sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to''' ''' incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face''' ''' Hub, it would be very nice if you could open a Pull request for the''' ''' `scheduler/scheduler_config.json` file''' ) deprecate('''skip_prk_steps not set''' , '''1.0.0''' , __a , standard_warn=__a) _UpperCamelCase = dict(scheduler.config) _UpperCamelCase = True _UpperCamelCase = FrozenDict(__a) if safety_checker is None: logger.warning( F'''You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure''' ''' that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered''' ''' results in services or applications open to the public. Both the diffusers team and Hugging Face''' ''' strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling''' ''' it only for use-cases that involve analyzing network behavior or auditing its results. For more''' ''' information, please have a look at https://github.com/huggingface/diffusers/pull/254 .''') self.register_modules( segmentation_model=__a , segmentation_processor=__a , vae=__a , text_encoder=__a , tokenizer=__a , unet=__a , scheduler=__a , safety_checker=__a , feature_extractor=__a , ) def UpperCAmelCase ( self , __a = "auto") -> Union[str, Any]: '''simple docstring''' if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory _UpperCamelCase = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(__a) def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' self.enable_attention_slicing(__a) def UpperCAmelCase ( self) -> Dict: '''simple docstring''' if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('''Please install accelerate via `pip install accelerate`''') _UpperCamelCase = torch.device('''cuda''') for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(__a , __a) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' if self.device != torch.device('''meta''') or not hasattr(self.unet , '''_hf_hook'''): return self.device for module in self.unet.modules(): if ( hasattr(__a , '''_hf_hook''') and hasattr(module._hf_hook , '''execution_device''') and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device) return self.device @torch.no_grad() def __call__( self , __a , __a , __a , __a = 5_12 , __a = 5_12 , __a = 50 , __a = 7.5 , __a = None , __a = 1 , __a = 0.0 , __a = None , __a = None , __a = "pil" , __a = True , __a = None , __a = 1 , **__a , ) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = self.segmentation_processor( text=[text] , images=[image] , padding='''max_length''' , return_tensors='''pt''').to(self.device) _UpperCamelCase = self.segmentation_model(**__a) _UpperCamelCase = torch.sigmoid(outputs.logits).cpu().detach().unsqueeze(-1).numpy() _UpperCamelCase = self.numpy_to_pil(__a)[0].resize(image.size) # Run inpainting pipeline with the generated mask _UpperCamelCase = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=__a , image=__a , mask_image=__a , height=__a , width=__a , num_inference_steps=__a , guidance_scale=__a , negative_prompt=__a , num_images_per_prompt=__a , eta=__a , generator=__a , latents=__a , output_type=__a , return_dict=__a , callback=__a , callback_steps=__a , )
78
"""simple docstring""" import copy import re class _UpperCAmelCase: lowercase__ = 'hp' lowercase__ = {} lowercase__ = None @classmethod def UpperCAmelCase ( cls , __a , __a) -> Dict: '''simple docstring''' _UpperCamelCase = prefix _UpperCamelCase = defaults cls.build_naming_info() @staticmethod def UpperCAmelCase ( __a , __a) -> Union[str, Any]: '''simple docstring''' if len(__a) == 0: return "" _UpperCamelCase = None if any(char.isdigit() for char in word): raise Exception(F'''Parameters should not contain numbers: \'{word}\' contains a number''') if word in info["short_word"]: return info["short_word"][word] for prefix_len in range(1 , len(__a) + 1): _UpperCamelCase = word[:prefix_len] if prefix in info["reverse_short_word"]: continue else: _UpperCamelCase = prefix break if short_word is None: # Paranoid fallback def int_to_alphabetic(__a): _UpperCamelCase = '''''' while integer != 0: _UpperCamelCase = chr(ord('''A''') + integer % 10) + s integer //= 10 return s _UpperCamelCase = 0 while True: _UpperCamelCase = word + '''#''' + int_to_alphabetic(__a) if sword in info["reverse_short_word"]: continue else: _UpperCamelCase = sword break _UpperCamelCase = short_word _UpperCamelCase = word return short_word @staticmethod def UpperCAmelCase ( __a , __a) -> Any: '''simple docstring''' _UpperCamelCase = param_name.split('''_''') _UpperCamelCase = [TrialShortNamer.shortname_for_word(__a , __a) for word in words] # We try to create a separatorless short name, but if there is a collision we have to fallback # to a separated short name _UpperCamelCase = ['''''', '''_'''] for separator in separators: _UpperCamelCase = separator.join(__a) if shortname not in info["reverse_short_param"]: _UpperCamelCase = shortname _UpperCamelCase = param_name return shortname return param_name @staticmethod def UpperCAmelCase ( __a , __a) -> List[str]: '''simple docstring''' _UpperCamelCase = TrialShortNamer.shortname_for_key(__a , __a) _UpperCamelCase = short_name _UpperCamelCase = param_name @classmethod def UpperCAmelCase ( cls) -> Any: '''simple docstring''' if cls.NAMING_INFO is not None: return _UpperCamelCase = { '''short_word''': {}, '''reverse_short_word''': {}, '''short_param''': {}, '''reverse_short_param''': {}, } _UpperCamelCase = list(cls.DEFAULTS.keys()) for k in field_keys: cls.add_new_param_name(__a , __a) _UpperCamelCase = info @classmethod def UpperCAmelCase ( cls , __a) -> Optional[Any]: '''simple docstring''' cls.build_naming_info() assert cls.PREFIX is not None _UpperCamelCase = [copy.copy(cls.PREFIX)] for k, v in params.items(): if k not in cls.DEFAULTS: raise Exception(F'''You should provide a default value for the param name {k} with value {v}''') if v == cls.DEFAULTS[k]: # The default value is not added to the name continue _UpperCamelCase = cls.NAMING_INFO['''short_param'''][k] if isinstance(__a , __a): _UpperCamelCase = 1 if v else 0 _UpperCamelCase = '''''' if isinstance(__a , (int, float)) else '''-''' _UpperCamelCase = F'''{key}{sep}{v}''' name.append(__a) return "_".join(__a) @classmethod def UpperCAmelCase ( cls , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = repr[len(cls.PREFIX) + 1 :] if repr == "": _UpperCamelCase = [] else: _UpperCamelCase = repr.split('''_''') _UpperCamelCase = {} for value in values: if "-" in value: _UpperCamelCase , _UpperCamelCase = value.split('''-''') else: _UpperCamelCase = re.sub('''[0-9.]''' , '''''' , __a) _UpperCamelCase = float(re.sub('''[^0-9.]''' , '''''' , __a)) _UpperCamelCase = cls.NAMING_INFO['''reverse_short_param'''][p_k] _UpperCamelCase = p_v for k in cls.DEFAULTS: if k not in parameters: _UpperCamelCase = cls.DEFAULTS[k] return parameters
78
1
"""simple docstring""" from __future__ import annotations def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> int | float: """simple docstring""" if len(__snake_case ) == 0: raise ValueError('''find_max() arg is an empty sequence''' ) if ( left >= len(__snake_case ) or left < -len(__snake_case ) or right >= len(__snake_case ) or right < -len(__snake_case ) ): raise IndexError('''list index out of range''' ) if left == right: return nums[left] _UpperCamelCase = (left + right) >> 1 # the middle _UpperCamelCase = find_max(__snake_case, __snake_case, __snake_case ) # find max in range[left, mid] _UpperCamelCase = find_max(__snake_case, mid + 1, __snake_case ) # find max in range[mid + 1, right] return left_max if left_max >= right_max else right_max if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
78
"""simple docstring""" import os import time import pytest from datasets.utils.filelock import FileLock, Timeout def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = FileLock(str(tmpdir / '''foo.lock''' ) ) _UpperCamelCase = FileLock(str(tmpdir / '''foo.lock''' ) ) _UpperCamelCase = 0.01 with locka.acquire(): with pytest.raises(__snake_case ): _UpperCamelCase = time.time() locka.acquire(__snake_case ) assert time.time() - _start > timeout def lowerCamelCase__ ( __snake_case ) -> List[Any]: """simple docstring""" _UpperCamelCase = '''a''' * 10_00 + '''.lock''' _UpperCamelCase = FileLock(str(tmpdir / filename ) ) assert locka._lock_file.endswith('''.lock''' ) assert not locka._lock_file.endswith(__snake_case ) assert len(os.path.basename(locka._lock_file ) ) <= 2_55 _UpperCamelCase = FileLock(tmpdir / filename ) with locka.acquire(): with pytest.raises(__snake_case ): locka.acquire(0 )
78
1
"""simple docstring""" import inspect import unittest from math import floor from transformers import CvtConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import CvtForImageClassification, CvtModel from transformers.models.cvt.modeling_cvt import CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _UpperCAmelCase( lowerCamelCase ): def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = self.config_class(**self.inputs_dict) self.parent.assertTrue(hasattr(__a , '''embed_dim''')) self.parent.assertTrue(hasattr(__a , '''num_heads''')) class _UpperCAmelCase: def __init__( self , __a , __a=13 , __a=64 , __a=3 , __a=[16, 48, 96] , __a=[1, 3, 6] , __a=[1, 2, 10] , __a=[7, 3, 3] , __a=[4, 2, 2] , __a=[2, 1, 1] , __a=[2, 2, 2] , __a=[False, False, True] , __a=[0.0, 0.0, 0.0] , __a=0.02 , __a=1e-12 , __a=True , __a=True , __a=2 , ) -> Optional[int]: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = image_size _UpperCamelCase = patch_sizes _UpperCamelCase = patch_stride _UpperCamelCase = patch_padding _UpperCamelCase = is_training _UpperCamelCase = use_labels _UpperCamelCase = num_labels _UpperCamelCase = num_channels _UpperCamelCase = embed_dim _UpperCamelCase = num_heads _UpperCamelCase = stride_kv _UpperCamelCase = depth _UpperCamelCase = cls_token _UpperCamelCase = attention_drop_rate _UpperCamelCase = initializer_range _UpperCamelCase = layer_norm_eps def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size] , self.num_labels) _UpperCamelCase = self.get_config() return config, pixel_values, labels def UpperCAmelCase ( self) -> Dict: '''simple docstring''' return CvtConfig( image_size=self.image_size , num_labels=self.num_labels , num_channels=self.num_channels , embed_dim=self.embed_dim , num_heads=self.num_heads , patch_sizes=self.patch_sizes , patch_padding=self.patch_padding , patch_stride=self.patch_stride , stride_kv=self.stride_kv , depth=self.depth , cls_token=self.cls_token , attention_drop_rate=self.attention_drop_rate , initializer_range=self.initializer_range , ) def UpperCAmelCase ( self , __a , __a , __a) -> Optional[int]: '''simple docstring''' _UpperCamelCase = CvtModel(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a) _UpperCamelCase = (self.image_size, self.image_size) _UpperCamelCase , _UpperCamelCase = image_size[0], image_size[1] for i in range(len(self.depth)): _UpperCamelCase = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1) _UpperCamelCase = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dim[-1], height, width)) def UpperCAmelCase ( self , __a , __a , __a) -> List[str]: '''simple docstring''' _UpperCamelCase = self.num_labels _UpperCamelCase = CvtForImageClassification(__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , labels=__a) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = self.prepare_config_and_inputs() _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = config_and_inputs _UpperCamelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class _UpperCAmelCase( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase__ = (CvtModel, CvtForImageClassification) if is_torch_available() else () lowercase__ = ( {'feature-extraction': CvtModel, 'image-classification': CvtForImageClassification} if is_torch_available() else {} ) lowercase__ = False lowercase__ = False lowercase__ = False lowercase__ = False lowercase__ = False def UpperCAmelCase ( self) -> Dict: '''simple docstring''' _UpperCamelCase = CvtModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=__a , has_text_modality=__a , hidden_size=37) def UpperCAmelCase ( self) -> Any: '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCAmelCase ( self) -> int: '''simple docstring''' return @unittest.skip(reason='''Cvt does not output attentions''') def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' pass @unittest.skip(reason='''Cvt does not use inputs_embeds''') def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' pass @unittest.skip(reason='''Cvt does not support input and output embeddings''') def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' pass def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = model_class(__a) _UpperCamelCase = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCamelCase = [*signature.parameters.keys()] _UpperCamelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , __a) def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' def check_hidden_states_output(__a , __a , __a): _UpperCamelCase = model_class(__a) model.to(__a) model.eval() with torch.no_grad(): _UpperCamelCase = model(**self._prepare_for_class(__a , __a)) _UpperCamelCase = outputs.hidden_states _UpperCamelCase = len(self.model_tester.depth) self.assertEqual(len(__a) , __a) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:]) , [ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = True check_hidden_states_output(__a , __a , __a) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCamelCase = True check_hidden_states_output(__a , __a , __a) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__a) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''') def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' pass @slow def UpperCAmelCase ( self) -> Any: '''simple docstring''' for model_name in CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase = CvtModel.from_pretrained(__a) self.assertIsNotNone(__a) def lowerCamelCase__ ( ) -> Tuple: """simple docstring""" _UpperCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class _UpperCAmelCase( unittest.TestCase ): @cached_property def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' return AutoImageProcessor.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0]) @slow def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = CvtForImageClassification.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0]).to(__a) _UpperCamelCase = self.default_image_processor _UpperCamelCase = prepare_img() _UpperCamelCase = image_processor(images=__a , return_tensors='''pt''').to(__a) # forward pass with torch.no_grad(): _UpperCamelCase = model(**__a) # verify the logits _UpperCamelCase = torch.Size((1, 10_00)) self.assertEqual(outputs.logits.shape , __a) _UpperCamelCase = torch.tensor([0.9285, 0.9015, -0.3150]).to(__a) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __a , atol=1e-4))
78
"""simple docstring""" from math import sqrt def lowerCamelCase__ ( __snake_case ) -> bool: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and ( number >= 0 ), "'number' must been an int and positive" _UpperCamelCase = True # 0 and 1 are none primes. if number <= 1: _UpperCamelCase = False for divisor in range(2, int(round(sqrt(__snake_case ) ) ) + 1 ): # if 'number' divisible by 'divisor' then sets 'status' # of false and break up the loop. if number % divisor == 0: _UpperCamelCase = False break # precondition assert isinstance(__snake_case, __snake_case ), "'status' must been from type bool" return status def lowerCamelCase__ ( __snake_case ) -> Dict: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n > 2), "'N' must been an int and > 2" # beginList: contains all natural numbers from 2 up to N _UpperCamelCase = list(range(2, n + 1 ) ) _UpperCamelCase = [] # this list will be returns. # actual sieve of erathostenes for i in range(len(__snake_case ) ): for j in range(i + 1, len(__snake_case ) ): if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0): _UpperCamelCase = 0 # filters actual prime numbers. _UpperCamelCase = [x for x in begin_list if x != 0] # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type list" return ans def lowerCamelCase__ ( __snake_case ) -> List[str]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n > 2), "'N' must been an int and > 2" _UpperCamelCase = [] # iterates over all numbers between 2 up to N+1 # if a number is prime then appends to list 'ans' for number in range(2, n + 1 ): if is_prime(__snake_case ): ans.append(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type list" return ans def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and number >= 0, "'number' must been an int and >= 0" _UpperCamelCase = [] # this list will be returns of the function. # potential prime number factors. _UpperCamelCase = 2 _UpperCamelCase = number if number == 0 or number == 1: ans.append(__snake_case ) # if 'number' not prime then builds the prime factorization of 'number' elif not is_prime(__snake_case ): while quotient != 1: if is_prime(__snake_case ) and (quotient % factor == 0): ans.append(__snake_case ) quotient /= factor else: factor += 1 else: ans.append(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type list" return ans def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and ( number >= 0 ), "'number' bust been an int and >= 0" _UpperCamelCase = 0 # prime factorization of 'number' _UpperCamelCase = prime_factorization(__snake_case ) _UpperCamelCase = max(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type int" return ans def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and ( number >= 0 ), "'number' bust been an int and >= 0" _UpperCamelCase = 0 # prime factorization of 'number' _UpperCamelCase = prime_factorization(__snake_case ) _UpperCamelCase = min(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type int" return ans def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ), "'number' must been an int" assert isinstance(number % 2 == 0, __snake_case ), "compare bust been from type bool" return number % 2 == 0 def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ), "'number' must been an int" assert isinstance(number % 2 != 0, __snake_case ), "compare bust been from type bool" return number % 2 != 0 def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" assert ( isinstance(__snake_case, __snake_case ) and (number > 2) and is_even(__snake_case ) ), "'number' must been an int, even and > 2" _UpperCamelCase = [] # this list will returned # creates a list of prime numbers between 2 up to 'number' _UpperCamelCase = get_prime_numbers(__snake_case ) _UpperCamelCase = len(__snake_case ) # run variable for while-loops. _UpperCamelCase = 0 _UpperCamelCase = None # exit variable. for break up the loops _UpperCamelCase = True while i < len_pn and loop: _UpperCamelCase = i + 1 while j < len_pn and loop: if prime_numbers[i] + prime_numbers[j] == number: _UpperCamelCase = False ans.append(prime_numbers[i] ) ans.append(prime_numbers[j] ) j += 1 i += 1 # precondition assert ( isinstance(__snake_case, __snake_case ) and (len(__snake_case ) == 2) and (ans[0] + ans[1] == number) and is_prime(ans[0] ) and is_prime(ans[1] ) ), "'ans' must contains two primes. And sum of elements must been eq 'number'" return ans def lowerCamelCase__ ( __snake_case, __snake_case ) -> str: """simple docstring""" assert ( isinstance(__snake_case, __snake_case ) and isinstance(__snake_case, __snake_case ) and (numbera >= 0) and (numbera >= 0) ), "'number1' and 'number2' must been positive integer." _UpperCamelCase = 0 while numbera != 0: _UpperCamelCase = numbera % numbera _UpperCamelCase = numbera _UpperCamelCase = rest # precondition assert isinstance(__snake_case, __snake_case ) and ( numbera >= 0 ), "'number' must been from type int and positive" return numbera def lowerCamelCase__ ( __snake_case, __snake_case ) -> List[str]: """simple docstring""" assert ( isinstance(__snake_case, __snake_case ) and isinstance(__snake_case, __snake_case ) and (numbera >= 1) and (numbera >= 1) ), "'number1' and 'number2' must been positive integer." _UpperCamelCase = 1 # actual answer that will be return. # for kgV (x,1) if numbera > 1 and numbera > 1: # builds the prime factorization of 'number1' and 'number2' _UpperCamelCase = prime_factorization(__snake_case ) _UpperCamelCase = prime_factorization(__snake_case ) elif numbera == 1 or numbera == 1: _UpperCamelCase = [] _UpperCamelCase = [] _UpperCamelCase = max(__snake_case, __snake_case ) _UpperCamelCase = 0 _UpperCamelCase = 0 _UpperCamelCase = [] # captured numbers int both 'primeFac1' and 'primeFac2' # iterates through primeFac1 for n in prime_fac_a: if n not in done: if n in prime_fac_a: _UpperCamelCase = prime_fac_a.count(__snake_case ) _UpperCamelCase = prime_fac_a.count(__snake_case ) for _ in range(max(__snake_case, __snake_case ) ): ans *= n else: _UpperCamelCase = prime_fac_a.count(__snake_case ) for _ in range(__snake_case ): ans *= n done.append(__snake_case ) # iterates through primeFac2 for n in prime_fac_a: if n not in done: _UpperCamelCase = prime_fac_a.count(__snake_case ) for _ in range(__snake_case ): ans *= n done.append(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ) and ( ans >= 0 ), "'ans' must been from type int and positive" return ans def lowerCamelCase__ ( __snake_case ) -> Tuple: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n >= 0), "'number' must been a positive int" _UpperCamelCase = 0 _UpperCamelCase = 2 # this variable holds the answer while index < n: index += 1 ans += 1 # counts to the next number # if ans not prime then # runs to the next prime number. while not is_prime(__snake_case ): ans += 1 # precondition assert isinstance(__snake_case, __snake_case ) and is_prime( __snake_case ), "'ans' must been a prime number and from type int" return ans def lowerCamelCase__ ( __snake_case, __snake_case ) -> Tuple: """simple docstring""" assert ( is_prime(__snake_case ) and is_prime(__snake_case ) and (p_number_a < p_number_a) ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'" _UpperCamelCase = p_number_a + 1 # jump to the next number _UpperCamelCase = [] # this list will be returns. # if number is not prime then # fetch the next prime number. while not is_prime(__snake_case ): number += 1 while number < p_number_a: ans.append(__snake_case ) number += 1 # fetch the next prime number. while not is_prime(__snake_case ): number += 1 # precondition assert ( isinstance(__snake_case, __snake_case ) and ans[0] != p_number_a and ans[len(__snake_case ) - 1] != p_number_a ), "'ans' must been a list without the arguments" # 'ans' contains not 'pNumber1' and 'pNumber2' ! return ans def lowerCamelCase__ ( __snake_case ) -> List[Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n >= 1), "'n' must been int and >= 1" _UpperCamelCase = [] # will be returned. for divisor in range(1, n + 1 ): if n % divisor == 0: ans.append(__snake_case ) # precondition assert ans[0] == 1 and ans[len(__snake_case ) - 1] == n, "Error in function getDivisiors(...)" return ans def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and ( number > 1 ), "'number' must been an int and >= 1" _UpperCamelCase = get_divisors(__snake_case ) # precondition assert ( isinstance(__snake_case, __snake_case ) and (divisors[0] == 1) and (divisors[len(__snake_case ) - 1] == number) ), "Error in help-function getDivisiors(...)" # summed all divisors up to 'number' (exclusive), hence [:-1] return sum(divisors[:-1] ) == number def lowerCamelCase__ ( __snake_case, __snake_case ) -> Optional[Any]: """simple docstring""" assert ( isinstance(__snake_case, __snake_case ) and isinstance(__snake_case, __snake_case ) and (denominator != 0) ), "The arguments must been from type int and 'denominator' != 0" # build the greatest common divisor of numerator and denominator. _UpperCamelCase = gcd(abs(__snake_case ), abs(__snake_case ) ) # precondition assert ( isinstance(__snake_case, __snake_case ) and (numerator % gcd_of_fraction == 0) and (denominator % gcd_of_fraction == 0) ), "Error in function gcd(...,...)" return (numerator // gcd_of_fraction, denominator // gcd_of_fraction) def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n >= 0), "'n' must been a int and >= 0" _UpperCamelCase = 1 # this will be return. for factor in range(1, n + 1 ): ans *= factor return ans def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n >= 0), "'n' must been an int and >= 0" _UpperCamelCase = 0 _UpperCamelCase = 1 _UpperCamelCase = 1 # this will be return for _ in range(n - 1 ): _UpperCamelCase = ans ans += fiba _UpperCamelCase = tmp return ans
78
1
"""simple docstring""" from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging _a = logging.get_logger(__name__) _a = { """EleutherAI/gpt-neo-1.3B""": """https://huggingface.co/EleutherAI/gpt-neo-1.3B/resolve/main/config.json""", # See all GPTNeo models at https://huggingface.co/models?filter=gpt_neo } class _UpperCAmelCase( lowerCamelCase ): lowercase__ = 'gpt_neo' lowercase__ = ['past_key_values'] lowercase__ = {'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers'} def __init__( self , __a=5_02_57 , __a=20_48 , __a=20_48 , __a=24 , __a=[[["global", "local"], 12]] , __a=16 , __a=None , __a=2_56 , __a="gelu_new" , __a=0.0 , __a=0.0 , __a=0.0 , __a=0.1 , __a=1e-5 , __a=0.02 , __a=True , __a=5_02_56 , __a=5_02_56 , **__a , ) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = vocab_size _UpperCamelCase = max_position_embeddings _UpperCamelCase = hidden_size _UpperCamelCase = num_layers _UpperCamelCase = num_heads _UpperCamelCase = intermediate_size _UpperCamelCase = window_size _UpperCamelCase = activation_function _UpperCamelCase = resid_dropout _UpperCamelCase = embed_dropout _UpperCamelCase = attention_dropout _UpperCamelCase = classifier_dropout _UpperCamelCase = layer_norm_epsilon _UpperCamelCase = initializer_range _UpperCamelCase = use_cache _UpperCamelCase = bos_token_id _UpperCamelCase = eos_token_id _UpperCamelCase = attention_types _UpperCamelCase = self.expand_attention_types_params(__a) if len(self.attention_layers) != self.num_layers: raise ValueError( '''Configuration for convolutional module is incorrect. ''' '''It is required that `len(config.attention_layers)` == `config.num_layers` ''' F'''but is `len(config.attention_layers) = {len(self.attention_layers)}`, ''' F'''`config.num_layers = {self.num_layers}`. ''' '''`config.attention_layers` is prepared using `config.attention_types`. ''' '''Please verify the value of `config.attention_types` argument.''') super().__init__(bos_token_id=__a , eos_token_id=__a , **__a) @staticmethod def UpperCAmelCase ( __a) -> int: '''simple docstring''' _UpperCamelCase = [] for item in attention_types: for _ in range(item[1]): attentions.extend(item[0]) return attentions def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case, __snake_case ) -> str: """simple docstring""" import torch _UpperCamelCase = input.size() _UpperCamelCase = len(__snake_case ) _UpperCamelCase = shape[dimension] _UpperCamelCase = torch.arange(0, __snake_case, __snake_case ) _UpperCamelCase = torch.div(sizedim - size, __snake_case, rounding_mode='''floor''' ) + 1 _UpperCamelCase = torch.arange(__snake_case ) + low_indices[:min_length][:, None] _UpperCamelCase = [slice(__snake_case )] * rank _UpperCamelCase = indices _UpperCamelCase = input[s] _UpperCamelCase = list(range(0, rank + 1 ) ) perm.append(perm.pop(dimension + 1 ) ) return sliced.permute(__snake_case ) def lowerCamelCase__ ( __snake_case, __snake_case ) -> str: """simple docstring""" import torch _UpperCamelCase = torch.arange(1, __snake_case ) _UpperCamelCase = torch.remainder(__snake_case, __snake_case ) _UpperCamelCase = remainders == 0 _UpperCamelCase = candidates[divisor_indices] _UpperCamelCase = torch.max(__snake_case ) return largest_divisor, torch.div(__snake_case, __snake_case, rounding_mode='''floor''' ) class _UpperCAmelCase( lowerCamelCase ): @property def UpperCAmelCase ( self) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' _UpperCamelCase = OrderedDict({'''input_ids''': {0: '''batch''', 1: '''sequence'''}}) if self.use_past: self.fill_with_past_key_values_(__a , direction='''inputs''') _UpperCamelCase = {0: '''batch''', 1: '''past_sequence + sequence'''} else: _UpperCamelCase = {0: '''batch''', 1: '''sequence'''} return common_inputs @property def UpperCAmelCase ( self) -> int: '''simple docstring''' return self._config.num_heads def UpperCAmelCase ( self , __a , __a = -1 , __a = -1 , __a = False , __a = None , ) -> Mapping[str, Any]: '''simple docstring''' _UpperCamelCase = super(__a , self).generate_dummy_inputs( __a , batch_size=__a , seq_length=__a , is_pair=__a , framework=__a) # We need to order the input in the way they appears in the forward() _UpperCamelCase = OrderedDict({'''input_ids''': common_inputs['''input_ids''']}) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''') else: import torch _UpperCamelCase , _UpperCamelCase = common_inputs['''input_ids'''].shape # Not using the same length for past_key_values _UpperCamelCase = seqlen + 2 _UpperCamelCase = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) _UpperCamelCase = [ (torch.zeros(__a), torch.zeros(__a)) for _ in range(self.num_layers) ] _UpperCamelCase = common_inputs['''attention_mask'''] if self.use_past: _UpperCamelCase = ordered_inputs['''attention_mask'''].dtype _UpperCamelCase = torch.cat( [ordered_inputs['''attention_mask'''], torch.ones(__a , __a , dtype=__a)] , dim=1) return ordered_inputs @property def UpperCAmelCase ( self) -> int: '''simple docstring''' return 13
78
"""simple docstring""" import logging from dataclasses import dataclass, field from pathlib import Path from typing import Optional, Union from .generation.configuration_utils import GenerationConfig from .training_args import TrainingArguments from .utils import add_start_docstrings _a = logging.getLogger(__name__) @dataclass @add_start_docstrings(TrainingArguments.__doc__ ) class _UpperCAmelCase( lowerCamelCase ): lowercase__ = field(default=lowerCamelCase , metadata={'help': 'Whether to use SortishSampler or not.'} ) lowercase__ = field( default=lowerCamelCase , metadata={'help': 'Whether to use generate to calculate generative metrics (ROUGE, BLEU).'} ) lowercase__ = field( default=lowerCamelCase , metadata={ 'help': ( 'The `max_length` to use on each evaluation loop when `predict_with_generate=True`. Will default ' 'to the `max_length` value of the model configuration.' ) } , ) lowercase__ = field( default=lowerCamelCase , metadata={ 'help': ( 'The `num_beams` to use on each evaluation loop when `predict_with_generate=True`. Will default ' 'to the `num_beams` value of the model configuration.' ) } , ) lowercase__ = field( default=lowerCamelCase , metadata={ 'help': 'Model id, file path or url pointing to a GenerationConfig json file, to use during prediction.' } , ) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = super().to_dict() for k, v in d.items(): if isinstance(__a , __a): _UpperCamelCase = v.to_dict() return d
78
1
"""simple docstring""" import unittest import numpy as np import torch from diffusers import KarrasVePipeline, KarrasVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class _UpperCAmelCase( unittest.TestCase ): @property def UpperCAmelCase ( self) -> str: '''simple docstring''' torch.manual_seed(0) _UpperCamelCase = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('''DownBlock2D''', '''AttnDownBlock2D''') , up_block_types=('''AttnUpBlock2D''', '''UpBlock2D''') , ) return model def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.dummy_uncond_unet _UpperCamelCase = KarrasVeScheduler() _UpperCamelCase = KarrasVePipeline(unet=__a , scheduler=__a) pipe.to(__a) pipe.set_progress_bar_config(disable=__a) _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe(num_inference_steps=2 , generator=__a , output_type='''numpy''').images _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe(num_inference_steps=2 , generator=__a , output_type='''numpy''' , return_dict=__a)[0] _UpperCamelCase = image[0, -3:, -3:, -1] _UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _UpperCamelCase = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 @slow @require_torch class _UpperCAmelCase( unittest.TestCase ): def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = '''google/ncsnpp-celebahq-256''' _UpperCamelCase = UNetaDModel.from_pretrained(__a) _UpperCamelCase = KarrasVeScheduler() _UpperCamelCase = KarrasVePipeline(unet=__a , scheduler=__a) pipe.to(__a) pipe.set_progress_bar_config(disable=__a) _UpperCamelCase = torch.manual_seed(0) _UpperCamelCase = pipe(num_inference_steps=20 , generator=__a , output_type='''numpy''').images _UpperCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 2_56, 2_56, 3) _UpperCamelCase = np.array([0.578, 0.5811, 0.5924, 0.5809, 0.587, 0.5886, 0.5861, 0.5802, 0.586]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2
78
"""simple docstring""" import argparse import torch from transformers import BlenderbotConfig, BlenderbotForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) _a = [ ["""attention""", """attn"""], ["""encoder_attention""", """encoder_attn"""], ["""q_lin""", """q_proj"""], ["""k_lin""", """k_proj"""], ["""v_lin""", """v_proj"""], ["""out_lin""", """out_proj"""], ["""norm_embeddings""", """layernorm_embedding"""], ["""position_embeddings""", """embed_positions"""], ["""embeddings""", """embed_tokens"""], ["""ffn.lin""", """fc"""], ] def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" if k == "embeddings.weight": return "shared.weight" for parlai_name, hf_name in PATTERNS: _UpperCamelCase = k.replace(__snake_case, __snake_case ) if k.startswith('''encoder''' ): _UpperCamelCase = k.replace('''.attn''', '''.self_attn''' ) _UpperCamelCase = k.replace('''norm1''', '''self_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm2''', '''final_layer_norm''' ) elif k.startswith('''decoder''' ): _UpperCamelCase = k.replace('''norm1''', '''self_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm2''', '''encoder_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm3''', '''final_layer_norm''' ) return k def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" _UpperCamelCase = [ '''model.encoder.layernorm_embedding.weight''', '''model.encoder.layernorm_embedding.bias''', '''model.decoder.layernorm_embedding.weight''', '''model.decoder.layernorm_embedding.bias''', ] for k in keys: _UpperCamelCase = sd.pop(__snake_case ) _UpperCamelCase = k.replace('''layernorm_embedding''', '''layer_norm''' ) assert new_k not in sd _UpperCamelCase = v _a = ["""START"""] @torch.no_grad() def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> int: """simple docstring""" _UpperCamelCase = torch.load(__snake_case, map_location='''cpu''' ) _UpperCamelCase = model['''model'''] _UpperCamelCase = BlenderbotConfig.from_json_file(__snake_case ) _UpperCamelCase = BlenderbotForConditionalGeneration(__snake_case ) _UpperCamelCase = m.model.state_dict().keys() _UpperCamelCase = [] _UpperCamelCase = {} for k, v in sd.items(): if k in IGNORE_KEYS: continue _UpperCamelCase = rename_state_dict_key(__snake_case ) if new_k not in valid_keys: failures.append([k, new_k] ) else: _UpperCamelCase = v if cfg.normalize_before: # Blenderbot-3B checkpoints. Rename layernorm_embedding -> layer_norm rename_layernorm_keys(__snake_case ) m.model.load_state_dict(__snake_case, strict=__snake_case ) m.half() m.save_pretrained(__snake_case ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument("""--src_path""", type=str, help="""like blenderbot-model.bin""") parser.add_argument("""--save_dir""", default="""hf_blenderbot""", type=str, help="""Where to save converted model.""") parser.add_argument( """--hf_config_json""", default="""blenderbot-3b-config.json""", type=str, help="""Path to config to use""" ) _a = parser.parse_args() convert_parlai_checkpoint(args.src_path, args.save_dir, args.hf_config_json)
78
1
"""simple docstring""" import unittest from transformers import SPIECE_UNDERLINE, ReformerTokenizer, ReformerTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _a = get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece @require_tokenizers class _UpperCAmelCase( lowerCamelCase , unittest.TestCase ): lowercase__ = ReformerTokenizer lowercase__ = ReformerTokenizerFast lowercase__ = True lowercase__ = False lowercase__ = True def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' super().setUp() _UpperCamelCase = ReformerTokenizer(__a , keep_accents=__a) tokenizer.save_pretrained(self.tmpdirname) def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = '''<s>''' _UpperCamelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__a) , __a) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__a) , __a) def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = list(self.get_tokenizer().get_vocab().keys()) self.assertEqual(vocab_keys[0] , '''<unk>''') self.assertEqual(vocab_keys[1] , '''<s>''') self.assertEqual(vocab_keys[-1] , '''j''') self.assertEqual(len(__a) , 10_00) def UpperCAmelCase ( self) -> Dict: '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 10_00) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' if not self.test_rust_tokenizer: return _UpperCamelCase = self.get_tokenizer() _UpperCamelCase = self.get_rust_tokenizer() _UpperCamelCase = '''I was born in 92000, and this is falsé.''' _UpperCamelCase = tokenizer.tokenize(__a) _UpperCamelCase = rust_tokenizer.tokenize(__a) self.assertListEqual(__a , __a) _UpperCamelCase = tokenizer.encode(__a , add_special_tokens=__a) _UpperCamelCase = rust_tokenizer.encode(__a , add_special_tokens=__a) self.assertListEqual(__a , __a) _UpperCamelCase = self.get_rust_tokenizer() _UpperCamelCase = tokenizer.encode(__a) _UpperCamelCase = rust_tokenizer.encode(__a) self.assertListEqual(__a , __a) def UpperCAmelCase ( self , __a=15) -> Optional[Any]: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})'''): _UpperCamelCase = self.rust_tokenizer_class.from_pretrained(__a , **__a) # Simple input _UpperCamelCase = '''This is a simple input''' _UpperCamelCase = ['''This is a simple input 1''', '''This is a simple input 2'''] _UpperCamelCase = ('''This is a simple input''', '''This is a pair''') _UpperCamelCase = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(__a , tokenizer_r.encode , __a , max_length=__a , padding='''max_length''') # Simple input self.assertRaises(__a , tokenizer_r.encode_plus , __a , max_length=__a , padding='''max_length''') # Simple input self.assertRaises( __a , tokenizer_r.batch_encode_plus , __a , max_length=__a , padding='''max_length''' , ) # Pair input self.assertRaises(__a , tokenizer_r.encode , __a , max_length=__a , padding='''max_length''') # Pair input self.assertRaises(__a , tokenizer_r.encode_plus , __a , max_length=__a , padding='''max_length''') # Pair input self.assertRaises( __a , tokenizer_r.batch_encode_plus , __a , max_length=__a , padding='''max_length''' , ) def UpperCAmelCase ( self) -> Dict: '''simple docstring''' pass def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = ReformerTokenizer(__a , keep_accents=__a) _UpperCamelCase = tokenizer.tokenize('''This is a test''') self.assertListEqual(__a , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est''']) self.assertListEqual( tokenizer.convert_tokens_to_ids(__a) , [2_85, 46, 10, 1_70, 3_82] , ) _UpperCamelCase = tokenizer.tokenize('''I was born in 92000, and this is falsé.''') self.assertListEqual( __a , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) _UpperCamelCase = tokenizer.convert_tokens_to_ids(__a) self.assertListEqual( __a , [8, 21, 84, 55, 24, 19, 7, 0, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) _UpperCamelCase = tokenizer.convert_ids_to_tokens(__a) self.assertListEqual( __a , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) @cached_property def UpperCAmelCase ( self) -> Dict: '''simple docstring''' return ReformerTokenizer.from_pretrained('''google/reformer-crime-and-punishment''') @slow def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = '''Hello World!''' _UpperCamelCase = [1_26, 32, 2_62, 1_52, 38, 72, 2_87] self.assertListEqual(__a , self.big_tokenizer.encode(__a)) @slow def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) _UpperCamelCase = [ 1_08, 2_65, 24, 1_11, 4, 2_58, 1_56, 35, 28, 2_75, 3, 2_59, 2_97, 2_60, 84, 4, 35, 1_10, 44, 8, 2_59, 91, 2_68, 21, 11, 2_09, 2_74, 1_09, 2_66, 2_77, 1_17, 86, 93, 3_15, 2_58, 2_78, 2_58, 2_77, 2_58, 0, 2_58, 2_88, 2_58, 3_19, 2_58, 0, 2_58, 0, 2_58, 0, 2_58, 0, 2_58, 2_87, 2_58, 3_15, 2_58, 2_89, 2_58, 2_78, 99, 2_69, 2_66, 2_62, 8, 2_59, 2_41, 4, 2_17, 2_30, 2_68, 2_66, 55, 1_68, 1_06, 75, 1_93, 2_66, 2_23, 27, 49, 26, 2_82, 25, 2_64, 2_99, 19, 26, 0, 2_58, 2_77, 1_17, 86, 93, 1_76, 1_83, 2_70, 11, 2_62, 42, 61, 2_65, ] self.assertListEqual(__a , self.big_tokenizer.encode(__a)) @require_torch @slow def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' import torch from transformers import ReformerConfig, ReformerModel # Build sequence _UpperCamelCase = list(self.big_tokenizer.get_vocab().keys())[:10] _UpperCamelCase = ''' '''.join(__a) _UpperCamelCase = self.big_tokenizer.encode_plus(__a , return_tensors='''pt''') _UpperCamelCase = self.big_tokenizer.batch_encode_plus([sequence, sequence] , return_tensors='''pt''') _UpperCamelCase = ReformerConfig() # The input gets padded during training so adjust the axial position encodings from the pretrained model value of (512, 1024) _UpperCamelCase = encoded_sequence['''input_ids'''].shape _UpperCamelCase = ReformerModel(__a) # Reformer has config.vocab_size == tokenizer.vocab_size == len(tokenizer) - 1 = 320; len(tokenizer) is 321 (including a pad token with id 320) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**__a) model(**__a) @slow def UpperCAmelCase ( self) -> int: '''simple docstring''' # fmt: off _UpperCamelCase = {'''input_ids''': [[1_08, 2_65, 24, 1_11, 4, 2_58, 1_56, 7, 51, 2_79, 58, 7, 76, 25, 69, 2_78], [1_40, 2_43, 2_64, 1_34, 17, 2_67, 77, 2_63, 22, 2_62, 2_97, 2_58, 3_04, 1_77, 2_79, 2_66, 14, 89, 13, 35, 2_61, 2_99, 2_72, 1_37, 2_75, 2_78]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # This tokenizer does not know some characters like ")". # That is the reason why we use very simple texts here. # Also see https://github.com/huggingface/transformers/pull/11737#issuecomment-850769064 _UpperCamelCase = [ '''This is a very simple sentence.''', '''The quick brown fox jumps over the lazy dog.''', ] self.tokenizer_integration_test_util( expected_encoding=__a , model_name='''google/reformer-crime-and-punishment''' , revision='''0e6c3decb8211d49bf881013425dc8b0448b3f5a''' , padding=__a , sequences=__a , )
78
"""simple docstring""" import argparse import os.path as osp import re import torch from safetensors.torch import load_file, save_file # =================# # UNet Conversion # # =================# _a = [ # (stable-diffusion, HF Diffusers) ("""time_embed.0.weight""", """time_embedding.linear_1.weight"""), ("""time_embed.0.bias""", """time_embedding.linear_1.bias"""), ("""time_embed.2.weight""", """time_embedding.linear_2.weight"""), ("""time_embed.2.bias""", """time_embedding.linear_2.bias"""), ("""input_blocks.0.0.weight""", """conv_in.weight"""), ("""input_blocks.0.0.bias""", """conv_in.bias"""), ("""out.0.weight""", """conv_norm_out.weight"""), ("""out.0.bias""", """conv_norm_out.bias"""), ("""out.2.weight""", """conv_out.weight"""), ("""out.2.bias""", """conv_out.bias"""), ] _a = [ # (stable-diffusion, HF Diffusers) ("""in_layers.0""", """norm1"""), ("""in_layers.2""", """conv1"""), ("""out_layers.0""", """norm2"""), ("""out_layers.3""", """conv2"""), ("""emb_layers.1""", """time_emb_proj"""), ("""skip_connection""", """conv_shortcut"""), ] _a = [] # hardcoded number of downblocks and resnets/attentions... # would need smarter logic for other networks. for i in range(4): # loop over downblocks/upblocks for j in range(2): # loop over resnets/attentions for downblocks _a = F"""down_blocks.{i}.resnets.{j}.""" _a = F"""input_blocks.{3*i + j + 1}.0.""" unet_conversion_map_layer.append((sd_down_res_prefix, hf_down_res_prefix)) if i < 3: # no attention layers in down_blocks.3 _a = F"""down_blocks.{i}.attentions.{j}.""" _a = F"""input_blocks.{3*i + j + 1}.1.""" unet_conversion_map_layer.append((sd_down_atn_prefix, hf_down_atn_prefix)) for j in range(3): # loop over resnets/attentions for upblocks _a = F"""up_blocks.{i}.resnets.{j}.""" _a = F"""output_blocks.{3*i + j}.0.""" unet_conversion_map_layer.append((sd_up_res_prefix, hf_up_res_prefix)) if i > 0: # no attention layers in up_blocks.0 _a = F"""up_blocks.{i}.attentions.{j}.""" _a = F"""output_blocks.{3*i + j}.1.""" unet_conversion_map_layer.append((sd_up_atn_prefix, hf_up_atn_prefix)) if i < 3: # no downsample in down_blocks.3 _a = F"""down_blocks.{i}.downsamplers.0.conv.""" _a = F"""input_blocks.{3*(i+1)}.0.op.""" unet_conversion_map_layer.append((sd_downsample_prefix, hf_downsample_prefix)) # no upsample in up_blocks.3 _a = F"""up_blocks.{i}.upsamplers.0.""" _a = F"""output_blocks.{3*i + 2}.{1 if i == 0 else 2}.""" unet_conversion_map_layer.append((sd_upsample_prefix, hf_upsample_prefix)) _a = """mid_block.attentions.0.""" _a = """middle_block.1.""" unet_conversion_map_layer.append((sd_mid_atn_prefix, hf_mid_atn_prefix)) for j in range(2): _a = F"""mid_block.resnets.{j}.""" _a = F"""middle_block.{2*j}.""" unet_conversion_map_layer.append((sd_mid_res_prefix, hf_mid_res_prefix)) def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" _UpperCamelCase = {k: k for k in unet_state_dict.keys()} for sd_name, hf_name in unet_conversion_map: _UpperCamelCase = sd_name for k, v in mapping.items(): if "resnets" in k: for sd_part, hf_part in unet_conversion_map_resnet: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v for k, v in mapping.items(): for sd_part, hf_part in unet_conversion_map_layer: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v _UpperCamelCase = {v: unet_state_dict[k] for k, v in mapping.items()} return new_state_dict # ================# # VAE Conversion # # ================# _a = [ # (stable-diffusion, HF Diffusers) ("""nin_shortcut""", """conv_shortcut"""), ("""norm_out""", """conv_norm_out"""), ("""mid.attn_1.""", """mid_block.attentions.0."""), ] for i in range(4): # down_blocks have two resnets for j in range(2): _a = F"""encoder.down_blocks.{i}.resnets.{j}.""" _a = F"""encoder.down.{i}.block.{j}.""" vae_conversion_map.append((sd_down_prefix, hf_down_prefix)) if i < 3: _a = F"""down_blocks.{i}.downsamplers.0.""" _a = F"""down.{i}.downsample.""" vae_conversion_map.append((sd_downsample_prefix, hf_downsample_prefix)) _a = F"""up_blocks.{i}.upsamplers.0.""" _a = F"""up.{3-i}.upsample.""" vae_conversion_map.append((sd_upsample_prefix, hf_upsample_prefix)) # up_blocks have three resnets # also, up blocks in hf are numbered in reverse from sd for j in range(3): _a = F"""decoder.up_blocks.{i}.resnets.{j}.""" _a = F"""decoder.up.{3-i}.block.{j}.""" vae_conversion_map.append((sd_up_prefix, hf_up_prefix)) # this part accounts for mid blocks in both the encoder and the decoder for i in range(2): _a = F"""mid_block.resnets.{i}.""" _a = F"""mid.block_{i+1}.""" vae_conversion_map.append((sd_mid_res_prefix, hf_mid_res_prefix)) _a = [ # (stable-diffusion, HF Diffusers) ("""norm.""", """group_norm."""), ("""q.""", """query."""), ("""k.""", """key."""), ("""v.""", """value."""), ("""proj_out.""", """proj_attn."""), ] def lowerCamelCase__ ( __snake_case ) -> List[str]: """simple docstring""" return w.reshape(*w.shape, 1, 1 ) def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" _UpperCamelCase = {k: k for k in vae_state_dict.keys()} for k, v in mapping.items(): for sd_part, hf_part in vae_conversion_map: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v for k, v in mapping.items(): if "attentions" in k: for sd_part, hf_part in vae_conversion_map_attn: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v _UpperCamelCase = {v: vae_state_dict[k] for k, v in mapping.items()} _UpperCamelCase = ['''q''', '''k''', '''v''', '''proj_out'''] for k, v in new_state_dict.items(): for weight_name in weights_to_convert: if F'''mid.attn_1.{weight_name}.weight''' in k: print(F'''Reshaping {k} for SD format''' ) _UpperCamelCase = reshape_weight_for_sd(__snake_case ) return new_state_dict # =========================# # Text Encoder Conversion # # =========================# _a = [ # (stable-diffusion, HF Diffusers) ("""resblocks.""", """text_model.encoder.layers."""), ("""ln_1""", """layer_norm1"""), ("""ln_2""", """layer_norm2"""), (""".c_fc.""", """.fc1."""), (""".c_proj.""", """.fc2."""), (""".attn""", """.self_attn"""), ("""ln_final.""", """transformer.text_model.final_layer_norm."""), ("""token_embedding.weight""", """transformer.text_model.embeddings.token_embedding.weight"""), ("""positional_embedding""", """transformer.text_model.embeddings.position_embedding.weight"""), ] _a = {re.escape(x[1]): x[0] for x in textenc_conversion_lst} _a = re.compile("""|""".join(protected.keys())) # Ordering is from https://github.com/pytorch/pytorch/blob/master/test/cpp/api/modules.cpp _a = {"""q""": 0, """k""": 1, """v""": 2} def lowerCamelCase__ ( __snake_case ) -> Any: """simple docstring""" _UpperCamelCase = {} _UpperCamelCase = {} _UpperCamelCase = {} for k, v in text_enc_dict.items(): if ( k.endswith('''.self_attn.q_proj.weight''' ) or k.endswith('''.self_attn.k_proj.weight''' ) or k.endswith('''.self_attn.v_proj.weight''' ) ): _UpperCamelCase = k[: -len('''.q_proj.weight''' )] _UpperCamelCase = k[-len('''q_proj.weight''' )] if k_pre not in capture_qkv_weight: _UpperCamelCase = [None, None, None] _UpperCamelCase = v continue if ( k.endswith('''.self_attn.q_proj.bias''' ) or k.endswith('''.self_attn.k_proj.bias''' ) or k.endswith('''.self_attn.v_proj.bias''' ) ): _UpperCamelCase = k[: -len('''.q_proj.bias''' )] _UpperCamelCase = k[-len('''q_proj.bias''' )] if k_pre not in capture_qkv_bias: _UpperCamelCase = [None, None, None] _UpperCamelCase = v continue _UpperCamelCase = textenc_pattern.sub(lambda __snake_case : protected[re.escape(m.group(0 ) )], __snake_case ) _UpperCamelCase = v for k_pre, tensors in capture_qkv_weight.items(): if None in tensors: raise Exception('''CORRUPTED MODEL: one of the q-k-v values for the text encoder was missing''' ) _UpperCamelCase = textenc_pattern.sub(lambda __snake_case : protected[re.escape(m.group(0 ) )], __snake_case ) _UpperCamelCase = torch.cat(__snake_case ) for k_pre, tensors in capture_qkv_bias.items(): if None in tensors: raise Exception('''CORRUPTED MODEL: one of the q-k-v values for the text encoder was missing''' ) _UpperCamelCase = textenc_pattern.sub(lambda __snake_case : protected[re.escape(m.group(0 ) )], __snake_case ) _UpperCamelCase = torch.cat(__snake_case ) return new_state_dict def lowerCamelCase__ ( __snake_case ) -> Tuple: """simple docstring""" return text_enc_dict if __name__ == "__main__": _a = argparse.ArgumentParser() parser.add_argument("""--model_path""", default=None, type=str, required=True, help="""Path to the model to convert.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, required=True, help="""Path to the output model.""") parser.add_argument("""--half""", action="""store_true""", help="""Save weights in half precision.""") parser.add_argument( """--use_safetensors""", action="""store_true""", help="""Save weights use safetensors, default is ckpt.""" ) _a = parser.parse_args() assert args.model_path is not None, "Must provide a model path!" assert args.checkpoint_path is not None, "Must provide a checkpoint path!" # Path for safetensors _a = osp.join(args.model_path, """unet""", """diffusion_pytorch_model.safetensors""") _a = osp.join(args.model_path, """vae""", """diffusion_pytorch_model.safetensors""") _a = osp.join(args.model_path, """text_encoder""", """model.safetensors""") # Load models from safetensors if it exists, if it doesn't pytorch if osp.exists(unet_path): _a = load_file(unet_path, device="""cpu""") else: _a = osp.join(args.model_path, """unet""", """diffusion_pytorch_model.bin""") _a = torch.load(unet_path, map_location="""cpu""") if osp.exists(vae_path): _a = load_file(vae_path, device="""cpu""") else: _a = osp.join(args.model_path, """vae""", """diffusion_pytorch_model.bin""") _a = torch.load(vae_path, map_location="""cpu""") if osp.exists(text_enc_path): _a = load_file(text_enc_path, device="""cpu""") else: _a = osp.join(args.model_path, """text_encoder""", """pytorch_model.bin""") _a = torch.load(text_enc_path, map_location="""cpu""") # Convert the UNet model _a = convert_unet_state_dict(unet_state_dict) _a = {"""model.diffusion_model.""" + k: v for k, v in unet_state_dict.items()} # Convert the VAE model _a = convert_vae_state_dict(vae_state_dict) _a = {"""first_stage_model.""" + k: v for k, v in vae_state_dict.items()} # Easiest way to identify v2.0 model seems to be that the text encoder (OpenCLIP) is deeper _a = """text_model.encoder.layers.22.layer_norm2.bias""" in text_enc_dict if is_vaa_model: # Need to add the tag 'transformer' in advance so we can knock it out from the final layer-norm _a = {"""transformer.""" + k: v for k, v in text_enc_dict.items()} _a = convert_text_enc_state_dict_vaa(text_enc_dict) _a = {"""cond_stage_model.model.""" + k: v for k, v in text_enc_dict.items()} else: _a = convert_text_enc_state_dict(text_enc_dict) _a = {"""cond_stage_model.transformer.""" + k: v for k, v in text_enc_dict.items()} # Put together new checkpoint _a = {**unet_state_dict, **vae_state_dict, **text_enc_dict} if args.half: _a = {k: v.half() for k, v in state_dict.items()} if args.use_safetensors: save_file(state_dict, args.checkpoint_path) else: _a = {"""state_dict""": state_dict} torch.save(state_dict, args.checkpoint_path)
78
1
"""simple docstring""" import sys from collections import defaultdict class _UpperCAmelCase: def __init__( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = [] def UpperCAmelCase ( self , __a) -> Optional[Any]: '''simple docstring''' return self.node_position[vertex] def UpperCAmelCase ( self , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = pos def UpperCAmelCase ( self , __a , __a , __a , __a) -> Tuple: '''simple docstring''' if start > size // 2 - 1: return else: if 2 * start + 2 >= size: _UpperCamelCase = 2 * start + 1 else: if heap[2 * start + 1] < heap[2 * start + 2]: _UpperCamelCase = 2 * start + 1 else: _UpperCamelCase = 2 * start + 2 if heap[smallest_child] < heap[start]: _UpperCamelCase , _UpperCamelCase = heap[smallest_child], positions[smallest_child] _UpperCamelCase , _UpperCamelCase = ( heap[start], positions[start], ) _UpperCamelCase , _UpperCamelCase = temp, tempa _UpperCamelCase = self.get_position(positions[smallest_child]) self.set_position( positions[smallest_child] , self.get_position(positions[start])) self.set_position(positions[start] , __a) self.top_to_bottom(__a , __a , __a , __a) def UpperCAmelCase ( self , __a , __a , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = position[index] while index != 0: _UpperCamelCase = int((index - 2) / 2) if index % 2 == 0 else int((index - 1) / 2) if val < heap[parent]: _UpperCamelCase = heap[parent] _UpperCamelCase = position[parent] self.set_position(position[parent] , __a) else: _UpperCamelCase = val _UpperCamelCase = temp self.set_position(__a , __a) break _UpperCamelCase = parent else: _UpperCamelCase = val _UpperCamelCase = temp self.set_position(__a , 0) def UpperCAmelCase ( self , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = len(__a) // 2 - 1 for i in range(__a , -1 , -1): self.top_to_bottom(__a , __a , len(__a) , __a) def UpperCAmelCase ( self , __a , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = positions[0] _UpperCamelCase = sys.maxsize self.top_to_bottom(__a , 0 , len(__a) , __a) return temp def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" _UpperCamelCase = Heap() _UpperCamelCase = [0] * len(__snake_case ) _UpperCamelCase = [-1] * len(__snake_case ) # Neighboring Tree Vertex of selected vertex # Minimum Distance of explored vertex with neighboring vertex of partial tree # formed in graph _UpperCamelCase = [] # Heap of Distance of vertices from their neighboring vertex _UpperCamelCase = [] for vertex in range(len(__snake_case ) ): distance_tv.append(sys.maxsize ) positions.append(__snake_case ) heap.node_position.append(__snake_case ) _UpperCamelCase = [] _UpperCamelCase = 1 _UpperCamelCase = sys.maxsize for neighbor, distance in adjacency_list[0]: _UpperCamelCase = 0 _UpperCamelCase = distance heap.heapify(__snake_case, __snake_case ) for _ in range(1, len(__snake_case ) ): _UpperCamelCase = heap.delete_minimum(__snake_case, __snake_case ) if visited[vertex] == 0: tree_edges.append((nbr_tv[vertex], vertex) ) _UpperCamelCase = 1 for neighbor, distance in adjacency_list[vertex]: if ( visited[neighbor] == 0 and distance < distance_tv[heap.get_position(__snake_case )] ): _UpperCamelCase = distance heap.bottom_to_top( __snake_case, heap.get_position(__snake_case ), __snake_case, __snake_case ) _UpperCamelCase = vertex return tree_edges if __name__ == "__main__": # pragma: no cover # < --------- Prims Algorithm --------- > _a = int(input("""Enter number of edges: """).strip()) _a = defaultdict(list) for _ in range(edges_number): _a = [int(x) for x in input().strip().split()] adjacency_list[edge[0]].append([edge[1], edge[2]]) adjacency_list[edge[1]].append([edge[0], edge[2]]) print(prisms_algorithm(adjacency_list))
78
"""simple docstring""" import argparse import torch from transformers import OpenAIGPTConfig, OpenAIGPTModel, load_tf_weights_in_openai_gpt from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> Union[str, Any]: """simple docstring""" if openai_config_file == "": _UpperCamelCase = OpenAIGPTConfig() else: _UpperCamelCase = OpenAIGPTConfig.from_json_file(__snake_case ) _UpperCamelCase = OpenAIGPTModel(__snake_case ) # Load weights from numpy load_tf_weights_in_openai_gpt(__snake_case, __snake_case, __snake_case ) # Save pytorch-model _UpperCamelCase = pytorch_dump_folder_path + '''/''' + WEIGHTS_NAME _UpperCamelCase = pytorch_dump_folder_path + '''/''' + CONFIG_NAME print(F'''Save PyTorch model to {pytorch_weights_dump_path}''' ) torch.save(model.state_dict(), __snake_case ) print(F'''Save configuration file to {pytorch_config_dump_path}''' ) with open(__snake_case, '''w''', encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( """--openai_checkpoint_folder_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--openai_config_file""", default="""""", type=str, help=( """An optional config json file corresponding to the pre-trained OpenAI model. \n""" """This specifies the model architecture.""" ), ) _a = parser.parse_args() convert_openai_checkpoint_to_pytorch( args.openai_checkpoint_folder_path, args.openai_config_file, args.pytorch_dump_folder_path )
78
1
"""simple docstring""" from .constants import ( MODEL_NAME, OPTIMIZER_NAME, RNG_STATE_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SCALER_NAME, SCHEDULER_NAME, TORCH_LAUNCH_PARAMS, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .dataclasses import ( BnbQuantizationConfig, ComputeEnvironment, CustomDtype, DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, DynamoBackend, FPaRecipeKwargs, FullyShardedDataParallelPlugin, GradientAccumulationPlugin, GradScalerKwargs, InitProcessGroupKwargs, KwargsHandler, LoggerType, MegatronLMPlugin, PrecisionType, ProjectConfiguration, RNGType, SageMakerDistributedType, TensorInformation, TorchDynamoPlugin, ) from .environment import get_int_from_env, parse_choice_from_env, parse_flag_from_env from .imports import ( get_ccl_version, is_abit_bnb_available, is_abit_bnb_available, is_aim_available, is_bfaa_available, is_bnb_available, is_botoa_available, is_ccl_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_fpa_available, is_ipex_available, is_megatron_lm_available, is_mlflow_available, is_mps_available, is_npu_available, is_rich_available, is_safetensors_available, is_sagemaker_available, is_tensorboard_available, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) from .modeling import ( check_device_map, check_tied_parameters_in_config, check_tied_parameters_on_same_device, compute_module_sizes, convert_file_size_to_int, dtype_byte_size, find_tied_parameters, get_balanced_memory, get_max_layer_size, get_max_memory, get_mixed_precision_context_manager, id_tensor_storage, infer_auto_device_map, load_checkpoint_in_model, load_offloaded_weights, load_state_dict, named_module_tensors, retie_parameters, set_module_tensor_to_device, shard_checkpoint, ) from .offload import ( OffloadedWeightsLoader, PrefixedDataset, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, save_offload_index, ) from .operations import ( broadcast, broadcast_object_list, concatenate, convert_outputs_to_fpaa, convert_to_fpaa, find_batch_size, find_device, gather, gather_object, get_data_structure, honor_type, initialize_tensors, is_namedtuple, is_tensor_information, is_torch_tensor, listify, pad_across_processes, recursively_apply, reduce, send_to_device, slice_tensors, ) from .versions import compare_versions, is_torch_version if is_deepspeed_available(): from .deepspeed import ( DeepSpeedEngineWrapper, DeepSpeedOptimizerWrapper, DeepSpeedSchedulerWrapper, DummyOptim, DummyScheduler, HfDeepSpeedConfig, ) from .bnb import has_abit_bnb_layers, load_and_quantize_model from .fsdp_utils import load_fsdp_model, load_fsdp_optimizer, save_fsdp_model, save_fsdp_optimizer from .launch import ( PrepareForLaunch, _filter_args, prepare_deepspeed_cmd_env, prepare_multi_gpu_env, prepare_sagemager_args_inputs, prepare_simple_launcher_cmd_env, prepare_tpu, ) from .megatron_lm import ( AbstractTrainStep, BertTrainStep, GPTTrainStep, MegatronEngine, MegatronLMDummyDataLoader, MegatronLMDummyScheduler, MegatronLMOptimizerWrapper, MegatronLMSchedulerWrapper, TaTrainStep, avg_losses_across_data_parallel_group, gather_across_data_parallel_groups, ) from .megatron_lm import initialize as megatron_lm_initialize from .megatron_lm import prepare_data_loader as megatron_lm_prepare_data_loader from .megatron_lm import prepare_model as megatron_lm_prepare_model from .megatron_lm import prepare_optimizer as megatron_lm_prepare_optimizer from .megatron_lm import prepare_scheduler as megatron_lm_prepare_scheduler from .memory import find_executable_batch_size, release_memory from .other import ( extract_model_from_parallel, get_pretty_name, is_port_in_use, merge_dicts, patch_environment, save, wait_for_everyone, write_basic_config, ) from .random import set_seed, synchronize_rng_state, synchronize_rng_states from .torch_xla import install_xla from .tqdm import tqdm from .transformer_engine import convert_model, has_transformer_engine_layers
78
"""simple docstring""" from __future__ import annotations import unittest from transformers import AutoTokenizer, MBartConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFMBartForConditionalGeneration, TFMBartModel @require_tf class _UpperCAmelCase: lowercase__ = MBartConfig lowercase__ = {} lowercase__ = 'gelu' def __init__( self , __a , __a=13 , __a=7 , __a=True , __a=False , __a=99 , __a=32 , __a=2 , __a=4 , __a=37 , __a=0.1 , __a=0.1 , __a=20 , __a=2 , __a=1 , __a=0 , ) -> Any: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = seq_length _UpperCamelCase = is_training _UpperCamelCase = use_labels _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = eos_token_id _UpperCamelCase = pad_token_id _UpperCamelCase = bos_token_id def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size) _UpperCamelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size) , 1) _UpperCamelCase = tf.concat([input_ids, eos_tensor] , axis=1) _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) _UpperCamelCase = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) _UpperCamelCase = prepare_mbart_inputs_dict(__a , __a , __a) return config, inputs_dict def UpperCAmelCase ( self , __a , __a) -> Optional[int]: '''simple docstring''' _UpperCamelCase = TFMBartModel(config=__a).get_decoder() _UpperCamelCase = inputs_dict['''input_ids'''] _UpperCamelCase = input_ids[:1, :] _UpperCamelCase = inputs_dict['''attention_mask'''][:1, :] _UpperCamelCase = inputs_dict['''head_mask'''] _UpperCamelCase = 1 # first forward pass _UpperCamelCase = model(__a , attention_mask=__a , head_mask=__a , use_cache=__a) _UpperCamelCase , _UpperCamelCase = outputs.to_tuple() _UpperCamelCase = past_key_values[1] def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case, __snake_case=None, __snake_case=None, __snake_case=None, __snake_case=None, __snake_case=None, ) -> Optional[int]: """simple docstring""" if attention_mask is None: _UpperCamelCase = tf.cast(tf.math.not_equal(__snake_case, config.pad_token_id ), tf.inta ) if decoder_attention_mask is None: _UpperCamelCase = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape, dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:], config.pad_token_id ), tf.inta ), ], axis=-1, ) if head_mask is None: _UpperCamelCase = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _UpperCamelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: _UpperCamelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class _UpperCAmelCase( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase__ = (TFMBartForConditionalGeneration, TFMBartModel) if is_tf_available() else () lowercase__ = (TFMBartForConditionalGeneration,) if is_tf_available() else () lowercase__ = ( { 'conversational': TFMBartForConditionalGeneration, 'feature-extraction': TFMBartModel, 'summarization': TFMBartForConditionalGeneration, 'text2text-generation': TFMBartForConditionalGeneration, 'translation': TFMBartForConditionalGeneration, } if is_tf_available() else {} ) lowercase__ = True lowercase__ = False lowercase__ = False def UpperCAmelCase ( self , __a , __a , __a , __a , __a) -> Dict: '''simple docstring''' if pipeline_test_casse_name != "FeatureExtractionPipelineTests": # Exception encountered when calling layer '...' return True return False def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = TFMBartModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=__a) def UpperCAmelCase ( self) -> str: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__a) @require_sentencepiece @require_tokenizers @require_tf class _UpperCAmelCase( unittest.TestCase ): lowercase__ = [ ' UN Chief Says There Is No Military Solution in Syria', ] lowercase__ = [ 'Şeful ONU declară că nu există o soluţie militară în Siria', ] lowercase__ = 'facebook/mbart-large-en-ro' @cached_property def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' return AutoTokenizer.from_pretrained(self.model_name) @cached_property def UpperCAmelCase ( self) -> str: '''simple docstring''' _UpperCamelCase = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name) return model def UpperCAmelCase ( self , **__a) -> List[str]: '''simple docstring''' _UpperCamelCase = self.translate_src_text(**__a) self.assertListEqual(self.expected_text , __a) def UpperCAmelCase ( self , **__a) -> Dict: '''simple docstring''' _UpperCamelCase = self.tokenizer(self.src_text , **__a , return_tensors='''tf''') _UpperCamelCase = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2) _UpperCamelCase = self.tokenizer.batch_decode(__a , skip_special_tokens=__a) return generated_words @slow def UpperCAmelCase ( self) -> Any: '''simple docstring''' self._assert_generated_batch_equal_expected()
78
1
"""simple docstring""" import contextlib import importlib import io import unittest import transformers # Try to import everything from transformers to ensure every object can be loaded. from transformers import * # noqa F406 from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, require_flax, require_tf, require_torch from transformers.utils import ContextManagers, find_labels, is_flax_available, is_tf_available, is_torch_available if is_torch_available(): from transformers import BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification if is_tf_available(): from transformers import TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification if is_flax_available(): from transformers import FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification _a = DUMMY_UNKNOWN_IDENTIFIER # An actual model hosted on huggingface.co _a = """main""" # Default branch name _a = """f2c752cfc5c0ab6f4bdec59acea69eefbee381c2""" # One particular commit (not the top of `main`) _a = """aaaaaaa""" # This commit does not exist, so we should 404. _a = """d9e9f15bc825e4b2c9249e9578f884bbcb5e3684""" # Sha-1 of config.json on the top of `main`, for checking purposes _a = """4b243c475af8d0a7754e87d7d096c92e5199ec2fe168a2ee7998e3b8e9bcb1d3""" @contextlib.contextmanager def lowerCamelCase__ ( ) -> str: """simple docstring""" print('''Welcome!''' ) yield print('''Bye!''' ) @contextlib.contextmanager def lowerCamelCase__ ( ) -> List[Any]: """simple docstring""" print('''Bonjour!''' ) yield print('''Au revoir!''' ) class _UpperCAmelCase( unittest.TestCase ): def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' # If the spec is missing, importlib would not be able to import the module dynamically. assert transformers.__spec__ is not None assert importlib.util.find_spec('''transformers''') is not None class _UpperCAmelCase( unittest.TestCase ): @unittest.mock.patch('''sys.stdout''' , new_callable=io.StringIO) def UpperCAmelCase ( self , __a) -> Dict: '''simple docstring''' with ContextManagers([]): print('''Transformers are awesome!''') # The print statement adds a new line at the end of the output self.assertEqual(mock_stdout.getvalue() , '''Transformers are awesome!\n''') @unittest.mock.patch('''sys.stdout''' , new_callable=io.StringIO) def UpperCAmelCase ( self , __a) -> Optional[Any]: '''simple docstring''' with ContextManagers([context_en()]): print('''Transformers are awesome!''') # The output should be wrapped with an English welcome and goodbye self.assertEqual(mock_stdout.getvalue() , '''Welcome!\nTransformers are awesome!\nBye!\n''') @unittest.mock.patch('''sys.stdout''' , new_callable=io.StringIO) def UpperCAmelCase ( self , __a) -> List[Any]: '''simple docstring''' with ContextManagers([context_fr(), context_en()]): print('''Transformers are awesome!''') # The output should be wrapped with an English and French welcome and goodbye self.assertEqual(mock_stdout.getvalue() , '''Bonjour!\nWelcome!\nTransformers are awesome!\nBye!\nAu revoir!\n''') @require_torch def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' self.assertEqual(find_labels(__a) , ['''labels''']) self.assertEqual(find_labels(__a) , ['''labels''', '''next_sentence_label''']) self.assertEqual(find_labels(__a) , ['''start_positions''', '''end_positions''']) class _UpperCAmelCase( lowerCamelCase ): pass self.assertEqual(find_labels(__a) , ['''labels''']) @require_tf def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' self.assertEqual(find_labels(__a) , ['''labels''']) self.assertEqual(find_labels(__a) , ['''labels''', '''next_sentence_label''']) self.assertEqual(find_labels(__a) , ['''start_positions''', '''end_positions''']) class _UpperCAmelCase( lowerCamelCase ): pass self.assertEqual(find_labels(__a) , ['''labels''']) @require_flax def UpperCAmelCase ( self) -> str: '''simple docstring''' # Flax models don't have labels self.assertEqual(find_labels(__a) , []) self.assertEqual(find_labels(__a) , []) self.assertEqual(find_labels(__a) , []) class _UpperCAmelCase( lowerCamelCase ): pass self.assertEqual(find_labels(__a) , [])
78
"""simple docstring""" from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _a = logging.get_logger(__name__) class _UpperCAmelCase( lowerCamelCase ): lowercase__ = ['pixel_values'] def __init__( self , __a = True , __a = 1 / 2_55 , __a = True , __a = 8 , **__a , ) -> None: '''simple docstring''' super().__init__(**__a) _UpperCamelCase = do_rescale _UpperCamelCase = rescale_factor _UpperCamelCase = do_pad _UpperCamelCase = pad_size def UpperCAmelCase ( self , __a , __a , __a = None , **__a) -> np.ndarray: '''simple docstring''' return rescale(__a , scale=__a , data_format=__a , **__a) def UpperCAmelCase ( self , __a , __a , __a = None) -> List[Any]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = get_image_size(__a) _UpperCamelCase = (old_height // size + 1) * size - old_height _UpperCamelCase = (old_width // size + 1) * size - old_width return pad(__a , ((0, pad_height), (0, pad_width)) , mode='''symmetric''' , data_format=__a) def UpperCAmelCase ( self , __a , __a = None , __a = None , __a = None , __a = None , __a = None , __a = ChannelDimension.FIRST , **__a , ) -> Tuple: '''simple docstring''' _UpperCamelCase = do_rescale if do_rescale is not None else self.do_rescale _UpperCamelCase = rescale_factor if rescale_factor is not None else self.rescale_factor _UpperCamelCase = do_pad if do_pad is not None else self.do_pad _UpperCamelCase = pad_size if pad_size is not None else self.pad_size _UpperCamelCase = make_list_of_images(__a) if not valid_images(__a): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''') if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''') # All transformations expect numpy arrays. _UpperCamelCase = [to_numpy_array(__a) for image in images] if do_rescale: _UpperCamelCase = [self.rescale(image=__a , scale=__a) for image in images] if do_pad: _UpperCamelCase = [self.pad(__a , size=__a) for image in images] _UpperCamelCase = [to_channel_dimension_format(__a , __a) for image in images] _UpperCamelCase = {'''pixel_values''': images} return BatchFeature(data=__a , tensor_type=__a)
78
1
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.utils import ComputeEnvironment from .cluster import get_cluster_input from .config_args import cache_dir, default_config_file, default_yaml_config_file, load_config_from_file # noqa: F401 from .config_utils import _ask_field, _ask_options, _convert_compute_environment # noqa: F401 from .sagemaker import get_sagemaker_input _a = """Launches a series of prompts to create and save a `default_config.yaml` configuration file for your training system. Should always be ran first on your machine""" def lowerCamelCase__ ( ) -> Any: """simple docstring""" _UpperCamelCase = _ask_options( '''In which compute environment are you running?''', ['''This machine''', '''AWS (Amazon SageMaker)'''], _convert_compute_environment, ) if compute_environment == ComputeEnvironment.AMAZON_SAGEMAKER: _UpperCamelCase = get_sagemaker_input() else: _UpperCamelCase = get_cluster_input() return config def lowerCamelCase__ ( __snake_case=None ) -> List[Any]: """simple docstring""" if subparsers is not None: _UpperCamelCase = subparsers.add_parser('''config''', description=__snake_case ) else: _UpperCamelCase = argparse.ArgumentParser('''Accelerate config command''', description=__snake_case ) parser.add_argument( '''--config_file''', default=__snake_case, help=( '''The path to use to store the config file. Will default to a file named default_config.yaml in the cache ''' '''location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ''' '''such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ''' '''with \'huggingface\'.''' ), ) if subparsers is not None: parser.set_defaults(func=__snake_case ) return parser def lowerCamelCase__ ( __snake_case ) -> Tuple: """simple docstring""" _UpperCamelCase = get_user_input() if args.config_file is not None: _UpperCamelCase = args.config_file else: if not os.path.isdir(__snake_case ): os.makedirs(__snake_case ) _UpperCamelCase = default_yaml_config_file if config_file.endswith('''.json''' ): config.to_json_file(__snake_case ) else: config.to_yaml_file(__snake_case ) print(F'''accelerate configuration saved at {config_file}''' ) def lowerCamelCase__ ( ) -> Optional[int]: """simple docstring""" _UpperCamelCase = config_command_parser() _UpperCamelCase = parser.parse_args() config_command(__snake_case ) if __name__ == "__main__": main()
78
"""simple docstring""" from importlib import import_module from .logging import get_logger _a = get_logger(__name__) class _UpperCAmelCase: def __init__( self , __a , __a=None) -> Dict: '''simple docstring''' _UpperCamelCase = attrs or [] if module is not None: for key in module.__dict__: if key in attrs or not key.startswith('''__'''): setattr(self , __a , getattr(__a , __a)) _UpperCamelCase = module._original_module if isinstance(__a , _PatchedModuleObj) else module class _UpperCAmelCase: lowercase__ = [] def __init__( self , __a , __a , __a , __a=None) -> List[str]: '''simple docstring''' _UpperCamelCase = obj _UpperCamelCase = target _UpperCamelCase = new _UpperCamelCase = target.split('''.''')[0] _UpperCamelCase = {} _UpperCamelCase = attrs or [] def __enter__( self) -> int: '''simple docstring''' *_UpperCamelCase , _UpperCamelCase = self.target.split('''.''') # Patch modules: # it's used to patch attributes of submodules like "os.path.join"; # in this case we need to patch "os" and "os.path" for i in range(len(__a)): try: _UpperCamelCase = import_module('''.'''.join(submodules[: i + 1])) except ModuleNotFoundError: continue # We iterate over all the globals in self.obj in case we find "os" or "os.path" for attr in self.obj.__dir__(): _UpperCamelCase = getattr(self.obj , __a) # We don't check for the name of the global, but rather if its value *is* "os" or "os.path". # This allows to patch renamed modules like "from os import path as ospath". if obj_attr is submodule or ( (isinstance(__a , _PatchedModuleObj) and obj_attr._original_module is submodule) ): _UpperCamelCase = obj_attr # patch at top level setattr(self.obj , __a , _PatchedModuleObj(__a , attrs=self.attrs)) _UpperCamelCase = getattr(self.obj , __a) # construct lower levels patches for key in submodules[i + 1 :]: setattr(__a , __a , _PatchedModuleObj(getattr(__a , __a , __a) , attrs=self.attrs)) _UpperCamelCase = getattr(__a , __a) # finally set the target attribute setattr(__a , __a , self.new) # Patch attribute itself: # it's used for builtins like "open", # and also to patch "os.path.join" we may also need to patch "join" # itself if it was imported as "from os.path import join". if submodules: # if it's an attribute of a submodule like "os.path.join" try: _UpperCamelCase = getattr(import_module('''.'''.join(__a)) , __a) except (AttributeError, ModuleNotFoundError): return # We iterate over all the globals in self.obj in case we find "os.path.join" for attr in self.obj.__dir__(): # We don't check for the name of the global, but rather if its value *is* "os.path.join". # This allows to patch renamed attributes like "from os.path import join as pjoin". if getattr(self.obj , __a) is attr_value: _UpperCamelCase = getattr(self.obj , __a) setattr(self.obj , __a , self.new) elif target_attr in globals()["__builtins__"]: # if it'a s builtin like "open" _UpperCamelCase = globals()['''__builtins__'''][target_attr] setattr(self.obj , __a , self.new) else: raise RuntimeError(F'''Tried to patch attribute {target_attr} instead of a submodule.''') def __exit__( self , *__a) -> Tuple: '''simple docstring''' for attr in list(self.original): setattr(self.obj , __a , self.original.pop(__a)) def UpperCAmelCase ( self) -> Dict: '''simple docstring''' self.__enter__() self._active_patches.append(self) def UpperCAmelCase ( self) -> str: '''simple docstring''' try: self._active_patches.remove(self) except ValueError: # If the patch hasn't been started this will fail return None return self.__exit__()
78
1
"""simple docstring""" import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import BatchEncoding, PreTrainedTokenizer from ...utils import logging _a = logging.get_logger(__name__) _a = """▁""" _a = { """vocab_file""": """vocab.json""", """spm_file""": """sentencepiece.bpe.model""", """tokenizer_config_file""": """tokenizer_config.json""", } _a = { """vocab_file""": { """facebook/m2m100_418M""": """https://huggingface.co/facebook/m2m100_418M/resolve/main/vocab.json""", """facebook/m2m100_1.2B""": """https://huggingface.co/facebook/m2m100_1.2B/resolve/main/vocab.json""", }, """spm_file""": { """facebook/m2m100_418M""": """https://huggingface.co/facebook/m2m100_418M/resolve/main/sentencepiece.bpe.model""", """facebook/m2m100_1.2B""": """https://huggingface.co/facebook/m2m100_1.2B/resolve/main/sentencepiece.bpe.model""", }, """tokenizer_config_file""": { """facebook/m2m100_418M""": """https://huggingface.co/facebook/m2m100_418M/resolve/main/tokenizer_config.json""", """facebook/m2m100_1.2B""": """https://huggingface.co/facebook/m2m100_1.2B/resolve/main/tokenizer_config.json""", }, } _a = { """facebook/m2m100_418M""": 1024, } # fmt: off _a = { """m2m100""": ["""af""", """am""", """ar""", """ast""", """az""", """ba""", """be""", """bg""", """bn""", """br""", """bs""", """ca""", """ceb""", """cs""", """cy""", """da""", """de""", """el""", """en""", """es""", """et""", """fa""", """ff""", """fi""", """fr""", """fy""", """ga""", """gd""", """gl""", """gu""", """ha""", """he""", """hi""", """hr""", """ht""", """hu""", """hy""", """id""", """ig""", """ilo""", """is""", """it""", """ja""", """jv""", """ka""", """kk""", """km""", """kn""", """ko""", """lb""", """lg""", """ln""", """lo""", """lt""", """lv""", """mg""", """mk""", """ml""", """mn""", """mr""", """ms""", """my""", """ne""", """nl""", """no""", """ns""", """oc""", """or""", """pa""", """pl""", """ps""", """pt""", """ro""", """ru""", """sd""", """si""", """sk""", """sl""", """so""", """sq""", """sr""", """ss""", """su""", """sv""", """sw""", """ta""", """th""", """tl""", """tn""", """tr""", """uk""", """ur""", """uz""", """vi""", """wo""", """xh""", """yi""", """yo""", """zh""", """zu"""], """wmt21""": ["""en""", """ha""", """is""", """ja""", """cs""", """ru""", """zh""", """de"""] } class _UpperCAmelCase( lowerCamelCase ): lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = ['input_ids', 'attention_mask'] lowercase__ = [] lowercase__ = [] def __init__( self , __a , __a , __a=None , __a=None , __a="<s>" , __a="</s>" , __a="</s>" , __a="<pad>" , __a="<unk>" , __a="m2m100" , __a = None , __a=8 , **__a , ) -> None: '''simple docstring''' _UpperCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs _UpperCamelCase = language_codes _UpperCamelCase = FAIRSEQ_LANGUAGE_CODES[language_codes] _UpperCamelCase = {lang_code: F'''__{lang_code}__''' for lang_code in fairseq_language_code} _UpperCamelCase = kwargs.get('''additional_special_tokens''' , []) kwargs["additional_special_tokens"] += [ self.get_lang_token(__a) for lang_code in fairseq_language_code if self.get_lang_token(__a) not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=__a , tgt_lang=__a , bos_token=__a , eos_token=__a , sep_token=__a , unk_token=__a , pad_token=__a , language_codes=__a , sp_model_kwargs=self.sp_model_kwargs , num_madeup_words=__a , **__a , ) _UpperCamelCase = vocab_file _UpperCamelCase = load_json(__a) _UpperCamelCase = {v: k for k, v in self.encoder.items()} _UpperCamelCase = spm_file _UpperCamelCase = load_spm(__a , self.sp_model_kwargs) _UpperCamelCase = len(self.encoder) _UpperCamelCase = { self.get_lang_token(__a): self.encoder_size + i for i, lang_code in enumerate(__a) } _UpperCamelCase = {lang_code: self.encoder_size + i for i, lang_code in enumerate(__a)} _UpperCamelCase = {v: k for k, v in self.lang_token_to_id.items()} _UpperCamelCase = src_lang if src_lang is not None else '''en''' _UpperCamelCase = tgt_lang _UpperCamelCase = self.get_lang_id(self._src_lang) self.set_src_lang_special_tokens(self._src_lang) _UpperCamelCase = num_madeup_words @property def UpperCAmelCase ( self) -> int: '''simple docstring''' return len(self.encoder) + len(self.lang_token_to_id) @property def UpperCAmelCase ( self) -> str: '''simple docstring''' return self._src_lang @src_lang.setter def UpperCAmelCase ( self , __a) -> None: '''simple docstring''' _UpperCamelCase = new_src_lang self.set_src_lang_special_tokens(self._src_lang) def UpperCAmelCase ( self , __a) -> List[str]: '''simple docstring''' return self.sp_model.encode(__a , out_type=__a) def UpperCAmelCase ( self , __a) -> Tuple: '''simple docstring''' if token in self.lang_token_to_id: return self.lang_token_to_id[token] return self.encoder.get(__a , self.encoder[self.unk_token]) def UpperCAmelCase ( self , __a) -> str: '''simple docstring''' if index in self.id_to_lang_token: return self.id_to_lang_token[index] return self.decoder.get(__a , self.unk_token) def UpperCAmelCase ( self , __a) -> Optional[int]: '''simple docstring''' _UpperCamelCase = [] _UpperCamelCase = '''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(__a) + token _UpperCamelCase = [] else: current_sub_tokens.append(__a) out_string += self.sp_model.decode(__a) return out_string.strip() def UpperCAmelCase ( self , __a , __a = None , __a = False) -> List[int]: '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__a , token_ids_a=__a , already_has_special_tokens=__a) _UpperCamelCase = [1] * len(self.prefix_tokens) _UpperCamelCase = [1] * len(self.suffix_tokens) if token_ids_a is None: return prefix_ones + ([0] * len(__a)) + suffix_ones return prefix_ones + ([0] * len(__a)) + ([0] * len(__a)) + suffix_ones def UpperCAmelCase ( self , __a , __a = None) -> List[int]: '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def UpperCAmelCase ( self) -> Dict: '''simple docstring''' _UpperCamelCase = {self.convert_ids_to_tokens(__a): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __getstate__( self) -> Dict: '''simple docstring''' _UpperCamelCase = self.__dict__.copy() _UpperCamelCase = None return state def __setstate__( self , __a) -> None: '''simple docstring''' _UpperCamelCase = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs'''): _UpperCamelCase = {} _UpperCamelCase = load_spm(self.spm_file , self.sp_model_kwargs) def UpperCAmelCase ( self , __a , __a = None) -> Tuple[str]: '''simple docstring''' _UpperCamelCase = Path(__a) if not save_dir.is_dir(): raise OSError(F'''{save_directory} should be a directory''') _UpperCamelCase = save_dir / ( (filename_prefix + '''-''' if filename_prefix else '''''') + self.vocab_files_names['''vocab_file'''] ) _UpperCamelCase = save_dir / ( (filename_prefix + '''-''' if filename_prefix else '''''') + self.vocab_files_names['''spm_file'''] ) save_json(self.encoder , __a) if os.path.abspath(self.spm_file) != os.path.abspath(__a) and os.path.isfile(self.spm_file): copyfile(self.spm_file , __a) elif not os.path.isfile(self.spm_file): with open(__a , '''wb''') as fi: _UpperCamelCase = self.sp_model.serialized_model_proto() fi.write(__a) return (str(__a), str(__a)) def UpperCAmelCase ( self , __a , __a = "en" , __a = None , __a = "ro" , **__a , ) -> BatchEncoding: '''simple docstring''' _UpperCamelCase = src_lang _UpperCamelCase = tgt_lang self.set_src_lang_special_tokens(self.src_lang) return super().prepare_seqaseq_batch(__a , __a , **__a) def UpperCAmelCase ( self , __a , __a , __a , **__a) -> int: '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError('''Translation requires a `src_lang` and a `tgt_lang` for this model''') _UpperCamelCase = src_lang _UpperCamelCase = self(__a , add_special_tokens=__a , **__a) _UpperCamelCase = self.get_lang_id(__a) _UpperCamelCase = tgt_lang_id return inputs def UpperCAmelCase ( self) -> int: '''simple docstring''' self.set_src_lang_special_tokens(self.src_lang) def UpperCAmelCase ( self) -> str: '''simple docstring''' self.set_tgt_lang_special_tokens(self.tgt_lang) def UpperCAmelCase ( self , __a) -> None: '''simple docstring''' _UpperCamelCase = self.get_lang_token(__a) _UpperCamelCase = self.lang_token_to_id[lang_token] _UpperCamelCase = [self.cur_lang_id] _UpperCamelCase = [self.eos_token_id] def UpperCAmelCase ( self , __a) -> None: '''simple docstring''' _UpperCamelCase = self.get_lang_token(__a) _UpperCamelCase = self.lang_token_to_id[lang_token] _UpperCamelCase = [self.cur_lang_id] _UpperCamelCase = [self.eos_token_id] def UpperCAmelCase ( self , __a) -> str: '''simple docstring''' return self.lang_code_to_token[lang] def UpperCAmelCase ( self , __a) -> int: '''simple docstring''' _UpperCamelCase = self.get_lang_token(__a) return self.lang_token_to_id[lang_token] def lowerCamelCase__ ( __snake_case, __snake_case ) -> sentencepiece.SentencePieceProcessor: """simple docstring""" _UpperCamelCase = sentencepiece.SentencePieceProcessor(**__snake_case ) spm.Load(str(__snake_case ) ) return spm def lowerCamelCase__ ( __snake_case ) -> Union[Dict, List]: """simple docstring""" with open(__snake_case, '''r''' ) as f: return json.load(__snake_case ) def lowerCamelCase__ ( __snake_case, __snake_case ) -> None: """simple docstring""" with open(__snake_case, '''w''' ) as f: json.dump(__snake_case, __snake_case, indent=2 )
78
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_flax_available, is_torch_available, is_transformers_available, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .multicontrolnet import MultiControlNetModel from .pipeline_controlnet import StableDiffusionControlNetPipeline from .pipeline_controlnet_imgaimg import StableDiffusionControlNetImgaImgPipeline from .pipeline_controlnet_inpaint import StableDiffusionControlNetInpaintPipeline if is_transformers_available() and is_flax_available(): from .pipeline_flax_controlnet import FlaxStableDiffusionControlNetPipeline
78
1
"""simple docstring""" import os import unittest from transformers import BatchEncoding from transformers.models.bert.tokenization_bert import ( BasicTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.models.prophetnet.tokenization_prophetnet import VOCAB_FILES_NAMES, ProphetNetTokenizer from transformers.testing_utils import require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin class _UpperCAmelCase( lowerCamelCase , unittest.TestCase ): lowercase__ = ProphetNetTokenizer lowercase__ = False def UpperCAmelCase ( self) -> Dict: '''simple docstring''' super().setUp() _UpperCamelCase = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] _UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file''']) with open(self.vocab_file , '''w''' , encoding='''utf-8''') as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens])) def UpperCAmelCase ( self , __a) -> Any: '''simple docstring''' _UpperCamelCase = '''UNwant\u00E9d,running''' _UpperCamelCase = '''unwanted, running''' return input_text, output_text def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.tokenizer_class(self.vocab_file) _UpperCamelCase = tokenizer.tokenize('''UNwant\u00E9d,running''') self.assertListEqual(__a , ['''un''', '''##want''', '''##ed''', ''',''', '''runn''', '''##ing''']) self.assertListEqual(tokenizer.convert_tokens_to_ids(__a) , [9, 6, 7, 12, 10, 11]) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = BasicTokenizer() self.assertListEqual(tokenizer.tokenize('''ah\u535A\u63A8zz''') , ['''ah''', '''\u535A''', '''\u63A8''', '''zz''']) def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = BasicTokenizer(do_lower_case=__a) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? ''') , ['''hello''', '''!''', '''how''', '''are''', '''you''', '''?''']) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''') , ['''hello''']) def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = BasicTokenizer(do_lower_case=__a , strip_accents=__a) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''') , ['''hällo''', '''!''', '''how''', '''are''', '''you''', '''?''']) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''') , ['''h\u00E9llo''']) def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = BasicTokenizer(do_lower_case=__a , strip_accents=__a) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''') , ['''hallo''', '''!''', '''how''', '''are''', '''you''', '''?''']) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''') , ['''hello''']) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = BasicTokenizer(do_lower_case=__a) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''') , ['''hallo''', '''!''', '''how''', '''are''', '''you''', '''?''']) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''') , ['''hello''']) def UpperCAmelCase ( self) -> str: '''simple docstring''' _UpperCamelCase = BasicTokenizer(do_lower_case=__a) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? ''') , ['''HeLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?''']) def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = BasicTokenizer(do_lower_case=__a , strip_accents=__a) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''') , ['''HäLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?''']) def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = BasicTokenizer(do_lower_case=__a , strip_accents=__a) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''') , ['''HaLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?''']) def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = BasicTokenizer(do_lower_case=__a , never_split=['''[UNK]''']) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? [UNK]''') , ['''HeLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?''', '''[UNK]''']) def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing'''] _UpperCamelCase = {} for i, token in enumerate(__a): _UpperCamelCase = i _UpperCamelCase = WordpieceTokenizer(vocab=__a , unk_token='''[UNK]''') self.assertListEqual(tokenizer.tokenize('''''') , []) self.assertListEqual(tokenizer.tokenize('''unwanted running''') , ['''un''', '''##want''', '''##ed''', '''runn''', '''##ing''']) self.assertListEqual(tokenizer.tokenize('''unwantedX running''') , ['''[UNK]''', '''runn''', '''##ing''']) @require_torch def UpperCAmelCase ( self) -> str: '''simple docstring''' _UpperCamelCase = self.tokenizer_class.from_pretrained('''microsoft/prophetnet-large-uncased''') _UpperCamelCase = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] _UpperCamelCase = [10_37, 21_46, 2_04_23, 20_05, 76_80, 78_49, 39_89, 10_12, 1_02] _UpperCamelCase = tokenizer(__a , padding=__a , return_tensors='''pt''') self.assertIsInstance(__a , __a) _UpperCamelCase = list(batch.input_ids.numpy()[0]) self.assertListEqual(__a , __a) self.assertEqual((2, 9) , batch.input_ids.shape) self.assertEqual((2, 9) , batch.attention_mask.shape) def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' self.assertTrue(_is_whitespace(''' ''')) self.assertTrue(_is_whitespace('''\t''')) self.assertTrue(_is_whitespace('''\r''')) self.assertTrue(_is_whitespace('''\n''')) self.assertTrue(_is_whitespace('''\u00A0''')) self.assertFalse(_is_whitespace('''A''')) self.assertFalse(_is_whitespace('''-''')) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' self.assertTrue(_is_control('''\u0005''')) self.assertFalse(_is_control('''A''')) self.assertFalse(_is_control(''' ''')) self.assertFalse(_is_control('''\t''')) self.assertFalse(_is_control('''\r''')) def UpperCAmelCase ( self) -> int: '''simple docstring''' self.assertTrue(_is_punctuation('''-''')) self.assertTrue(_is_punctuation('''$''')) self.assertTrue(_is_punctuation('''`''')) self.assertTrue(_is_punctuation('''.''')) self.assertFalse(_is_punctuation('''A''')) self.assertFalse(_is_punctuation(''' ''')) @slow def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = self.tokenizer_class.from_pretrained('''microsoft/prophetnet-large-uncased''') _UpperCamelCase = tokenizer.encode('''sequence builders''' , add_special_tokens=__a) _UpperCamelCase = tokenizer.encode('''multi-sequence build''' , add_special_tokens=__a) _UpperCamelCase = tokenizer.build_inputs_with_special_tokens(__a) _UpperCamelCase = tokenizer.build_inputs_with_special_tokens(__a , __a) assert encoded_sentence == text + [1_02] assert encoded_pair == text + [1_02] + text_a + [1_02]
78
"""simple docstring""" from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging _a = logging.get_logger(__name__) _a = { """EleutherAI/gpt-neo-1.3B""": """https://huggingface.co/EleutherAI/gpt-neo-1.3B/resolve/main/config.json""", # See all GPTNeo models at https://huggingface.co/models?filter=gpt_neo } class _UpperCAmelCase( lowerCamelCase ): lowercase__ = 'gpt_neo' lowercase__ = ['past_key_values'] lowercase__ = {'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers'} def __init__( self , __a=5_02_57 , __a=20_48 , __a=20_48 , __a=24 , __a=[[["global", "local"], 12]] , __a=16 , __a=None , __a=2_56 , __a="gelu_new" , __a=0.0 , __a=0.0 , __a=0.0 , __a=0.1 , __a=1e-5 , __a=0.02 , __a=True , __a=5_02_56 , __a=5_02_56 , **__a , ) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = vocab_size _UpperCamelCase = max_position_embeddings _UpperCamelCase = hidden_size _UpperCamelCase = num_layers _UpperCamelCase = num_heads _UpperCamelCase = intermediate_size _UpperCamelCase = window_size _UpperCamelCase = activation_function _UpperCamelCase = resid_dropout _UpperCamelCase = embed_dropout _UpperCamelCase = attention_dropout _UpperCamelCase = classifier_dropout _UpperCamelCase = layer_norm_epsilon _UpperCamelCase = initializer_range _UpperCamelCase = use_cache _UpperCamelCase = bos_token_id _UpperCamelCase = eos_token_id _UpperCamelCase = attention_types _UpperCamelCase = self.expand_attention_types_params(__a) if len(self.attention_layers) != self.num_layers: raise ValueError( '''Configuration for convolutional module is incorrect. ''' '''It is required that `len(config.attention_layers)` == `config.num_layers` ''' F'''but is `len(config.attention_layers) = {len(self.attention_layers)}`, ''' F'''`config.num_layers = {self.num_layers}`. ''' '''`config.attention_layers` is prepared using `config.attention_types`. ''' '''Please verify the value of `config.attention_types` argument.''') super().__init__(bos_token_id=__a , eos_token_id=__a , **__a) @staticmethod def UpperCAmelCase ( __a) -> int: '''simple docstring''' _UpperCamelCase = [] for item in attention_types: for _ in range(item[1]): attentions.extend(item[0]) return attentions def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case, __snake_case ) -> str: """simple docstring""" import torch _UpperCamelCase = input.size() _UpperCamelCase = len(__snake_case ) _UpperCamelCase = shape[dimension] _UpperCamelCase = torch.arange(0, __snake_case, __snake_case ) _UpperCamelCase = torch.div(sizedim - size, __snake_case, rounding_mode='''floor''' ) + 1 _UpperCamelCase = torch.arange(__snake_case ) + low_indices[:min_length][:, None] _UpperCamelCase = [slice(__snake_case )] * rank _UpperCamelCase = indices _UpperCamelCase = input[s] _UpperCamelCase = list(range(0, rank + 1 ) ) perm.append(perm.pop(dimension + 1 ) ) return sliced.permute(__snake_case ) def lowerCamelCase__ ( __snake_case, __snake_case ) -> str: """simple docstring""" import torch _UpperCamelCase = torch.arange(1, __snake_case ) _UpperCamelCase = torch.remainder(__snake_case, __snake_case ) _UpperCamelCase = remainders == 0 _UpperCamelCase = candidates[divisor_indices] _UpperCamelCase = torch.max(__snake_case ) return largest_divisor, torch.div(__snake_case, __snake_case, rounding_mode='''floor''' ) class _UpperCAmelCase( lowerCamelCase ): @property def UpperCAmelCase ( self) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' _UpperCamelCase = OrderedDict({'''input_ids''': {0: '''batch''', 1: '''sequence'''}}) if self.use_past: self.fill_with_past_key_values_(__a , direction='''inputs''') _UpperCamelCase = {0: '''batch''', 1: '''past_sequence + sequence'''} else: _UpperCamelCase = {0: '''batch''', 1: '''sequence'''} return common_inputs @property def UpperCAmelCase ( self) -> int: '''simple docstring''' return self._config.num_heads def UpperCAmelCase ( self , __a , __a = -1 , __a = -1 , __a = False , __a = None , ) -> Mapping[str, Any]: '''simple docstring''' _UpperCamelCase = super(__a , self).generate_dummy_inputs( __a , batch_size=__a , seq_length=__a , is_pair=__a , framework=__a) # We need to order the input in the way they appears in the forward() _UpperCamelCase = OrderedDict({'''input_ids''': common_inputs['''input_ids''']}) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''') else: import torch _UpperCamelCase , _UpperCamelCase = common_inputs['''input_ids'''].shape # Not using the same length for past_key_values _UpperCamelCase = seqlen + 2 _UpperCamelCase = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) _UpperCamelCase = [ (torch.zeros(__a), torch.zeros(__a)) for _ in range(self.num_layers) ] _UpperCamelCase = common_inputs['''attention_mask'''] if self.use_past: _UpperCamelCase = ordered_inputs['''attention_mask'''].dtype _UpperCamelCase = torch.cat( [ordered_inputs['''attention_mask'''], torch.ones(__a , __a , dtype=__a)] , dim=1) return ordered_inputs @property def UpperCAmelCase ( self) -> int: '''simple docstring''' return 13
78
1
"""simple docstring""" from PIL import Image def lowerCamelCase__ ( __snake_case ) -> Image: """simple docstring""" _UpperCamelCase , _UpperCamelCase = image.size _UpperCamelCase = 0 _UpperCamelCase = image.load() for i in range(__snake_case ): for j in range(__snake_case ): _UpperCamelCase = pixels[j, i] mean += pixel mean //= width * height for j in range(__snake_case ): for i in range(__snake_case ): _UpperCamelCase = 2_55 if pixels[i, j] > mean else 0 return image if __name__ == "__main__": _a = mean_threshold(Image.open("""path_to_image""").convert("""L""")) image.save("""output_image_path""")
78
"""simple docstring""" import sys from collections import defaultdict class _UpperCAmelCase: def __init__( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = [] def UpperCAmelCase ( self , __a) -> Optional[Any]: '''simple docstring''' return self.node_position[vertex] def UpperCAmelCase ( self , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = pos def UpperCAmelCase ( self , __a , __a , __a , __a) -> Tuple: '''simple docstring''' if start > size // 2 - 1: return else: if 2 * start + 2 >= size: _UpperCamelCase = 2 * start + 1 else: if heap[2 * start + 1] < heap[2 * start + 2]: _UpperCamelCase = 2 * start + 1 else: _UpperCamelCase = 2 * start + 2 if heap[smallest_child] < heap[start]: _UpperCamelCase , _UpperCamelCase = heap[smallest_child], positions[smallest_child] _UpperCamelCase , _UpperCamelCase = ( heap[start], positions[start], ) _UpperCamelCase , _UpperCamelCase = temp, tempa _UpperCamelCase = self.get_position(positions[smallest_child]) self.set_position( positions[smallest_child] , self.get_position(positions[start])) self.set_position(positions[start] , __a) self.top_to_bottom(__a , __a , __a , __a) def UpperCAmelCase ( self , __a , __a , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = position[index] while index != 0: _UpperCamelCase = int((index - 2) / 2) if index % 2 == 0 else int((index - 1) / 2) if val < heap[parent]: _UpperCamelCase = heap[parent] _UpperCamelCase = position[parent] self.set_position(position[parent] , __a) else: _UpperCamelCase = val _UpperCamelCase = temp self.set_position(__a , __a) break _UpperCamelCase = parent else: _UpperCamelCase = val _UpperCamelCase = temp self.set_position(__a , 0) def UpperCAmelCase ( self , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = len(__a) // 2 - 1 for i in range(__a , -1 , -1): self.top_to_bottom(__a , __a , len(__a) , __a) def UpperCAmelCase ( self , __a , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = positions[0] _UpperCamelCase = sys.maxsize self.top_to_bottom(__a , 0 , len(__a) , __a) return temp def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" _UpperCamelCase = Heap() _UpperCamelCase = [0] * len(__snake_case ) _UpperCamelCase = [-1] * len(__snake_case ) # Neighboring Tree Vertex of selected vertex # Minimum Distance of explored vertex with neighboring vertex of partial tree # formed in graph _UpperCamelCase = [] # Heap of Distance of vertices from their neighboring vertex _UpperCamelCase = [] for vertex in range(len(__snake_case ) ): distance_tv.append(sys.maxsize ) positions.append(__snake_case ) heap.node_position.append(__snake_case ) _UpperCamelCase = [] _UpperCamelCase = 1 _UpperCamelCase = sys.maxsize for neighbor, distance in adjacency_list[0]: _UpperCamelCase = 0 _UpperCamelCase = distance heap.heapify(__snake_case, __snake_case ) for _ in range(1, len(__snake_case ) ): _UpperCamelCase = heap.delete_minimum(__snake_case, __snake_case ) if visited[vertex] == 0: tree_edges.append((nbr_tv[vertex], vertex) ) _UpperCamelCase = 1 for neighbor, distance in adjacency_list[vertex]: if ( visited[neighbor] == 0 and distance < distance_tv[heap.get_position(__snake_case )] ): _UpperCamelCase = distance heap.bottom_to_top( __snake_case, heap.get_position(__snake_case ), __snake_case, __snake_case ) _UpperCamelCase = vertex return tree_edges if __name__ == "__main__": # pragma: no cover # < --------- Prims Algorithm --------- > _a = int(input("""Enter number of edges: """).strip()) _a = defaultdict(list) for _ in range(edges_number): _a = [int(x) for x in input().strip().split()] adjacency_list[edge[0]].append([edge[1], edge[2]]) adjacency_list[edge[1]].append([edge[0], edge[2]]) print(prisms_algorithm(adjacency_list))
78
1
"""simple docstring""" import numpy as np class _UpperCAmelCase: def __init__( self) -> int: '''simple docstring''' _UpperCamelCase = (0, 0) _UpperCamelCase = None _UpperCamelCase = 0 _UpperCamelCase = 0 _UpperCamelCase = 0 def __eq__( self , __a) -> Dict: '''simple docstring''' return self.position == cell.position def UpperCAmelCase ( self) -> str: '''simple docstring''' print(self.position) class _UpperCAmelCase: def __init__( self , __a=(5, 5)) -> Dict: '''simple docstring''' _UpperCamelCase = np.zeros(__a) _UpperCamelCase = world_size[0] _UpperCamelCase = world_size[1] def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' print(self.w) def UpperCAmelCase ( self , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = [ (-1, -1), (-1, 0), (-1, 1), (0, -1), (0, 1), (1, -1), (1, 0), (1, 1), ] _UpperCamelCase = cell.position[0] _UpperCamelCase = cell.position[1] _UpperCamelCase = [] for n in neughbour_cord: _UpperCamelCase = current_x + n[0] _UpperCamelCase = current_y + n[1] if 0 <= x < self.world_x_limit and 0 <= y < self.world_y_limit: _UpperCamelCase = Cell() _UpperCamelCase = (x, y) _UpperCamelCase = cell neighbours.append(__a) return neighbours def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> List[Any]: """simple docstring""" _UpperCamelCase = [] _UpperCamelCase = [] _open.append(__snake_case ) while _open: _UpperCamelCase = np.argmin([n.f for n in _open] ) _UpperCamelCase = _open[min_f] _closed.append(_open.pop(__snake_case ) ) if current == goal: break for n in world.get_neigbours(__snake_case ): for c in _closed: if c == n: continue _UpperCamelCase = current.g + 1 _UpperCamelCase , _UpperCamelCase = n.position _UpperCamelCase , _UpperCamelCase = goal.position _UpperCamelCase = (ya - ya) ** 2 + (xa - xa) ** 2 _UpperCamelCase = n.h + n.g for c in _open: if c == n and c.f < n.f: continue _open.append(__snake_case ) _UpperCamelCase = [] while current.parent is not None: path.append(current.position ) _UpperCamelCase = current.parent path.append(current.position ) return path[::-1] if __name__ == "__main__": _a = Gridworld() # Start position and goal _a = Cell() _a = (0, 0) _a = Cell() _a = (4, 4) print(F"""path from {start.position} to {goal.position}""") _a = astar(world, start, goal) # Just for visual reasons. for i in s: _a = 1 print(world.w)
78
"""simple docstring""" import json import sys def lowerCamelCase__ ( __snake_case, __snake_case ) -> Union[str, Any]: """simple docstring""" with open(__snake_case, encoding='''utf-8''' ) as f: _UpperCamelCase = json.load(__snake_case ) _UpperCamelCase = ['''<details>''', '''<summary>Show updated benchmarks!</summary>''', ''' '''] for benchmark_name in sorted(__snake_case ): _UpperCamelCase = results[benchmark_name] _UpperCamelCase = benchmark_name.split('''/''' )[-1] output_md.append(F'''### Benchmark: {benchmark_file_name}''' ) _UpperCamelCase = '''| metric |''' _UpperCamelCase = '''|--------|''' _UpperCamelCase = '''| new / old (diff) |''' for metric_name in sorted(__snake_case ): _UpperCamelCase = benchmark_res[metric_name] _UpperCamelCase = metric_vals['''new'''] _UpperCamelCase = metric_vals.get('''old''', __snake_case ) _UpperCamelCase = metric_vals.get('''diff''', __snake_case ) _UpperCamelCase = F''' {new_val:f}''' if isinstance(__snake_case, (int, float) ) else '''None''' if old_val is not None: val_str += F''' / {old_val:f}''' if isinstance(__snake_case, (int, float) ) else "None" if dif_val is not None: val_str += F''' ({dif_val:f})''' if isinstance(__snake_case, (int, float) ) else "None" title += " " + metric_name + " |" lines += "---|" value += val_str + " |" output_md += [title, lines, value, " "] output_md.append('''</details>''' ) with open(__snake_case, '''w''', encoding='''utf-8''' ) as f: f.writelines('''\n'''.join(__snake_case ) ) if __name__ == "__main__": _a = sys.argv[1] _a = sys.argv[2] format_json_to_md(input_json_file, output_md_file)
78
1
"""simple docstring""" import logging from transformers import PretrainedConfig _a = logging.getLogger(__name__) _a = { """bertabs-finetuned-cnndm""": """https://huggingface.co/remi/bertabs-finetuned-cnndm-extractive-abstractive-summarization/resolve/main/config.json""", } class _UpperCAmelCase( lowerCamelCase ): lowercase__ = 'bertabs' def __init__( self , __a=3_05_22 , __a=5_12 , __a=6 , __a=5_12 , __a=8 , __a=5_12 , __a=0.2 , __a=6 , __a=7_68 , __a=8 , __a=20_48 , __a=0.2 , **__a , ) -> Any: '''simple docstring''' super().__init__(**__a) _UpperCamelCase = vocab_size _UpperCamelCase = max_pos _UpperCamelCase = enc_layers _UpperCamelCase = enc_hidden_size _UpperCamelCase = enc_heads _UpperCamelCase = enc_ff_size _UpperCamelCase = enc_dropout _UpperCamelCase = dec_layers _UpperCamelCase = dec_hidden_size _UpperCamelCase = dec_heads _UpperCamelCase = dec_ff_size _UpperCamelCase = dec_dropout
78
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTImageProcessor, ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) def lowerCamelCase__ ( __snake_case, __snake_case=False ) -> Tuple: """simple docstring""" _UpperCamelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'''blocks.{i}.norm1.weight''', F'''vit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((F'''blocks.{i}.norm1.bias''', F'''vit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((F'''blocks.{i}.attn.proj.weight''', F'''vit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.attn.proj.bias''', F'''vit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((F'''blocks.{i}.norm2.weight''', F'''vit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((F'''blocks.{i}.norm2.bias''', F'''vit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.weight''', F'''vit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.bias''', F'''vit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.weight''', F'''vit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.bias''', F'''vit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ('''cls_token''', '''vit.embeddings.cls_token'''), ('''patch_embed.proj.weight''', '''vit.embeddings.patch_embeddings.projection.weight'''), ('''patch_embed.proj.bias''', '''vit.embeddings.patch_embeddings.projection.bias'''), ('''pos_embed''', '''vit.embeddings.position_embeddings'''), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ('''pre_logits.fc.weight''', '''pooler.dense.weight'''), ('''pre_logits.fc.bias''', '''pooler.dense.bias'''), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _UpperCamelCase = [(pair[0], pair[1][4:]) if pair[1].startswith('''vit''' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('''norm.weight''', '''vit.layernorm.weight'''), ('''norm.bias''', '''vit.layernorm.bias'''), ('''head.weight''', '''classifier.weight'''), ('''head.bias''', '''classifier.bias'''), ] ) return rename_keys def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case=False ) -> str: """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: _UpperCamelCase = '''''' else: _UpperCamelCase = '''vit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _UpperCamelCase = state_dict.pop(F'''blocks.{i}.attn.qkv.weight''' ) _UpperCamelCase = state_dict.pop(F'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict _UpperCamelCase = in_proj_weight[ : config.hidden_size, : ] _UpperCamelCase = in_proj_bias[: config.hidden_size] _UpperCamelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _UpperCamelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _UpperCamelCase = in_proj_weight[ -config.hidden_size :, : ] _UpperCamelCase = in_proj_bias[-config.hidden_size :] def lowerCamelCase__ ( __snake_case ) -> Dict: """simple docstring""" _UpperCamelCase = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(__snake_case, __snake_case ) def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> Any: """simple docstring""" _UpperCamelCase = dct.pop(__snake_case ) _UpperCamelCase = val def lowerCamelCase__ ( ) -> Dict: """simple docstring""" _UpperCamelCase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _UpperCamelCase = Image.open(requests.get(__snake_case, stream=__snake_case ).raw ) return im @torch.no_grad() def lowerCamelCase__ ( __snake_case, __snake_case ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = ViTConfig() _UpperCamelCase = False # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size if vit_name[-5:] == "in21k": _UpperCamelCase = True _UpperCamelCase = int(vit_name[-12:-10] ) _UpperCamelCase = int(vit_name[-9:-6] ) else: _UpperCamelCase = 10_00 _UpperCamelCase = '''huggingface/label-files''' _UpperCamelCase = '''imagenet-1k-id2label.json''' _UpperCamelCase = json.load(open(hf_hub_download(__snake_case, __snake_case, repo_type='''dataset''' ), '''r''' ) ) _UpperCamelCase = {int(__snake_case ): v for k, v in idalabel.items()} _UpperCamelCase = idalabel _UpperCamelCase = {v: k for k, v in idalabel.items()} _UpperCamelCase = int(vit_name[-6:-4] ) _UpperCamelCase = int(vit_name[-3:] ) # size of the architecture if "deit" in vit_name: if vit_name[9:].startswith('''tiny''' ): _UpperCamelCase = 1_92 _UpperCamelCase = 7_68 _UpperCamelCase = 12 _UpperCamelCase = 3 elif vit_name[9:].startswith('''small''' ): _UpperCamelCase = 3_84 _UpperCamelCase = 15_36 _UpperCamelCase = 12 _UpperCamelCase = 6 else: pass else: if vit_name[4:].startswith('''small''' ): _UpperCamelCase = 7_68 _UpperCamelCase = 23_04 _UpperCamelCase = 8 _UpperCamelCase = 8 elif vit_name[4:].startswith('''base''' ): pass elif vit_name[4:].startswith('''large''' ): _UpperCamelCase = 10_24 _UpperCamelCase = 40_96 _UpperCamelCase = 24 _UpperCamelCase = 16 elif vit_name[4:].startswith('''huge''' ): _UpperCamelCase = 12_80 _UpperCamelCase = 51_20 _UpperCamelCase = 32 _UpperCamelCase = 16 # load original model from timm _UpperCamelCase = timm.create_model(__snake_case, pretrained=__snake_case ) timm_model.eval() # load state_dict of original model, remove and rename some keys _UpperCamelCase = timm_model.state_dict() if base_model: remove_classification_head_(__snake_case ) _UpperCamelCase = create_rename_keys(__snake_case, __snake_case ) for src, dest in rename_keys: rename_key(__snake_case, __snake_case, __snake_case ) read_in_q_k_v(__snake_case, __snake_case, __snake_case ) # load HuggingFace model if vit_name[-5:] == "in21k": _UpperCamelCase = ViTModel(__snake_case ).eval() else: _UpperCamelCase = ViTForImageClassification(__snake_case ).eval() model.load_state_dict(__snake_case ) # Check outputs on an image, prepared by ViTImageProcessor/DeiTImageProcessor if "deit" in vit_name: _UpperCamelCase = DeiTImageProcessor(size=config.image_size ) else: _UpperCamelCase = ViTImageProcessor(size=config.image_size ) _UpperCamelCase = image_processor(images=prepare_img(), return_tensors='''pt''' ) _UpperCamelCase = encoding['''pixel_values'''] _UpperCamelCase = model(__snake_case ) if base_model: _UpperCamelCase = timm_model.forward_features(__snake_case ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(__snake_case, outputs.pooler_output, atol=1e-3 ) else: _UpperCamelCase = timm_model(__snake_case ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__snake_case, outputs.logits, atol=1e-3 ) Path(__snake_case ).mkdir(exist_ok=__snake_case ) print(F'''Saving model {vit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(__snake_case ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(__snake_case ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( """--vit_name""", default="""vit_base_patch16_224""", type=str, help="""Name of the ViT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) _a = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path)
78
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _a = logging.get_logger(__name__) _a = { """hustvl/yolos-small""": """https://huggingface.co/hustvl/yolos-small/resolve/main/config.json""", # See all YOLOS models at https://huggingface.co/models?filter=yolos } class _UpperCAmelCase( lowerCamelCase ): lowercase__ = 'yolos' def __init__( self , __a=7_68 , __a=12 , __a=12 , __a=30_72 , __a="gelu" , __a=0.0 , __a=0.0 , __a=0.02 , __a=1e-12 , __a=[5_12, 8_64] , __a=16 , __a=3 , __a=True , __a=1_00 , __a=True , __a=False , __a=1 , __a=5 , __a=2 , __a=5 , __a=2 , __a=0.1 , **__a , ) -> Any: '''simple docstring''' super().__init__(**__a) _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = initializer_range _UpperCamelCase = layer_norm_eps _UpperCamelCase = image_size _UpperCamelCase = patch_size _UpperCamelCase = num_channels _UpperCamelCase = qkv_bias _UpperCamelCase = num_detection_tokens _UpperCamelCase = use_mid_position_embeddings _UpperCamelCase = auxiliary_loss # Hungarian matcher _UpperCamelCase = class_cost _UpperCamelCase = bbox_cost _UpperCamelCase = giou_cost # Loss coefficients _UpperCamelCase = bbox_loss_coefficient _UpperCamelCase = giou_loss_coefficient _UpperCamelCase = eos_coefficient class _UpperCAmelCase( lowerCamelCase ): lowercase__ = version.parse('1.11' ) @property def UpperCAmelCase ( self) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ]) @property def UpperCAmelCase ( self) -> float: '''simple docstring''' return 1e-4 @property def UpperCAmelCase ( self) -> int: '''simple docstring''' return 12
78
"""simple docstring""" import unittest from transformers import AlbertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForPreTraining, AlbertForQuestionAnswering, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertModel, ) from transformers.models.albert.modeling_albert import ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST class _UpperCAmelCase: def __init__( self , __a , __a=13 , __a=7 , __a=True , __a=True , __a=True , __a=True , __a=99 , __a=16 , __a=36 , __a=6 , __a=6 , __a=6 , __a=37 , __a="gelu" , __a=0.1 , __a=0.1 , __a=5_12 , __a=16 , __a=2 , __a=0.02 , __a=3 , __a=4 , __a=None , ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = seq_length _UpperCamelCase = is_training _UpperCamelCase = use_input_mask _UpperCamelCase = use_token_type_ids _UpperCamelCase = use_labels _UpperCamelCase = vocab_size _UpperCamelCase = embedding_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_hidden_groups _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = type_vocab_size _UpperCamelCase = type_sequence_label_size _UpperCamelCase = initializer_range _UpperCamelCase = num_labels _UpperCamelCase = num_choices _UpperCamelCase = scope def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) _UpperCamelCase = None if self.use_input_mask: _UpperCamelCase = random_attention_mask([self.batch_size, self.seq_length]) _UpperCamelCase = None if self.use_token_type_ids: _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size) _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) _UpperCamelCase = ids_tensor([self.batch_size] , self.num_choices) _UpperCamelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase ( self) -> Dict: '''simple docstring''' return AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , num_hidden_groups=self.num_hidden_groups , ) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Optional[int]: '''simple docstring''' _UpperCamelCase = AlbertModel(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , token_type_ids=__a) _UpperCamelCase = model(__a , token_type_ids=__a) _UpperCamelCase = model(__a) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = AlbertForPreTraining(config=__a) model.to(__a) model.eval() _UpperCamelCase = model( __a , attention_mask=__a , token_type_ids=__a , labels=__a , sentence_order_label=__a , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) self.parent.assertEqual(result.sop_logits.shape , (self.batch_size, config.num_labels)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Optional[int]: '''simple docstring''' _UpperCamelCase = AlbertForMaskedLM(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Any: '''simple docstring''' _UpperCamelCase = AlbertForQuestionAnswering(config=__a) model.to(__a) model.eval() _UpperCamelCase = model( __a , attention_mask=__a , token_type_ids=__a , start_positions=__a , end_positions=__a , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> List[Any]: '''simple docstring''' _UpperCamelCase = self.num_labels _UpperCamelCase = AlbertForSequenceClassification(__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> List[str]: '''simple docstring''' _UpperCamelCase = self.num_labels _UpperCamelCase = AlbertForTokenClassification(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.num_choices _UpperCamelCase = AlbertForMultipleChoice(config=__a) model.to(__a) model.eval() _UpperCamelCase = input_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() _UpperCamelCase = token_type_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() _UpperCamelCase = input_mask.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() _UpperCamelCase = model( __a , attention_mask=__a , token_type_ids=__a , labels=__a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices)) def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.prepare_config_and_inputs() ( ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ) = config_and_inputs _UpperCamelCase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class _UpperCAmelCase( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase__ = ( ( AlbertModel, AlbertForPreTraining, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertForQuestionAnswering, ) if is_torch_available() else () ) lowercase__ = ( { 'feature-extraction': AlbertModel, 'fill-mask': AlbertForMaskedLM, 'question-answering': AlbertForQuestionAnswering, 'text-classification': AlbertForSequenceClassification, 'token-classification': AlbertForTokenClassification, 'zero-shot': AlbertForSequenceClassification, } if is_torch_available() else {} ) lowercase__ = True def UpperCAmelCase ( self , __a , __a , __a=False) -> List[str]: '''simple docstring''' _UpperCamelCase = super()._prepare_for_class(__a , __a , return_labels=__a) if return_labels: if model_class in get_values(__a): _UpperCamelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=__a) _UpperCamelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__a) return inputs_dict def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = AlbertModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=__a , hidden_size=37) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*__a) def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__a) def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__a) def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__a) def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__a) def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _UpperCamelCase = type self.model_tester.create_and_check_model(*__a) @slow def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' for model_name in ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase = AlbertModel.from_pretrained(__a) self.assertIsNotNone(__a) @require_torch class _UpperCAmelCase( unittest.TestCase ): @slow def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = AlbertModel.from_pretrained('''albert-base-v2''') _UpperCamelCase = torch.tensor([[0, 3_45, 2_32, 3_28, 7_40, 1_40, 16_95, 69, 60_78, 15_88, 2]]) _UpperCamelCase = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]) with torch.no_grad(): _UpperCamelCase = model(__a , attention_mask=__a)[0] _UpperCamelCase = torch.Size((1, 11, 7_68)) self.assertEqual(output.shape , __a) _UpperCamelCase = torch.tensor( [[[-0.6513, 1.5035, -0.2766], [-0.6515, 1.5046, -0.2780], [-0.6512, 1.5049, -0.2784]]]) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , __a , atol=1e-4))
78
1
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import ( BitConfig, ViTHybridConfig, ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel, ) from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) def lowerCamelCase__ ( __snake_case, __snake_case=False ) -> Any: """simple docstring""" _UpperCamelCase = [] # fmt: off # stem: rename_keys.append(('''cls_token''', '''vit.embeddings.cls_token''') ) rename_keys.append(('''pos_embed''', '''vit.embeddings.position_embeddings''') ) rename_keys.append(('''patch_embed.proj.weight''', '''vit.embeddings.patch_embeddings.projection.weight''') ) rename_keys.append(('''patch_embed.proj.bias''', '''vit.embeddings.patch_embeddings.projection.bias''') ) # backbone rename_keys.append(('''patch_embed.backbone.stem.conv.weight''', '''vit.embeddings.patch_embeddings.backbone.bit.embedder.convolution.weight''') ) rename_keys.append(('''patch_embed.backbone.stem.norm.weight''', '''vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.weight''') ) rename_keys.append(('''patch_embed.backbone.stem.norm.bias''', '''vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.bias''') ) for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv1.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv1.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.bias''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv2.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv2.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.bias''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv3.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv3.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.bias''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.conv.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.conv.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.bias''') ) # transformer encoder for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'''blocks.{i}.norm1.weight''', F'''vit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((F'''blocks.{i}.norm1.bias''', F'''vit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((F'''blocks.{i}.attn.proj.weight''', F'''vit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.attn.proj.bias''', F'''vit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((F'''blocks.{i}.norm2.weight''', F'''vit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((F'''blocks.{i}.norm2.bias''', F'''vit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.weight''', F'''vit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.bias''', F'''vit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.weight''', F'''vit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.bias''', F'''vit.encoder.layer.{i}.output.dense.bias''') ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ('''pre_logits.fc.weight''', '''pooler.dense.weight'''), ('''pre_logits.fc.bias''', '''pooler.dense.bias'''), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _UpperCamelCase = [(pair[0], pair[1][4:]) if pair[1].startswith('''vit''' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('''norm.weight''', '''vit.layernorm.weight'''), ('''norm.bias''', '''vit.layernorm.bias'''), ('''head.weight''', '''classifier.weight'''), ('''head.bias''', '''classifier.bias'''), ] ) # fmt: on return rename_keys def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case=False ) -> int: """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: _UpperCamelCase = '''''' else: _UpperCamelCase = '''vit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _UpperCamelCase = state_dict.pop(F'''blocks.{i}.attn.qkv.weight''' ) _UpperCamelCase = state_dict.pop(F'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict _UpperCamelCase = in_proj_weight[ : config.hidden_size, : ] _UpperCamelCase = in_proj_bias[: config.hidden_size] _UpperCamelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _UpperCamelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _UpperCamelCase = in_proj_weight[ -config.hidden_size :, : ] _UpperCamelCase = in_proj_bias[-config.hidden_size :] def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(__snake_case, __snake_case ) def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> int: """simple docstring""" _UpperCamelCase = dct.pop(__snake_case ) _UpperCamelCase = val def lowerCamelCase__ ( ) -> List[Any]: """simple docstring""" _UpperCamelCase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _UpperCamelCase = Image.open(requests.get(__snake_case, stream=__snake_case ).raw ) return im @torch.no_grad() def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case=False ) -> str: """simple docstring""" _UpperCamelCase = BitConfig( global_padding='''same''', layer_type='''bottleneck''', depths=(3, 4, 9), out_features=['''stage3'''], embedding_dynamic_padding=__snake_case, ) _UpperCamelCase = ViTHybridConfig(backbone_config=__snake_case, image_size=3_84, num_labels=10_00 ) _UpperCamelCase = False # load original model from timm _UpperCamelCase = timm.create_model(__snake_case, pretrained=__snake_case ) timm_model.eval() # load state_dict of original model, remove and rename some keys _UpperCamelCase = timm_model.state_dict() if base_model: remove_classification_head_(__snake_case ) _UpperCamelCase = create_rename_keys(__snake_case, __snake_case ) for src, dest in rename_keys: rename_key(__snake_case, __snake_case, __snake_case ) read_in_q_k_v(__snake_case, __snake_case, __snake_case ) _UpperCamelCase = '''huggingface/label-files''' _UpperCamelCase = '''imagenet-1k-id2label.json''' _UpperCamelCase = json.load(open(hf_hub_download(__snake_case, __snake_case, repo_type='''dataset''' ), '''r''' ) ) _UpperCamelCase = {int(__snake_case ): v for k, v in idalabel.items()} _UpperCamelCase = idalabel _UpperCamelCase = {v: k for k, v in idalabel.items()} # load HuggingFace model if vit_name[-5:] == "in21k": _UpperCamelCase = ViTHybridModel(__snake_case ).eval() else: _UpperCamelCase = ViTHybridForImageClassification(__snake_case ).eval() model.load_state_dict(__snake_case ) # create image processor _UpperCamelCase = create_transform(**resolve_data_config({}, model=__snake_case ) ) _UpperCamelCase = transform.transforms _UpperCamelCase = { '''bilinear''': PILImageResampling.BILINEAR, '''bicubic''': PILImageResampling.BICUBIC, '''nearest''': PILImageResampling.NEAREST, } _UpperCamelCase = ViTHybridImageProcessor( do_resize=__snake_case, size={'''shortest_edge''': timm_transforms[0].size}, resample=pillow_resamplings[timm_transforms[0].interpolation.value], do_center_crop=__snake_case, crop_size={'''height''': timm_transforms[1].size[0], '''width''': timm_transforms[1].size[1]}, do_normalize=__snake_case, image_mean=timm_transforms[-1].mean.tolist(), image_std=timm_transforms[-1].std.tolist(), ) _UpperCamelCase = prepare_img() _UpperCamelCase = transform(__snake_case ).unsqueeze(0 ) _UpperCamelCase = processor(__snake_case, return_tensors='''pt''' ).pixel_values # verify pixel values assert torch.allclose(__snake_case, __snake_case ) # verify logits with torch.no_grad(): _UpperCamelCase = model(__snake_case ) _UpperCamelCase = outputs.logits print('''Predicted class:''', logits.argmax(-1 ).item() ) if base_model: _UpperCamelCase = timm_model.forward_features(__snake_case ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(__snake_case, outputs.pooler_output, atol=1e-3 ) else: _UpperCamelCase = timm_model(__snake_case ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__snake_case, outputs.logits, atol=1e-3 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: Path(__snake_case ).mkdir(exist_ok=__snake_case ) print(F'''Saving model {vit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(__snake_case ) print(F'''Saving processor to {pytorch_dump_folder_path}''' ) processor.save_pretrained(__snake_case ) if push_to_hub: print(F'''Pushing model and processor to the hub {vit_name}''' ) model.push_to_hub(F'''ybelkada/{vit_name}''' ) processor.push_to_hub(F'''ybelkada/{vit_name}''' ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( """--vit_name""", default="""vit_base_r50_s16_384""", type=str, help="""Name of the hybrid ViT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to upload the model to the HuggingFace hub.""" ) _a = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path, args.push_to_hub)
78
"""simple docstring""" import os from typing import BinaryIO, Optional, Union import numpy as np import pyarrow.parquet as pq from .. import Audio, Dataset, Features, Image, NamedSplit, Value, config from ..features.features import FeatureType, _visit from ..formatting import query_table from ..packaged_modules import _PACKAGED_DATASETS_MODULES from ..packaged_modules.parquet.parquet import Parquet from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" _UpperCamelCase = np.inf def set_batch_size(__snake_case ) -> None: nonlocal batch_size if isinstance(__snake_case, __snake_case ): _UpperCamelCase = min(__snake_case, config.PARQUET_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS ) elif isinstance(__snake_case, __snake_case ): _UpperCamelCase = min(__snake_case, config.PARQUET_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS ) elif isinstance(__snake_case, __snake_case ) and feature.dtype == "binary": _UpperCamelCase = min(__snake_case, config.PARQUET_ROW_GROUP_SIZE_FOR_BINARY_DATASETS ) _visit(__snake_case, __snake_case ) return None if batch_size is np.inf else batch_size class _UpperCAmelCase( lowerCamelCase ): def __init__( self , __a , __a = None , __a = None , __a = None , __a = False , __a = False , __a = None , **__a , ) -> Dict: '''simple docstring''' super().__init__( __a , split=__a , features=__a , cache_dir=__a , keep_in_memory=__a , streaming=__a , num_proc=__a , **__a , ) _UpperCamelCase = path_or_paths if isinstance(__a , __a) else {self.split: path_or_paths} _UpperCamelCase = _PACKAGED_DATASETS_MODULES['''parquet'''][1] _UpperCamelCase = Parquet( cache_dir=__a , data_files=__a , features=__a , hash=__a , **__a , ) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' # Build iterable dataset if self.streaming: _UpperCamelCase = self.builder.as_streaming_dataset(split=self.split) # Build regular (map-style) dataset else: _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None self.builder.download_and_prepare( download_config=__a , download_mode=__a , verification_mode=__a , base_path=__a , num_proc=self.num_proc , ) _UpperCamelCase = self.builder.as_dataset( split=self.split , verification_mode=__a , in_memory=self.keep_in_memory) return dataset class _UpperCAmelCase: def __init__( self , __a , __a , __a = None , **__a , ) -> Dict: '''simple docstring''' _UpperCamelCase = dataset _UpperCamelCase = path_or_buf _UpperCamelCase = batch_size or get_writer_batch_size(dataset.features) _UpperCamelCase = parquet_writer_kwargs def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = self.batch_size if self.batch_size else config.DEFAULT_MAX_BATCH_SIZE if isinstance(self.path_or_buf , (str, bytes, os.PathLike)): with open(self.path_or_buf , '''wb+''') as buffer: _UpperCamelCase = self._write(file_obj=__a , batch_size=__a , **self.parquet_writer_kwargs) else: _UpperCamelCase = self._write(file_obj=self.path_or_buf , batch_size=__a , **self.parquet_writer_kwargs) return written def UpperCAmelCase ( self , __a , __a , **__a) -> int: '''simple docstring''' _UpperCamelCase = 0 _UpperCamelCase = parquet_writer_kwargs.pop('''path_or_buf''' , __a) _UpperCamelCase = self.dataset.features.arrow_schema _UpperCamelCase = pq.ParquetWriter(__a , schema=__a , **__a) for offset in logging.tqdm( range(0 , len(self.dataset) , __a) , unit='''ba''' , disable=not logging.is_progress_bar_enabled() , desc='''Creating parquet from Arrow format''' , ): _UpperCamelCase = query_table( table=self.dataset._data , key=slice(__a , offset + batch_size) , indices=self.dataset._indices if self.dataset._indices is not None else None , ) writer.write_table(__a) written += batch.nbytes writer.close() return written
78
1
"""simple docstring""" from torch import nn class _UpperCAmelCase( nn.Module ): def __init__( self , __a , __a) -> Optional[Any]: '''simple docstring''' super().__init__() _UpperCamelCase = class_size _UpperCamelCase = embed_size # self.mlp1 = nn.Linear(embed_size, embed_size) # self.mlp2 = (nn.Linear(embed_size, class_size)) _UpperCamelCase = nn.Linear(__a , __a) def UpperCAmelCase ( self , __a) -> Any: '''simple docstring''' # hidden_state = nn.functional.relu(self.mlp1(hidden_state)) # hidden_state = self.mlp2(hidden_state) _UpperCamelCase = self.mlp(__a) return logits
78
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class _UpperCAmelCase( unittest.TestCase ): def __init__( self , __a , __a=7 , __a=3 , __a=18 , __a=30 , __a=4_00 , __a=True , __a=None , __a=True , __a=None , __a=True , ) -> int: '''simple docstring''' _UpperCamelCase = size if size is not None else {'''shortest_edge''': 20} _UpperCamelCase = crop_size if crop_size is not None else {'''height''': 18, '''width''': 18} _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = num_channels _UpperCamelCase = image_size _UpperCamelCase = min_resolution _UpperCamelCase = max_resolution _UpperCamelCase = do_resize _UpperCamelCase = size _UpperCamelCase = do_center_crop _UpperCamelCase = crop_size _UpperCamelCase = do_flip_channel_order def UpperCAmelCase ( self) -> str: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_flip_channel_order": self.do_flip_channel_order, } @require_torch @require_vision class _UpperCAmelCase( lowerCamelCase , unittest.TestCase ): lowercase__ = MobileViTImageProcessor if is_vision_available() else None def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = MobileViTImageProcessingTester(self) @property def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(__a , '''do_resize''')) self.assertTrue(hasattr(__a , '''size''')) self.assertTrue(hasattr(__a , '''do_center_crop''')) self.assertTrue(hasattr(__a , '''center_crop''')) self.assertTrue(hasattr(__a , '''do_flip_channel_order''')) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size , {'''shortest_edge''': 20}) self.assertEqual(image_processor.crop_size , {'''height''': 18, '''width''': 18}) _UpperCamelCase = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84) self.assertEqual(image_processor.size , {'''shortest_edge''': 42}) self.assertEqual(image_processor.crop_size , {'''height''': 84, '''width''': 84}) def UpperCAmelCase ( self) -> Dict: '''simple docstring''' pass def UpperCAmelCase ( self) -> str: '''simple docstring''' # Initialize image_processing _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random PIL images _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__a) for image in image_inputs: self.assertIsInstance(__a , Image.Image) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(__a , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' # Initialize image_processing _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__a , numpify=__a) for image in image_inputs: self.assertIsInstance(__a , np.ndarray) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(__a , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def UpperCAmelCase ( self) -> int: '''simple docstring''' # Initialize image_processing _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__a , torchify=__a) for image in image_inputs: self.assertIsInstance(__a , torch.Tensor) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(__a , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
78
1
"""simple docstring""" import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert_fast import BertTokenizerFast from .tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer, DPRReaderTokenizer _a = logging.get_logger(__name__) _a = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} _a = { """vocab_file""": { """facebook/dpr-ctx_encoder-single-nq-base""": ( """https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/vocab.txt""" ), """facebook/dpr-ctx_encoder-multiset-base""": ( """https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """facebook/dpr-ctx_encoder-single-nq-base""": ( """https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/tokenizer.json""" ), """facebook/dpr-ctx_encoder-multiset-base""": ( """https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/tokenizer.json""" ), }, } _a = { """vocab_file""": { """facebook/dpr-question_encoder-single-nq-base""": ( """https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/vocab.txt""" ), """facebook/dpr-question_encoder-multiset-base""": ( """https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """facebook/dpr-question_encoder-single-nq-base""": ( """https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/tokenizer.json""" ), """facebook/dpr-question_encoder-multiset-base""": ( """https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/tokenizer.json""" ), }, } _a = { """vocab_file""": { """facebook/dpr-reader-single-nq-base""": ( """https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/vocab.txt""" ), """facebook/dpr-reader-multiset-base""": ( """https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """facebook/dpr-reader-single-nq-base""": ( """https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/tokenizer.json""" ), """facebook/dpr-reader-multiset-base""": ( """https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/tokenizer.json""" ), }, } _a = { """facebook/dpr-ctx_encoder-single-nq-base""": 512, """facebook/dpr-ctx_encoder-multiset-base""": 512, } _a = { """facebook/dpr-question_encoder-single-nq-base""": 512, """facebook/dpr-question_encoder-multiset-base""": 512, } _a = { """facebook/dpr-reader-single-nq-base""": 512, """facebook/dpr-reader-multiset-base""": 512, } _a = { """facebook/dpr-ctx_encoder-single-nq-base""": {"""do_lower_case""": True}, """facebook/dpr-ctx_encoder-multiset-base""": {"""do_lower_case""": True}, } _a = { """facebook/dpr-question_encoder-single-nq-base""": {"""do_lower_case""": True}, """facebook/dpr-question_encoder-multiset-base""": {"""do_lower_case""": True}, } _a = { """facebook/dpr-reader-single-nq-base""": {"""do_lower_case""": True}, """facebook/dpr-reader-multiset-base""": {"""do_lower_case""": True}, } class _UpperCAmelCase( lowerCamelCase ): lowercase__ = VOCAB_FILES_NAMES lowercase__ = CONTEXT_ENCODER_PRETRAINED_VOCAB_FILES_MAP lowercase__ = CONTEXT_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = CONTEXT_ENCODER_PRETRAINED_INIT_CONFIGURATION lowercase__ = DPRContextEncoderTokenizer class _UpperCAmelCase( lowerCamelCase ): lowercase__ = VOCAB_FILES_NAMES lowercase__ = QUESTION_ENCODER_PRETRAINED_VOCAB_FILES_MAP lowercase__ = QUESTION_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = QUESTION_ENCODER_PRETRAINED_INIT_CONFIGURATION lowercase__ = DPRQuestionEncoderTokenizer _a = collections.namedtuple( """DPRSpanPrediction""", ["""span_score""", """relevance_score""", """doc_id""", """start_index""", """end_index""", """text"""] ) _a = collections.namedtuple("""DPRReaderOutput""", ["""start_logits""", """end_logits""", """relevance_logits"""]) _a = R""" Return a dictionary with the token ids of the input strings and other information to give to `.decode_best_spans`. It converts the strings of a question and different passages (title and text) in a sequence of IDs (integers), using the tokenizer and vocabulary. The resulting `input_ids` is a matrix of size `(n_passages, sequence_length)` with the format: [CLS] <question token ids> [SEP] <titles ids> [SEP] <texts ids> Args: questions (`str` or `List[str]`): The questions to be encoded. You can specify one question for many passages. In this case, the question will be duplicated like `[questions] * n_passages`. Otherwise you have to specify as many questions as in `titles` or `texts`. titles (`str` or `List[str]`): The passages titles to be encoded. This can be a string or a list of strings if there are several passages. texts (`str` or `List[str]`): The passages texts to be encoded. This can be a string or a list of strings if there are several passages. padding (`bool`, `str` or [`~utils.PaddingStrategy`], *optional*, defaults to `False`): Activates and controls padding. Accepts the following values: - `True` or `'longest'`: Pad to the longest sequence in the batch (or no padding if only a single sequence if provided). - `'max_length'`: Pad to a maximum length specified with the argument `max_length` or to the maximum acceptable input length for the model if that argument is not provided. - `False` or `'do_not_pad'` (default): No padding (i.e., can output a batch with sequences of different lengths). truncation (`bool`, `str` or [`~tokenization_utils_base.TruncationStrategy`], *optional*, defaults to `False`): Activates and controls truncation. Accepts the following values: - `True` or `'longest_first'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum acceptable input length for the model if that argument is not provided. This will truncate token by token, removing a token from the longest sequence in the pair if a pair of sequences (or a batch of pairs) is provided. - `'only_first'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum acceptable input length for the model if that argument is not provided. This will only truncate the first sequence of a pair if a pair of sequences (or a batch of pairs) is provided. - `'only_second'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum acceptable input length for the model if that argument is not provided. This will only truncate the second sequence of a pair if a pair of sequences (or a batch of pairs) is provided. - `False` or `'do_not_truncate'` (default): No truncation (i.e., can output batch with sequence lengths greater than the model maximum admissible input size). max_length (`int`, *optional*): Controls the maximum length to use by one of the truncation/padding parameters. If left unset or set to `None`, this will use the predefined model maximum length if a maximum length is required by one of the truncation/padding parameters. If the model has no specific maximum input length (like XLNet) truncation/padding to a maximum length will be deactivated. return_tensors (`str` or [`~utils.TensorType`], *optional*): If set, will return tensors instead of list of python integers. Acceptable values are: - `'tf'`: Return TensorFlow `tf.constant` objects. - `'pt'`: Return PyTorch `torch.Tensor` objects. - `'np'`: Return Numpy `np.ndarray` objects. return_attention_mask (`bool`, *optional*): Whether or not to return the attention mask. If not set, will return the attention mask according to the specific tokenizer's default, defined by the `return_outputs` attribute. [What are attention masks?](../glossary#attention-mask) Return: `Dict[str, List[List[int]]]`: A dictionary with the following keys: - `input_ids`: List of token ids to be fed to a model. - `attention_mask`: List of indices specifying which tokens should be attended to by the model. """ @add_start_docstrings(lowerCamelCase ) class _UpperCAmelCase: def __call__( self , __a , __a = None , __a = None , __a = False , __a = False , __a = None , __a = None , __a = None , **__a , ) -> BatchEncoding: '''simple docstring''' if titles is None and texts is None: return super().__call__( __a , padding=__a , truncation=__a , max_length=__a , return_tensors=__a , return_attention_mask=__a , **__a , ) elif titles is None or texts is None: _UpperCamelCase = titles if texts is None else texts return super().__call__( __a , __a , padding=__a , truncation=__a , max_length=__a , return_tensors=__a , return_attention_mask=__a , **__a , ) _UpperCamelCase = titles if not isinstance(__a , __a) else [titles] _UpperCamelCase = texts if not isinstance(__a , __a) else [texts] _UpperCamelCase = len(__a) _UpperCamelCase = questions if not isinstance(__a , __a) else [questions] * n_passages assert len(__a) == len( __a), F'''There should be as many titles than texts but got {len(__a)} titles and {len(__a)} texts.''' _UpperCamelCase = super().__call__(__a , __a , padding=__a , truncation=__a)['''input_ids'''] _UpperCamelCase = super().__call__(__a , add_special_tokens=__a , padding=__a , truncation=__a)['''input_ids'''] _UpperCamelCase = { '''input_ids''': [ (encoded_question_and_title + encoded_text)[:max_length] if max_length is not None and truncation else encoded_question_and_title + encoded_text for encoded_question_and_title, encoded_text in zip(__a , __a) ] } if return_attention_mask is not False: _UpperCamelCase = [] for input_ids in encoded_inputs["input_ids"]: attention_mask.append([int(input_id != self.pad_token_id) for input_id in input_ids]) _UpperCamelCase = attention_mask return self.pad(__a , padding=__a , max_length=__a , return_tensors=__a) def UpperCAmelCase ( self , __a , __a , __a = 16 , __a = 64 , __a = 4 , ) -> List[DPRSpanPrediction]: '''simple docstring''' _UpperCamelCase = reader_input['''input_ids'''] _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = reader_output[:3] _UpperCamelCase = len(__a) _UpperCamelCase = sorted(range(__a) , reverse=__a , key=relevance_logits.__getitem__) _UpperCamelCase = [] for doc_id in sorted_docs: _UpperCamelCase = list(input_ids[doc_id]) # assuming question & title information is at the beginning of the sequence _UpperCamelCase = sequence_ids.index(self.sep_token_id , 2) + 1 # second sep id if sequence_ids[-1] == self.pad_token_id: _UpperCamelCase = sequence_ids.index(self.pad_token_id) else: _UpperCamelCase = len(__a) _UpperCamelCase = self._get_best_spans( start_logits=start_logits[doc_id][passage_offset:sequence_len] , end_logits=end_logits[doc_id][passage_offset:sequence_len] , max_answer_length=__a , top_spans=__a , ) for start_index, end_index in best_spans: start_index += passage_offset end_index += passage_offset nbest_spans_predictions.append( DPRSpanPrediction( span_score=start_logits[doc_id][start_index] + end_logits[doc_id][end_index] , relevance_score=relevance_logits[doc_id] , doc_id=__a , start_index=__a , end_index=__a , text=self.decode(sequence_ids[start_index : end_index + 1]) , )) if len(__a) >= num_spans: break return nbest_spans_predictions[:num_spans] def UpperCAmelCase ( self , __a , __a , __a , __a , ) -> List[DPRSpanPrediction]: '''simple docstring''' _UpperCamelCase = [] for start_index, start_score in enumerate(__a): for answer_length, end_score in enumerate(end_logits[start_index : start_index + max_answer_length]): scores.append(((start_index, start_index + answer_length), start_score + end_score)) _UpperCamelCase = sorted(__a , key=lambda __a: x[1] , reverse=__a) _UpperCamelCase = [] for (start_index, end_index), score in scores: assert start_index <= end_index, F'''Wrong span indices: [{start_index}:{end_index}]''' _UpperCamelCase = end_index - start_index + 1 assert length <= max_answer_length, F'''Span is too long: {length} > {max_answer_length}''' if any( start_index <= prev_start_index <= prev_end_index <= end_index or prev_start_index <= start_index <= end_index <= prev_end_index for (prev_start_index, prev_end_index) in chosen_span_intervals): continue chosen_span_intervals.append((start_index, end_index)) if len(__a) == top_spans: break return chosen_span_intervals @add_end_docstrings(lowerCamelCase ) class _UpperCAmelCase( lowerCamelCase , lowerCamelCase ): lowercase__ = VOCAB_FILES_NAMES lowercase__ = READER_PRETRAINED_VOCAB_FILES_MAP lowercase__ = READER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = READER_PRETRAINED_INIT_CONFIGURATION lowercase__ = ['input_ids', 'attention_mask'] lowercase__ = DPRReaderTokenizer
78
"""simple docstring""" import warnings from typing import List import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import is_flax_available, is_tf_available, is_torch_available class _UpperCAmelCase( lowerCamelCase ): lowercase__ = ['image_processor', 'tokenizer'] lowercase__ = 'OwlViTImageProcessor' lowercase__ = ('CLIPTokenizer', 'CLIPTokenizerFast') def __init__( self , __a=None , __a=None , **__a) -> List[Any]: '''simple docstring''' _UpperCamelCase = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , __a , ) _UpperCamelCase = kwargs.pop('''feature_extractor''') _UpperCamelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''') if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''') super().__init__(__a , __a) def __call__( self , __a=None , __a=None , __a=None , __a="max_length" , __a="np" , **__a) -> List[str]: '''simple docstring''' if text is None and query_images is None and images is None: raise ValueError( '''You have to specify at least one text or query image or image. All three cannot be none.''') if text is not None: if isinstance(__a , __a) or (isinstance(__a , __a) and not isinstance(text[0] , __a)): _UpperCamelCase = [self.tokenizer(__a , padding=__a , return_tensors=__a , **__a)] elif isinstance(__a , __a) and isinstance(text[0] , __a): _UpperCamelCase = [] # Maximum number of queries across batch _UpperCamelCase = max([len(__a) for t in text]) # Pad all batch samples to max number of text queries for t in text: if len(__a) != max_num_queries: _UpperCamelCase = t + [''' '''] * (max_num_queries - len(__a)) _UpperCamelCase = self.tokenizer(__a , padding=__a , return_tensors=__a , **__a) encodings.append(__a) else: raise TypeError('''Input text should be a string, a list of strings or a nested list of strings''') if return_tensors == "np": _UpperCamelCase = np.concatenate([encoding['''input_ids'''] for encoding in encodings] , axis=0) _UpperCamelCase = np.concatenate([encoding['''attention_mask'''] for encoding in encodings] , axis=0) elif return_tensors == "jax" and is_flax_available(): import jax.numpy as jnp _UpperCamelCase = jnp.concatenate([encoding['''input_ids'''] for encoding in encodings] , axis=0) _UpperCamelCase = jnp.concatenate([encoding['''attention_mask'''] for encoding in encodings] , axis=0) elif return_tensors == "pt" and is_torch_available(): import torch _UpperCamelCase = torch.cat([encoding['''input_ids'''] for encoding in encodings] , dim=0) _UpperCamelCase = torch.cat([encoding['''attention_mask'''] for encoding in encodings] , dim=0) elif return_tensors == "tf" and is_tf_available(): import tensorflow as tf _UpperCamelCase = tf.stack([encoding['''input_ids'''] for encoding in encodings] , axis=0) _UpperCamelCase = tf.stack([encoding['''attention_mask'''] for encoding in encodings] , axis=0) else: raise ValueError('''Target return tensor type could not be returned''') _UpperCamelCase = BatchEncoding() _UpperCamelCase = input_ids _UpperCamelCase = attention_mask if query_images is not None: _UpperCamelCase = BatchEncoding() _UpperCamelCase = self.image_processor( __a , return_tensors=__a , **__a).pixel_values _UpperCamelCase = query_pixel_values if images is not None: _UpperCamelCase = self.image_processor(__a , return_tensors=__a , **__a) if text is not None and images is not None: _UpperCamelCase = image_features.pixel_values return encoding elif query_images is not None and images is not None: _UpperCamelCase = image_features.pixel_values return encoding elif text is not None or query_images is not None: return encoding else: return BatchEncoding(data=dict(**__a) , tensor_type=__a) def UpperCAmelCase ( self , *__a , **__a) -> str: '''simple docstring''' return self.image_processor.post_process(*__a , **__a) def UpperCAmelCase ( self , *__a , **__a) -> Dict: '''simple docstring''' return self.image_processor.post_process_object_detection(*__a , **__a) def UpperCAmelCase ( self , *__a , **__a) -> Optional[Any]: '''simple docstring''' return self.image_processor.post_process_image_guided_detection(*__a , **__a) def UpperCAmelCase ( self , *__a , **__a) -> Optional[int]: '''simple docstring''' return self.tokenizer.batch_decode(*__a , **__a) def UpperCAmelCase ( self , *__a , **__a) -> Optional[Any]: '''simple docstring''' return self.tokenizer.decode(*__a , **__a) @property def UpperCAmelCase ( self) -> str: '''simple docstring''' warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , __a , ) return self.image_processor_class @property def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , __a , ) return self.image_processor
78
1
"""simple docstring""" import builtins import sys from ...utils.imports import _is_package_available from . import cursor, input from .helpers import Direction, clear_line, forceWrite, linebreak, move_cursor, reset_cursor, writeColor from .keymap import KEYMAP _a = False try: _a = _is_package_available("""google.colab""") except ModuleNotFoundError: pass @input.register class _UpperCAmelCase: def __init__( self , __a = None , __a = []) -> Optional[int]: '''simple docstring''' _UpperCamelCase = 0 _UpperCamelCase = choices _UpperCamelCase = prompt if sys.platform == "win32": _UpperCamelCase = '''*''' else: _UpperCamelCase = '''➔ ''' def UpperCAmelCase ( self , __a , __a = "") -> Tuple: '''simple docstring''' if sys.platform != "win32": writeColor(self.choices[index] , 32 , __a) else: forceWrite(self.choices[index] , __a) def UpperCAmelCase ( self , __a) -> str: '''simple docstring''' if index == self.position: forceWrite(F''' {self.arrow_char} ''') self.write_choice(__a) else: forceWrite(F''' {self.choices[index]}''') reset_cursor() def UpperCAmelCase ( self , __a , __a = 1) -> List[str]: '''simple docstring''' _UpperCamelCase = self.position if direction == Direction.DOWN: if self.position + 1 >= len(self.choices): return self.position += num_spaces else: if self.position - 1 < 0: return self.position -= num_spaces clear_line() self.print_choice(__a) move_cursor(__a , direction.name) self.print_choice(self.position) @input.mark(KEYMAP['''up''']) def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' self.move_direction(Direction.UP) @input.mark(KEYMAP['''down''']) def UpperCAmelCase ( self) -> Any: '''simple docstring''' self.move_direction(Direction.DOWN) @input.mark(KEYMAP['''newline''']) def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' move_cursor(len(self.choices) - self.position , '''DOWN''') return self.position @input.mark(KEYMAP['''interrupt''']) def UpperCAmelCase ( self) -> str: '''simple docstring''' move_cursor(len(self.choices) - self.position , '''DOWN''') raise KeyboardInterrupt @input.mark_multiple(*[KEYMAP[str(__a)] for number in range(10)]) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = int(chr(self.current_selection)) _UpperCamelCase = index - self.position if index == self.position: return if index < len(self.choices): if self.position > index: self.move_direction(Direction.UP , -movement) elif self.position < index: self.move_direction(Direction.DOWN , __a) else: return else: return def UpperCAmelCase ( self , __a = 0) -> Optional[Any]: '''simple docstring''' if self.prompt: linebreak() forceWrite(self.prompt , '''\n''') if in_colab: forceWrite('''Please input a choice index (starting from 0), and press enter''' , '''\n''') else: forceWrite('''Please select a choice using the arrow or number keys, and selecting with enter''' , '''\n''') _UpperCamelCase = default_choice for i in range(len(self.choices)): self.print_choice(__a) forceWrite('''\n''') move_cursor(len(self.choices) - self.position , '''UP''') with cursor.hide(): while True: if in_colab: try: _UpperCamelCase = int(builtins.input()) except ValueError: _UpperCamelCase = default_choice else: _UpperCamelCase = self.handle_input() if choice is not None: reset_cursor() for _ in range(len(self.choices) + 1): move_cursor(1 , '''UP''') clear_line() self.write_choice(__a , '''\n''') return choice
78
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) _a = { """configuration_perceiver""": ["""PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """PerceiverConfig""", """PerceiverOnnxConfig"""], """tokenization_perceiver""": ["""PerceiverTokenizer"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ["""PerceiverFeatureExtractor"""] _a = ["""PerceiverImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ """PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST""", """PerceiverForImageClassificationConvProcessing""", """PerceiverForImageClassificationFourier""", """PerceiverForImageClassificationLearned""", """PerceiverForMaskedLM""", """PerceiverForMultimodalAutoencoding""", """PerceiverForOpticalFlow""", """PerceiverForSequenceClassification""", """PerceiverLayer""", """PerceiverModel""", """PerceiverPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys _a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
78
1
"""simple docstring""" import unittest from transformers import MPNetConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) class _UpperCAmelCase: def __init__( self , __a , __a=13 , __a=7 , __a=True , __a=True , __a=False , __a=True , __a=99 , __a=64 , __a=5 , __a=4 , __a=64 , __a="gelu" , __a=0.1 , __a=0.1 , __a=5_12 , __a=16 , __a=2 , __a=0.02 , __a=3 , __a=4 , __a=None , ) -> Optional[int]: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = seq_length _UpperCamelCase = is_training _UpperCamelCase = use_input_mask _UpperCamelCase = use_token_type_ids _UpperCamelCase = use_labels _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = type_vocab_size _UpperCamelCase = type_sequence_label_size _UpperCamelCase = initializer_range _UpperCamelCase = num_labels _UpperCamelCase = num_choices _UpperCamelCase = scope def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' return MPNetConfig.from_pretrained('''microsoft/mpnet-base''') def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) _UpperCamelCase = None if self.use_input_mask: _UpperCamelCase = random_attention_mask([self.batch_size, self.seq_length]) _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size) _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) _UpperCamelCase = ids_tensor([self.batch_size] , self.num_choices) _UpperCamelCase = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase ( self) -> Dict: '''simple docstring''' return MPNetConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = MPNetModel(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , __a) _UpperCamelCase = model(__a) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a) -> str: '''simple docstring''' _UpperCamelCase = MPNetForQuestionAnswering(config=__a) model.to(__a) model.eval() _UpperCamelCase = model( __a , attention_mask=__a , start_positions=__a , end_positions=__a , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a) -> Tuple: '''simple docstring''' _UpperCamelCase = self.num_labels _UpperCamelCase = MPNetForSequenceClassification(__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , labels=__a) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a) -> Dict: '''simple docstring''' _UpperCamelCase = self.num_choices _UpperCamelCase = MPNetForMultipleChoice(config=__a) model.to(__a) model.eval() _UpperCamelCase = input_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() _UpperCamelCase = input_mask.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() _UpperCamelCase = model( __a , attention_mask=__a , labels=__a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a) -> Any: '''simple docstring''' _UpperCamelCase = self.num_labels _UpperCamelCase = MPNetForTokenClassification(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , labels=__a) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def UpperCAmelCase ( self) -> str: '''simple docstring''' _UpperCamelCase = self.prepare_config_and_inputs() ((_UpperCamelCase) , (_UpperCamelCase) , (_UpperCamelCase) , (_UpperCamelCase) , (_UpperCamelCase) , (_UpperCamelCase)) = config_and_inputs _UpperCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class _UpperCAmelCase( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase__ = ( ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) if is_torch_available() else () ) lowercase__ = ( { 'feature-extraction': MPNetModel, 'fill-mask': MPNetForMaskedLM, 'question-answering': MPNetForQuestionAnswering, 'text-classification': MPNetForSequenceClassification, 'token-classification': MPNetForTokenClassification, 'zero-shot': MPNetForSequenceClassification, } if is_torch_available() else {} ) lowercase__ = False lowercase__ = True def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = MPNetModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=__a , hidden_size=37) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_model(*__a) def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_sequence_classification(*__a) def UpperCAmelCase ( self) -> Dict: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_multiple_choice(*__a) def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_token_classification(*__a) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_question_answering(*__a) @require_torch class _UpperCAmelCase( unittest.TestCase ): @slow def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = MPNetModel.from_pretrained('''microsoft/mpnet-base''') _UpperCamelCase = torch.tensor([[0, 3_45, 2_32, 3_28, 7_40, 1_40, 16_95, 69, 60_78, 15_88, 2]]) _UpperCamelCase = model(__a)[0] _UpperCamelCase = torch.Size((1, 11, 7_68)) self.assertEqual(output.shape , __a) _UpperCamelCase = torch.tensor( [[[-0.0550, 0.1943, -0.0740], [-0.0562, 0.2211, -0.0579], [-0.0437, 0.3337, -0.0641]]]) # compare the actual values for a slice. self.assertTrue(torch.allclose(output[:, :3, :3] , __a , atol=1e-4))
78
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class _UpperCAmelCase: def __init__( self , __a , __a=13 , __a=2 , __a=24 , __a=16 , __a=True , __a=True , __a=32 , __a=5 , __a=4 , __a=37 , __a="gelu" , __a=0.1 , __a=0.1 , __a=10 , __a=0.02 , __a=None , __a=2 , __a=2 , ) -> List[str]: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = patch_size _UpperCamelCase = max_length _UpperCamelCase = num_mel_bins _UpperCamelCase = is_training _UpperCamelCase = use_labels _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = type_sequence_label_size _UpperCamelCase = initializer_range _UpperCamelCase = scope _UpperCamelCase = frequency_stride _UpperCamelCase = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) _UpperCamelCase = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 _UpperCamelCase = (self.max_length - self.patch_size) // self.time_stride + 1 _UpperCamelCase = frequency_out_dimension * time_out_dimension _UpperCamelCase = num_patches + 2 def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins]) _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size) _UpperCamelCase = self.get_config() return config, input_values, labels def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__a , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def UpperCAmelCase ( self , __a , __a , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = ASTModel(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.prepare_config_and_inputs() ( ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ) = config_and_inputs _UpperCamelCase = {'''input_values''': input_values} return config, inputs_dict @require_torch class _UpperCAmelCase( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase__ = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) lowercase__ = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) lowercase__ = False lowercase__ = False lowercase__ = False lowercase__ = False def UpperCAmelCase ( self , __a , __a , __a , __a , __a) -> Optional[Any]: '''simple docstring''' if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = ASTModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=__a , has_text_modality=__a , hidden_size=37) def UpperCAmelCase ( self) -> int: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='''AST does not use inputs_embeds''') def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' pass def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = model_class(__a) self.assertIsInstance(model.get_input_embeddings() , (nn.Module)) _UpperCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__a , nn.Linear)) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = model_class(__a) _UpperCamelCase = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCamelCase = [*signature.parameters.keys()] _UpperCamelCase = ['''input_values'''] self.assertListEqual(arg_names[:1] , __a) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a) @slow def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase = ASTModel.from_pretrained(__a) self.assertIsNotNone(__a) def lowerCamelCase__ ( ) -> List[str]: """simple docstring""" _UpperCamelCase = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''', filename='''sample_audio.flac''', repo_type='''dataset''' ) _UpperCamelCase , _UpperCamelCase = torchaudio.load(__snake_case ) return audio, sampling_rate @require_torch @require_torchaudio class _UpperCAmelCase( unittest.TestCase ): @cached_property def UpperCAmelCase ( self) -> Dict: '''simple docstring''' return ( ASTFeatureExtractor.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''') if is_torchaudio_available() else None ) @slow def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.default_feature_extractor _UpperCamelCase = ASTForAudioClassification.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''').to(__a) _UpperCamelCase = self.default_feature_extractor _UpperCamelCase , _UpperCamelCase = prepare_audio() _UpperCamelCase = audio.squeeze().numpy() _UpperCamelCase = feature_extractor(__a , sampling_rate=__a , return_tensors='''pt''').to(__a) # forward pass with torch.no_grad(): _UpperCamelCase = model(**__a) # verify the logits _UpperCamelCase = torch.Size((1, 5_27)) self.assertEqual(outputs.logits.shape , __a) _UpperCamelCase = torch.tensor([-0.8760, -7.0042, -8.6602]).to(__a) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __a , atol=1e-4))
78
1
"""simple docstring""" def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" assert isinstance(__snake_case, __snake_case ), F'''The input value of [n={number}] is not an integer''' if number == 1: return 2 elif number < 1: _UpperCamelCase = F'''The input value of [n={number}] has to be > 0''' raise ValueError(__snake_case ) else: _UpperCamelCase = sylvester(number - 1 ) _UpperCamelCase = num - 1 _UpperCamelCase = num return lower * upper + 1 if __name__ == "__main__": print(F"""The 8th number in Sylvester's sequence: {sylvester(8)}""")
78
"""simple docstring""" def lowerCamelCase__ ( ) -> list[list[int]]: """simple docstring""" return [list(range(10_00 - i, -10_00 - i, -1 ) ) for i in range(10_00 )] _a = generate_large_matrix() _a = ( [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]], [[3, 2], [1, 0]], [[7, 7, 6]], [[7, 7, 6], [-1, -2, -3]], grid, ) def lowerCamelCase__ ( __snake_case ) -> None: """simple docstring""" assert all(row == sorted(__snake_case, reverse=__snake_case ) for row in grid ) assert all(list(__snake_case ) == sorted(__snake_case, reverse=__snake_case ) for col in zip(*__snake_case ) ) def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" _UpperCamelCase = 0 _UpperCamelCase = len(__snake_case ) - 1 # Edge cases such as no values or all numbers are negative. if not array or array[0] < 0: return 0 while right + 1 > left: _UpperCamelCase = (left + right) // 2 _UpperCamelCase = array[mid] # Num must be negative and the index must be greater than or equal to 0. if num < 0 and array[mid - 1] >= 0: return mid if num >= 0: _UpperCamelCase = mid + 1 else: _UpperCamelCase = mid - 1 # No negative numbers so return the last index of the array + 1 which is the length. return len(__snake_case ) def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" _UpperCamelCase = 0 _UpperCamelCase = len(grid[0] ) for i in range(len(__snake_case ) ): _UpperCamelCase = find_negative_index(grid[i][:bound] ) total += bound return (len(__snake_case ) * len(grid[0] )) - total def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" return len([number for row in grid for number in row if number < 0] ) def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" _UpperCamelCase = 0 for row in grid: for i, number in enumerate(__snake_case ): if number < 0: total += len(__snake_case ) - i break return total def lowerCamelCase__ ( ) -> None: """simple docstring""" from timeit import timeit print('''Running benchmarks''' ) _UpperCamelCase = ( '''from __main__ import count_negatives_binary_search, ''' '''count_negatives_brute_force, count_negatives_brute_force_with_break, grid''' ) for func in ( "count_negatives_binary_search", # took 0.7727 seconds "count_negatives_brute_force_with_break", # took 4.6505 seconds "count_negatives_brute_force", # took 12.8160 seconds ): _UpperCamelCase = timeit(F'''{func}(grid=grid)''', setup=__snake_case, number=5_00 ) print(F'''{func}() took {time:0.4f} seconds''' ) if __name__ == "__main__": import doctest doctest.testmod() benchmark()
78
1
"""simple docstring""" import argparse import re import requests import torch # git clone https://github.com/salesforce/BLIP.git from models.blip import blip_decoder from models.blip_itm import blip_itm from models.blip_vqa import blip_vqa from PIL import Image from torchvision import transforms from torchvision.transforms.functional import InterpolationMode from transformers import ( BertTokenizer, BlipConfig, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, ) def lowerCamelCase__ ( __snake_case, __snake_case ) -> Optional[int]: """simple docstring""" _UpperCamelCase = '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg''' _UpperCamelCase = Image.open(requests.get(__snake_case, stream=__snake_case ).raw ).convert('''RGB''' ) _UpperCamelCase = transforms.Compose( [ transforms.Resize((image_size, image_size), interpolation=InterpolationMode.BICUBIC ), transforms.ToTensor(), transforms.Normalize((0.48145466, 0.4578275, 0.40821073), (0.26862954, 0.26130258, 0.27577711) ), ] ) _UpperCamelCase = transform(__snake_case ).unsqueeze(0 ).to(__snake_case ) return image def lowerCamelCase__ ( __snake_case ) -> Dict: """simple docstring""" if "visual_encoder" in key: _UpperCamelCase = re.sub('''visual_encoder*''', '''vision_model.encoder''', __snake_case ) if "blocks" in key: _UpperCamelCase = re.sub(r'''blocks''', '''layers''', __snake_case ) if "attn" in key: _UpperCamelCase = re.sub(r'''attn''', '''self_attn''', __snake_case ) if "norm1" in key: _UpperCamelCase = re.sub(r'''norm1''', '''layer_norm1''', __snake_case ) if "norm2" in key: _UpperCamelCase = re.sub(r'''norm2''', '''layer_norm2''', __snake_case ) if "encoder.norm" in key: _UpperCamelCase = re.sub(r'''encoder.norm''', '''post_layernorm''', __snake_case ) if "encoder.patch_embed.proj" in key: _UpperCamelCase = re.sub(r'''encoder.patch_embed.proj''', '''embeddings.patch_embedding''', __snake_case ) if "encoder.pos_embed" in key: _UpperCamelCase = re.sub(r'''encoder.pos_embed''', '''embeddings.position_embedding''', __snake_case ) if "encoder.cls_token" in key: _UpperCamelCase = re.sub(r'''encoder.cls_token''', '''embeddings.class_embedding''', __snake_case ) if "self_attn" in key: _UpperCamelCase = re.sub(r'''self_attn.proj''', '''self_attn.projection''', __snake_case ) return key @torch.no_grad() def lowerCamelCase__ ( __snake_case, __snake_case=None ) -> Union[str, Any]: """simple docstring""" if config_path is not None: _UpperCamelCase = BlipConfig.from_pretrained(__snake_case ) else: _UpperCamelCase = BlipConfig(projection_dim=5_12, text_config={}, vision_config={} ) _UpperCamelCase = BlipForConditionalGeneration(__snake_case ).eval() _UpperCamelCase = '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_capfilt_large.pth''' _UpperCamelCase = blip_decoder(pretrained=__snake_case, image_size=3_84, vit='''base''' ) _UpperCamelCase = pt_model.eval() _UpperCamelCase = pt_model.state_dict() for key in modified_state_dict.copy(): _UpperCamelCase = modified_state_dict.pop(__snake_case ) _UpperCamelCase = rename_key(__snake_case ) _UpperCamelCase = value hf_model.load_state_dict(__snake_case ) _UpperCamelCase = 3_84 _UpperCamelCase = load_demo_image(image_size=__snake_case, device='''cpu''' ) _UpperCamelCase = BertTokenizer.from_pretrained('''bert-base-uncased''' ) _UpperCamelCase = tokenizer(['''a picture of'''] ).input_ids _UpperCamelCase = hf_model.generate(__snake_case, __snake_case ) assert out[0].tolist() == [3_05_22, 10_37, 38_61, 19_97, 10_37, 24_50, 35_64, 20_06, 19_96, 35_09, 20_07, 20_14, 38_99, 1_02] _UpperCamelCase = hf_model.generate(__snake_case ) assert out[0].tolist() == [3_05_22, 10_37, 24_50, 35_64, 20_06, 19_96, 35_09, 20_07, 20_14, 38_99, 1_02] if pytorch_dump_folder_path is not None: hf_model.save_pretrained(__snake_case ) # model_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_vqa.pth' _UpperCamelCase = ( '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_vqa_capfilt_large.pth''' ) _UpperCamelCase = blip_vqa(pretrained=__snake_case, image_size=__snake_case, vit='''base''' ) vqa_model.eval() _UpperCamelCase = vqa_model.state_dict() for key in modified_state_dict.copy(): _UpperCamelCase = modified_state_dict.pop(__snake_case ) _UpperCamelCase = rename_key(__snake_case ) _UpperCamelCase = value _UpperCamelCase = BlipForQuestionAnswering(__snake_case ) hf_vqa_model.load_state_dict(__snake_case ) _UpperCamelCase = ['''How many dogs are in this image?'''] _UpperCamelCase = tokenizer(__snake_case, return_tensors='''pt''' ).input_ids _UpperCamelCase = hf_vqa_model.generate(__snake_case, __snake_case ) print(tokenizer.decode(answer[0] ) ) assert tokenizer.decode(answer[0] ) == "[UNK] 1 [SEP]" if pytorch_dump_folder_path is not None: hf_vqa_model.save_pretrained(pytorch_dump_folder_path + '''_vqa''' ) _UpperCamelCase = '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_retrieval_coco.pth''' _UpperCamelCase = blip_itm(pretrained=__snake_case, image_size=__snake_case, vit='''base''' ) itm_model.eval() _UpperCamelCase = itm_model.state_dict() for key in modified_state_dict.copy(): _UpperCamelCase = modified_state_dict.pop(__snake_case ) _UpperCamelCase = rename_key(__snake_case ) _UpperCamelCase = value _UpperCamelCase = BlipForImageTextRetrieval(__snake_case ) _UpperCamelCase = ['''A picture of a woman with a dog sitting in a beach'''] _UpperCamelCase = tokenizer( __snake_case, return_tensors='''pt''', padding='''max_length''', truncation=__snake_case, max_length=35, ).input_ids hf_itm_model.load_state_dict(__snake_case ) hf_itm_model.eval() _UpperCamelCase = hf_itm_model(__snake_case, __snake_case, use_itm_head=__snake_case ) _UpperCamelCase = hf_itm_model(__snake_case, __snake_case, use_itm_head=__snake_case ) assert out[0].item() == 0.2110687494277954 assert torch.nn.functional.softmax(out_itm[0], dim=1 )[:, 1].item() == 0.45698845386505127 if pytorch_dump_folder_path is not None: hf_itm_model.save_pretrained(pytorch_dump_folder_path + '''_itm''' ) if __name__ == "__main__": _a = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") _a = parser.parse_args() convert_blip_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
78
"""simple docstring""" import copy import re class _UpperCAmelCase: lowercase__ = 'hp' lowercase__ = {} lowercase__ = None @classmethod def UpperCAmelCase ( cls , __a , __a) -> Dict: '''simple docstring''' _UpperCamelCase = prefix _UpperCamelCase = defaults cls.build_naming_info() @staticmethod def UpperCAmelCase ( __a , __a) -> Union[str, Any]: '''simple docstring''' if len(__a) == 0: return "" _UpperCamelCase = None if any(char.isdigit() for char in word): raise Exception(F'''Parameters should not contain numbers: \'{word}\' contains a number''') if word in info["short_word"]: return info["short_word"][word] for prefix_len in range(1 , len(__a) + 1): _UpperCamelCase = word[:prefix_len] if prefix in info["reverse_short_word"]: continue else: _UpperCamelCase = prefix break if short_word is None: # Paranoid fallback def int_to_alphabetic(__a): _UpperCamelCase = '''''' while integer != 0: _UpperCamelCase = chr(ord('''A''') + integer % 10) + s integer //= 10 return s _UpperCamelCase = 0 while True: _UpperCamelCase = word + '''#''' + int_to_alphabetic(__a) if sword in info["reverse_short_word"]: continue else: _UpperCamelCase = sword break _UpperCamelCase = short_word _UpperCamelCase = word return short_word @staticmethod def UpperCAmelCase ( __a , __a) -> Any: '''simple docstring''' _UpperCamelCase = param_name.split('''_''') _UpperCamelCase = [TrialShortNamer.shortname_for_word(__a , __a) for word in words] # We try to create a separatorless short name, but if there is a collision we have to fallback # to a separated short name _UpperCamelCase = ['''''', '''_'''] for separator in separators: _UpperCamelCase = separator.join(__a) if shortname not in info["reverse_short_param"]: _UpperCamelCase = shortname _UpperCamelCase = param_name return shortname return param_name @staticmethod def UpperCAmelCase ( __a , __a) -> List[str]: '''simple docstring''' _UpperCamelCase = TrialShortNamer.shortname_for_key(__a , __a) _UpperCamelCase = short_name _UpperCamelCase = param_name @classmethod def UpperCAmelCase ( cls) -> Any: '''simple docstring''' if cls.NAMING_INFO is not None: return _UpperCamelCase = { '''short_word''': {}, '''reverse_short_word''': {}, '''short_param''': {}, '''reverse_short_param''': {}, } _UpperCamelCase = list(cls.DEFAULTS.keys()) for k in field_keys: cls.add_new_param_name(__a , __a) _UpperCamelCase = info @classmethod def UpperCAmelCase ( cls , __a) -> Optional[Any]: '''simple docstring''' cls.build_naming_info() assert cls.PREFIX is not None _UpperCamelCase = [copy.copy(cls.PREFIX)] for k, v in params.items(): if k not in cls.DEFAULTS: raise Exception(F'''You should provide a default value for the param name {k} with value {v}''') if v == cls.DEFAULTS[k]: # The default value is not added to the name continue _UpperCamelCase = cls.NAMING_INFO['''short_param'''][k] if isinstance(__a , __a): _UpperCamelCase = 1 if v else 0 _UpperCamelCase = '''''' if isinstance(__a , (int, float)) else '''-''' _UpperCamelCase = F'''{key}{sep}{v}''' name.append(__a) return "_".join(__a) @classmethod def UpperCAmelCase ( cls , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = repr[len(cls.PREFIX) + 1 :] if repr == "": _UpperCamelCase = [] else: _UpperCamelCase = repr.split('''_''') _UpperCamelCase = {} for value in values: if "-" in value: _UpperCamelCase , _UpperCamelCase = value.split('''-''') else: _UpperCamelCase = re.sub('''[0-9.]''' , '''''' , __a) _UpperCamelCase = float(re.sub('''[^0-9.]''' , '''''' , __a)) _UpperCamelCase = cls.NAMING_INFO['''reverse_short_param'''][p_k] _UpperCamelCase = p_v for k in cls.DEFAULTS: if k not in parameters: _UpperCamelCase = cls.DEFAULTS[k] return parameters
78
1
"""simple docstring""" import argparse import os.path as osp import re import torch from safetensors.torch import load_file, save_file # =================# # UNet Conversion # # =================# _a = [ # (stable-diffusion, HF Diffusers) ("""time_embed.0.weight""", """time_embedding.linear_1.weight"""), ("""time_embed.0.bias""", """time_embedding.linear_1.bias"""), ("""time_embed.2.weight""", """time_embedding.linear_2.weight"""), ("""time_embed.2.bias""", """time_embedding.linear_2.bias"""), ("""input_blocks.0.0.weight""", """conv_in.weight"""), ("""input_blocks.0.0.bias""", """conv_in.bias"""), ("""out.0.weight""", """conv_norm_out.weight"""), ("""out.0.bias""", """conv_norm_out.bias"""), ("""out.2.weight""", """conv_out.weight"""), ("""out.2.bias""", """conv_out.bias"""), ] _a = [ # (stable-diffusion, HF Diffusers) ("""in_layers.0""", """norm1"""), ("""in_layers.2""", """conv1"""), ("""out_layers.0""", """norm2"""), ("""out_layers.3""", """conv2"""), ("""emb_layers.1""", """time_emb_proj"""), ("""skip_connection""", """conv_shortcut"""), ] _a = [] # hardcoded number of downblocks and resnets/attentions... # would need smarter logic for other networks. for i in range(4): # loop over downblocks/upblocks for j in range(2): # loop over resnets/attentions for downblocks _a = F"""down_blocks.{i}.resnets.{j}.""" _a = F"""input_blocks.{3*i + j + 1}.0.""" unet_conversion_map_layer.append((sd_down_res_prefix, hf_down_res_prefix)) if i < 3: # no attention layers in down_blocks.3 _a = F"""down_blocks.{i}.attentions.{j}.""" _a = F"""input_blocks.{3*i + j + 1}.1.""" unet_conversion_map_layer.append((sd_down_atn_prefix, hf_down_atn_prefix)) for j in range(3): # loop over resnets/attentions for upblocks _a = F"""up_blocks.{i}.resnets.{j}.""" _a = F"""output_blocks.{3*i + j}.0.""" unet_conversion_map_layer.append((sd_up_res_prefix, hf_up_res_prefix)) if i > 0: # no attention layers in up_blocks.0 _a = F"""up_blocks.{i}.attentions.{j}.""" _a = F"""output_blocks.{3*i + j}.1.""" unet_conversion_map_layer.append((sd_up_atn_prefix, hf_up_atn_prefix)) if i < 3: # no downsample in down_blocks.3 _a = F"""down_blocks.{i}.downsamplers.0.conv.""" _a = F"""input_blocks.{3*(i+1)}.0.op.""" unet_conversion_map_layer.append((sd_downsample_prefix, hf_downsample_prefix)) # no upsample in up_blocks.3 _a = F"""up_blocks.{i}.upsamplers.0.""" _a = F"""output_blocks.{3*i + 2}.{1 if i == 0 else 2}.""" unet_conversion_map_layer.append((sd_upsample_prefix, hf_upsample_prefix)) _a = """mid_block.attentions.0.""" _a = """middle_block.1.""" unet_conversion_map_layer.append((sd_mid_atn_prefix, hf_mid_atn_prefix)) for j in range(2): _a = F"""mid_block.resnets.{j}.""" _a = F"""middle_block.{2*j}.""" unet_conversion_map_layer.append((sd_mid_res_prefix, hf_mid_res_prefix)) def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" _UpperCamelCase = {k: k for k in unet_state_dict.keys()} for sd_name, hf_name in unet_conversion_map: _UpperCamelCase = sd_name for k, v in mapping.items(): if "resnets" in k: for sd_part, hf_part in unet_conversion_map_resnet: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v for k, v in mapping.items(): for sd_part, hf_part in unet_conversion_map_layer: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v _UpperCamelCase = {v: unet_state_dict[k] for k, v in mapping.items()} return new_state_dict # ================# # VAE Conversion # # ================# _a = [ # (stable-diffusion, HF Diffusers) ("""nin_shortcut""", """conv_shortcut"""), ("""norm_out""", """conv_norm_out"""), ("""mid.attn_1.""", """mid_block.attentions.0."""), ] for i in range(4): # down_blocks have two resnets for j in range(2): _a = F"""encoder.down_blocks.{i}.resnets.{j}.""" _a = F"""encoder.down.{i}.block.{j}.""" vae_conversion_map.append((sd_down_prefix, hf_down_prefix)) if i < 3: _a = F"""down_blocks.{i}.downsamplers.0.""" _a = F"""down.{i}.downsample.""" vae_conversion_map.append((sd_downsample_prefix, hf_downsample_prefix)) _a = F"""up_blocks.{i}.upsamplers.0.""" _a = F"""up.{3-i}.upsample.""" vae_conversion_map.append((sd_upsample_prefix, hf_upsample_prefix)) # up_blocks have three resnets # also, up blocks in hf are numbered in reverse from sd for j in range(3): _a = F"""decoder.up_blocks.{i}.resnets.{j}.""" _a = F"""decoder.up.{3-i}.block.{j}.""" vae_conversion_map.append((sd_up_prefix, hf_up_prefix)) # this part accounts for mid blocks in both the encoder and the decoder for i in range(2): _a = F"""mid_block.resnets.{i}.""" _a = F"""mid.block_{i+1}.""" vae_conversion_map.append((sd_mid_res_prefix, hf_mid_res_prefix)) _a = [ # (stable-diffusion, HF Diffusers) ("""norm.""", """group_norm."""), ("""q.""", """query."""), ("""k.""", """key."""), ("""v.""", """value."""), ("""proj_out.""", """proj_attn."""), ] def lowerCamelCase__ ( __snake_case ) -> List[str]: """simple docstring""" return w.reshape(*w.shape, 1, 1 ) def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" _UpperCamelCase = {k: k for k in vae_state_dict.keys()} for k, v in mapping.items(): for sd_part, hf_part in vae_conversion_map: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v for k, v in mapping.items(): if "attentions" in k: for sd_part, hf_part in vae_conversion_map_attn: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v _UpperCamelCase = {v: vae_state_dict[k] for k, v in mapping.items()} _UpperCamelCase = ['''q''', '''k''', '''v''', '''proj_out'''] for k, v in new_state_dict.items(): for weight_name in weights_to_convert: if F'''mid.attn_1.{weight_name}.weight''' in k: print(F'''Reshaping {k} for SD format''' ) _UpperCamelCase = reshape_weight_for_sd(__snake_case ) return new_state_dict # =========================# # Text Encoder Conversion # # =========================# _a = [ # (stable-diffusion, HF Diffusers) ("""resblocks.""", """text_model.encoder.layers."""), ("""ln_1""", """layer_norm1"""), ("""ln_2""", """layer_norm2"""), (""".c_fc.""", """.fc1."""), (""".c_proj.""", """.fc2."""), (""".attn""", """.self_attn"""), ("""ln_final.""", """transformer.text_model.final_layer_norm."""), ("""token_embedding.weight""", """transformer.text_model.embeddings.token_embedding.weight"""), ("""positional_embedding""", """transformer.text_model.embeddings.position_embedding.weight"""), ] _a = {re.escape(x[1]): x[0] for x in textenc_conversion_lst} _a = re.compile("""|""".join(protected.keys())) # Ordering is from https://github.com/pytorch/pytorch/blob/master/test/cpp/api/modules.cpp _a = {"""q""": 0, """k""": 1, """v""": 2} def lowerCamelCase__ ( __snake_case ) -> Any: """simple docstring""" _UpperCamelCase = {} _UpperCamelCase = {} _UpperCamelCase = {} for k, v in text_enc_dict.items(): if ( k.endswith('''.self_attn.q_proj.weight''' ) or k.endswith('''.self_attn.k_proj.weight''' ) or k.endswith('''.self_attn.v_proj.weight''' ) ): _UpperCamelCase = k[: -len('''.q_proj.weight''' )] _UpperCamelCase = k[-len('''q_proj.weight''' )] if k_pre not in capture_qkv_weight: _UpperCamelCase = [None, None, None] _UpperCamelCase = v continue if ( k.endswith('''.self_attn.q_proj.bias''' ) or k.endswith('''.self_attn.k_proj.bias''' ) or k.endswith('''.self_attn.v_proj.bias''' ) ): _UpperCamelCase = k[: -len('''.q_proj.bias''' )] _UpperCamelCase = k[-len('''q_proj.bias''' )] if k_pre not in capture_qkv_bias: _UpperCamelCase = [None, None, None] _UpperCamelCase = v continue _UpperCamelCase = textenc_pattern.sub(lambda __snake_case : protected[re.escape(m.group(0 ) )], __snake_case ) _UpperCamelCase = v for k_pre, tensors in capture_qkv_weight.items(): if None in tensors: raise Exception('''CORRUPTED MODEL: one of the q-k-v values for the text encoder was missing''' ) _UpperCamelCase = textenc_pattern.sub(lambda __snake_case : protected[re.escape(m.group(0 ) )], __snake_case ) _UpperCamelCase = torch.cat(__snake_case ) for k_pre, tensors in capture_qkv_bias.items(): if None in tensors: raise Exception('''CORRUPTED MODEL: one of the q-k-v values for the text encoder was missing''' ) _UpperCamelCase = textenc_pattern.sub(lambda __snake_case : protected[re.escape(m.group(0 ) )], __snake_case ) _UpperCamelCase = torch.cat(__snake_case ) return new_state_dict def lowerCamelCase__ ( __snake_case ) -> Tuple: """simple docstring""" return text_enc_dict if __name__ == "__main__": _a = argparse.ArgumentParser() parser.add_argument("""--model_path""", default=None, type=str, required=True, help="""Path to the model to convert.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, required=True, help="""Path to the output model.""") parser.add_argument("""--half""", action="""store_true""", help="""Save weights in half precision.""") parser.add_argument( """--use_safetensors""", action="""store_true""", help="""Save weights use safetensors, default is ckpt.""" ) _a = parser.parse_args() assert args.model_path is not None, "Must provide a model path!" assert args.checkpoint_path is not None, "Must provide a checkpoint path!" # Path for safetensors _a = osp.join(args.model_path, """unet""", """diffusion_pytorch_model.safetensors""") _a = osp.join(args.model_path, """vae""", """diffusion_pytorch_model.safetensors""") _a = osp.join(args.model_path, """text_encoder""", """model.safetensors""") # Load models from safetensors if it exists, if it doesn't pytorch if osp.exists(unet_path): _a = load_file(unet_path, device="""cpu""") else: _a = osp.join(args.model_path, """unet""", """diffusion_pytorch_model.bin""") _a = torch.load(unet_path, map_location="""cpu""") if osp.exists(vae_path): _a = load_file(vae_path, device="""cpu""") else: _a = osp.join(args.model_path, """vae""", """diffusion_pytorch_model.bin""") _a = torch.load(vae_path, map_location="""cpu""") if osp.exists(text_enc_path): _a = load_file(text_enc_path, device="""cpu""") else: _a = osp.join(args.model_path, """text_encoder""", """pytorch_model.bin""") _a = torch.load(text_enc_path, map_location="""cpu""") # Convert the UNet model _a = convert_unet_state_dict(unet_state_dict) _a = {"""model.diffusion_model.""" + k: v for k, v in unet_state_dict.items()} # Convert the VAE model _a = convert_vae_state_dict(vae_state_dict) _a = {"""first_stage_model.""" + k: v for k, v in vae_state_dict.items()} # Easiest way to identify v2.0 model seems to be that the text encoder (OpenCLIP) is deeper _a = """text_model.encoder.layers.22.layer_norm2.bias""" in text_enc_dict if is_vaa_model: # Need to add the tag 'transformer' in advance so we can knock it out from the final layer-norm _a = {"""transformer.""" + k: v for k, v in text_enc_dict.items()} _a = convert_text_enc_state_dict_vaa(text_enc_dict) _a = {"""cond_stage_model.model.""" + k: v for k, v in text_enc_dict.items()} else: _a = convert_text_enc_state_dict(text_enc_dict) _a = {"""cond_stage_model.transformer.""" + k: v for k, v in text_enc_dict.items()} # Put together new checkpoint _a = {**unet_state_dict, **vae_state_dict, **text_enc_dict} if args.half: _a = {k: v.half() for k, v in state_dict.items()} if args.use_safetensors: save_file(state_dict, args.checkpoint_path) else: _a = {"""state_dict""": state_dict} torch.save(state_dict, args.checkpoint_path)
78
"""simple docstring""" import os import time import pytest from datasets.utils.filelock import FileLock, Timeout def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = FileLock(str(tmpdir / '''foo.lock''' ) ) _UpperCamelCase = FileLock(str(tmpdir / '''foo.lock''' ) ) _UpperCamelCase = 0.01 with locka.acquire(): with pytest.raises(__snake_case ): _UpperCamelCase = time.time() locka.acquire(__snake_case ) assert time.time() - _start > timeout def lowerCamelCase__ ( __snake_case ) -> List[Any]: """simple docstring""" _UpperCamelCase = '''a''' * 10_00 + '''.lock''' _UpperCamelCase = FileLock(str(tmpdir / filename ) ) assert locka._lock_file.endswith('''.lock''' ) assert not locka._lock_file.endswith(__snake_case ) assert len(os.path.basename(locka._lock_file ) ) <= 2_55 _UpperCamelCase = FileLock(tmpdir / filename ) with locka.acquire(): with pytest.raises(__snake_case ): locka.acquire(0 )
78
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) _a = { """configuration_funnel""": ["""FUNNEL_PRETRAINED_CONFIG_ARCHIVE_MAP""", """FunnelConfig"""], """convert_funnel_original_tf_checkpoint_to_pytorch""": [], """tokenization_funnel""": ["""FunnelTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ["""FunnelTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ """FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST""", """FunnelBaseModel""", """FunnelForMaskedLM""", """FunnelForMultipleChoice""", """FunnelForPreTraining""", """FunnelForQuestionAnswering""", """FunnelForSequenceClassification""", """FunnelForTokenClassification""", """FunnelModel""", """FunnelPreTrainedModel""", """load_tf_weights_in_funnel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ """TF_FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFFunnelBaseModel""", """TFFunnelForMaskedLM""", """TFFunnelForMultipleChoice""", """TFFunnelForPreTraining""", """TFFunnelForQuestionAnswering""", """TFFunnelForSequenceClassification""", """TFFunnelForTokenClassification""", """TFFunnelModel""", """TFFunnelPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_funnel import FUNNEL_PRETRAINED_CONFIG_ARCHIVE_MAP, FunnelConfig from .tokenization_funnel import FunnelTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_funnel_fast import FunnelTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_funnel import ( FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST, FunnelBaseModel, FunnelForMaskedLM, FunnelForMultipleChoice, FunnelForPreTraining, FunnelForQuestionAnswering, FunnelForSequenceClassification, FunnelForTokenClassification, FunnelModel, FunnelPreTrainedModel, load_tf_weights_in_funnel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_funnel import ( TF_FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST, TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, TFFunnelPreTrainedModel, ) else: import sys _a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
78
"""simple docstring""" from math import sqrt def lowerCamelCase__ ( __snake_case ) -> bool: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and ( number >= 0 ), "'number' must been an int and positive" _UpperCamelCase = True # 0 and 1 are none primes. if number <= 1: _UpperCamelCase = False for divisor in range(2, int(round(sqrt(__snake_case ) ) ) + 1 ): # if 'number' divisible by 'divisor' then sets 'status' # of false and break up the loop. if number % divisor == 0: _UpperCamelCase = False break # precondition assert isinstance(__snake_case, __snake_case ), "'status' must been from type bool" return status def lowerCamelCase__ ( __snake_case ) -> Dict: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n > 2), "'N' must been an int and > 2" # beginList: contains all natural numbers from 2 up to N _UpperCamelCase = list(range(2, n + 1 ) ) _UpperCamelCase = [] # this list will be returns. # actual sieve of erathostenes for i in range(len(__snake_case ) ): for j in range(i + 1, len(__snake_case ) ): if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0): _UpperCamelCase = 0 # filters actual prime numbers. _UpperCamelCase = [x for x in begin_list if x != 0] # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type list" return ans def lowerCamelCase__ ( __snake_case ) -> List[str]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n > 2), "'N' must been an int and > 2" _UpperCamelCase = [] # iterates over all numbers between 2 up to N+1 # if a number is prime then appends to list 'ans' for number in range(2, n + 1 ): if is_prime(__snake_case ): ans.append(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type list" return ans def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and number >= 0, "'number' must been an int and >= 0" _UpperCamelCase = [] # this list will be returns of the function. # potential prime number factors. _UpperCamelCase = 2 _UpperCamelCase = number if number == 0 or number == 1: ans.append(__snake_case ) # if 'number' not prime then builds the prime factorization of 'number' elif not is_prime(__snake_case ): while quotient != 1: if is_prime(__snake_case ) and (quotient % factor == 0): ans.append(__snake_case ) quotient /= factor else: factor += 1 else: ans.append(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type list" return ans def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and ( number >= 0 ), "'number' bust been an int and >= 0" _UpperCamelCase = 0 # prime factorization of 'number' _UpperCamelCase = prime_factorization(__snake_case ) _UpperCamelCase = max(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type int" return ans def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and ( number >= 0 ), "'number' bust been an int and >= 0" _UpperCamelCase = 0 # prime factorization of 'number' _UpperCamelCase = prime_factorization(__snake_case ) _UpperCamelCase = min(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type int" return ans def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ), "'number' must been an int" assert isinstance(number % 2 == 0, __snake_case ), "compare bust been from type bool" return number % 2 == 0 def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ), "'number' must been an int" assert isinstance(number % 2 != 0, __snake_case ), "compare bust been from type bool" return number % 2 != 0 def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" assert ( isinstance(__snake_case, __snake_case ) and (number > 2) and is_even(__snake_case ) ), "'number' must been an int, even and > 2" _UpperCamelCase = [] # this list will returned # creates a list of prime numbers between 2 up to 'number' _UpperCamelCase = get_prime_numbers(__snake_case ) _UpperCamelCase = len(__snake_case ) # run variable for while-loops. _UpperCamelCase = 0 _UpperCamelCase = None # exit variable. for break up the loops _UpperCamelCase = True while i < len_pn and loop: _UpperCamelCase = i + 1 while j < len_pn and loop: if prime_numbers[i] + prime_numbers[j] == number: _UpperCamelCase = False ans.append(prime_numbers[i] ) ans.append(prime_numbers[j] ) j += 1 i += 1 # precondition assert ( isinstance(__snake_case, __snake_case ) and (len(__snake_case ) == 2) and (ans[0] + ans[1] == number) and is_prime(ans[0] ) and is_prime(ans[1] ) ), "'ans' must contains two primes. And sum of elements must been eq 'number'" return ans def lowerCamelCase__ ( __snake_case, __snake_case ) -> str: """simple docstring""" assert ( isinstance(__snake_case, __snake_case ) and isinstance(__snake_case, __snake_case ) and (numbera >= 0) and (numbera >= 0) ), "'number1' and 'number2' must been positive integer." _UpperCamelCase = 0 while numbera != 0: _UpperCamelCase = numbera % numbera _UpperCamelCase = numbera _UpperCamelCase = rest # precondition assert isinstance(__snake_case, __snake_case ) and ( numbera >= 0 ), "'number' must been from type int and positive" return numbera def lowerCamelCase__ ( __snake_case, __snake_case ) -> List[str]: """simple docstring""" assert ( isinstance(__snake_case, __snake_case ) and isinstance(__snake_case, __snake_case ) and (numbera >= 1) and (numbera >= 1) ), "'number1' and 'number2' must been positive integer." _UpperCamelCase = 1 # actual answer that will be return. # for kgV (x,1) if numbera > 1 and numbera > 1: # builds the prime factorization of 'number1' and 'number2' _UpperCamelCase = prime_factorization(__snake_case ) _UpperCamelCase = prime_factorization(__snake_case ) elif numbera == 1 or numbera == 1: _UpperCamelCase = [] _UpperCamelCase = [] _UpperCamelCase = max(__snake_case, __snake_case ) _UpperCamelCase = 0 _UpperCamelCase = 0 _UpperCamelCase = [] # captured numbers int both 'primeFac1' and 'primeFac2' # iterates through primeFac1 for n in prime_fac_a: if n not in done: if n in prime_fac_a: _UpperCamelCase = prime_fac_a.count(__snake_case ) _UpperCamelCase = prime_fac_a.count(__snake_case ) for _ in range(max(__snake_case, __snake_case ) ): ans *= n else: _UpperCamelCase = prime_fac_a.count(__snake_case ) for _ in range(__snake_case ): ans *= n done.append(__snake_case ) # iterates through primeFac2 for n in prime_fac_a: if n not in done: _UpperCamelCase = prime_fac_a.count(__snake_case ) for _ in range(__snake_case ): ans *= n done.append(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ) and ( ans >= 0 ), "'ans' must been from type int and positive" return ans def lowerCamelCase__ ( __snake_case ) -> Tuple: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n >= 0), "'number' must been a positive int" _UpperCamelCase = 0 _UpperCamelCase = 2 # this variable holds the answer while index < n: index += 1 ans += 1 # counts to the next number # if ans not prime then # runs to the next prime number. while not is_prime(__snake_case ): ans += 1 # precondition assert isinstance(__snake_case, __snake_case ) and is_prime( __snake_case ), "'ans' must been a prime number and from type int" return ans def lowerCamelCase__ ( __snake_case, __snake_case ) -> Tuple: """simple docstring""" assert ( is_prime(__snake_case ) and is_prime(__snake_case ) and (p_number_a < p_number_a) ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'" _UpperCamelCase = p_number_a + 1 # jump to the next number _UpperCamelCase = [] # this list will be returns. # if number is not prime then # fetch the next prime number. while not is_prime(__snake_case ): number += 1 while number < p_number_a: ans.append(__snake_case ) number += 1 # fetch the next prime number. while not is_prime(__snake_case ): number += 1 # precondition assert ( isinstance(__snake_case, __snake_case ) and ans[0] != p_number_a and ans[len(__snake_case ) - 1] != p_number_a ), "'ans' must been a list without the arguments" # 'ans' contains not 'pNumber1' and 'pNumber2' ! return ans def lowerCamelCase__ ( __snake_case ) -> List[Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n >= 1), "'n' must been int and >= 1" _UpperCamelCase = [] # will be returned. for divisor in range(1, n + 1 ): if n % divisor == 0: ans.append(__snake_case ) # precondition assert ans[0] == 1 and ans[len(__snake_case ) - 1] == n, "Error in function getDivisiors(...)" return ans def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and ( number > 1 ), "'number' must been an int and >= 1" _UpperCamelCase = get_divisors(__snake_case ) # precondition assert ( isinstance(__snake_case, __snake_case ) and (divisors[0] == 1) and (divisors[len(__snake_case ) - 1] == number) ), "Error in help-function getDivisiors(...)" # summed all divisors up to 'number' (exclusive), hence [:-1] return sum(divisors[:-1] ) == number def lowerCamelCase__ ( __snake_case, __snake_case ) -> Optional[Any]: """simple docstring""" assert ( isinstance(__snake_case, __snake_case ) and isinstance(__snake_case, __snake_case ) and (denominator != 0) ), "The arguments must been from type int and 'denominator' != 0" # build the greatest common divisor of numerator and denominator. _UpperCamelCase = gcd(abs(__snake_case ), abs(__snake_case ) ) # precondition assert ( isinstance(__snake_case, __snake_case ) and (numerator % gcd_of_fraction == 0) and (denominator % gcd_of_fraction == 0) ), "Error in function gcd(...,...)" return (numerator // gcd_of_fraction, denominator // gcd_of_fraction) def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n >= 0), "'n' must been a int and >= 0" _UpperCamelCase = 1 # this will be return. for factor in range(1, n + 1 ): ans *= factor return ans def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n >= 0), "'n' must been an int and >= 0" _UpperCamelCase = 0 _UpperCamelCase = 1 _UpperCamelCase = 1 # this will be return for _ in range(n - 1 ): _UpperCamelCase = ans ans += fiba _UpperCamelCase = tmp return ans
78
1
"""simple docstring""" import importlib import os import sys # This is required to make the module import works (when the python process is running from the root of the repo) sys.path.append(""".""") def lowerCamelCase__ ( __snake_case ) -> Tuple: """simple docstring""" _UpperCamelCase = test_file.split(os.path.sep ) if components[0:2] != ["tests", "models"]: raise ValueError( '''`test_file` should start with `tests/models/` (with `/` being the OS specific path separator). Got ''' F'''{test_file} instead.''' ) _UpperCamelCase = components[-1] if not test_fn.endswith('''py''' ): raise ValueError(F'''`test_file` should be a python file. Got {test_fn} instead.''' ) if not test_fn.startswith('''test_modeling_''' ): raise ValueError( F'''`test_file` should point to a file name of the form `test_modeling_*.py`. Got {test_fn} instead.''' ) _UpperCamelCase = components[:-1] + [test_fn.replace('''.py''', '''''' )] _UpperCamelCase = '''.'''.join(__snake_case ) return test_module_path def lowerCamelCase__ ( __snake_case ) -> List[str]: """simple docstring""" _UpperCamelCase = get_module_path(__snake_case ) _UpperCamelCase = importlib.import_module(__snake_case ) return test_module def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" _UpperCamelCase = [] _UpperCamelCase = get_test_module(__snake_case ) for attr in dir(__snake_case ): if attr.endswith('''ModelTester''' ): tester_classes.append(getattr(__snake_case, __snake_case ) ) # sort with class names return sorted(__snake_case, key=lambda __snake_case : x.__name__ ) def lowerCamelCase__ ( __snake_case ) -> Dict: """simple docstring""" _UpperCamelCase = [] _UpperCamelCase = get_test_module(__snake_case ) for attr in dir(__snake_case ): _UpperCamelCase = getattr(__snake_case, __snake_case ) # (TF/Flax)ModelTesterMixin is also an attribute in specific model test module. Let's exclude them by checking # `all_model_classes` is not empty (which also excludes other special classes). _UpperCamelCase = getattr(__snake_case, '''all_model_classes''', [] ) if len(__snake_case ) > 0: test_classes.append(__snake_case ) # sort with class names return sorted(__snake_case, key=lambda __snake_case : x.__name__ ) def lowerCamelCase__ ( __snake_case ) -> Any: """simple docstring""" _UpperCamelCase = get_test_classes(__snake_case ) _UpperCamelCase = set() for test_class in test_classes: model_classes.update(test_class.all_model_classes ) # sort with class names return sorted(__snake_case, key=lambda __snake_case : x.__name__ ) def lowerCamelCase__ ( __snake_case ) -> Any: """simple docstring""" _UpperCamelCase = test_class() if hasattr(__snake_case, '''setUp''' ): test.setUp() _UpperCamelCase = None if hasattr(__snake_case, '''model_tester''' ): # `(TF/Flax)ModelTesterMixin` has this attribute default to `None`. Let's skip this case. if test.model_tester is not None: _UpperCamelCase = test.model_tester.__class__ return model_tester def lowerCamelCase__ ( __snake_case, __snake_case ) -> Any: """simple docstring""" _UpperCamelCase = get_test_classes(__snake_case ) _UpperCamelCase = [] for test_class in test_classes: if model_class in test_class.all_model_classes: target_test_classes.append(__snake_case ) # sort with class names return sorted(__snake_case, key=lambda __snake_case : x.__name__ ) def lowerCamelCase__ ( __snake_case, __snake_case ) -> List[str]: """simple docstring""" _UpperCamelCase = get_test_classes_for_model(__snake_case, __snake_case ) _UpperCamelCase = [] for test_class in test_classes: _UpperCamelCase = get_model_tester_from_test_class(__snake_case ) if tester_class is not None: tester_classes.append(__snake_case ) # sort with class names return sorted(__snake_case, key=lambda __snake_case : x.__name__ ) def lowerCamelCase__ ( __snake_case ) -> Tuple: """simple docstring""" _UpperCamelCase = get_test_classes(__snake_case ) _UpperCamelCase = {test_class: get_model_tester_from_test_class(__snake_case ) for test_class in test_classes} return test_tester_mapping def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" _UpperCamelCase = get_model_classes(__snake_case ) _UpperCamelCase = { model_class: get_test_classes_for_model(__snake_case, __snake_case ) for model_class in model_classes } return model_test_mapping def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" _UpperCamelCase = get_model_classes(__snake_case ) _UpperCamelCase = { model_class: get_tester_classes_for_model(__snake_case, __snake_case ) for model_class in model_classes } return model_to_tester_mapping def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" if isinstance(__snake_case, __snake_case ): return o elif isinstance(__snake_case, __snake_case ): return o.__name__ elif isinstance(__snake_case, (list, tuple) ): return [to_json(__snake_case ) for x in o] elif isinstance(__snake_case, __snake_case ): return {to_json(__snake_case ): to_json(__snake_case ) for k, v in o.items()} else: return o
78
"""simple docstring""" import logging from dataclasses import dataclass, field from pathlib import Path from typing import Optional, Union from .generation.configuration_utils import GenerationConfig from .training_args import TrainingArguments from .utils import add_start_docstrings _a = logging.getLogger(__name__) @dataclass @add_start_docstrings(TrainingArguments.__doc__ ) class _UpperCAmelCase( lowerCamelCase ): lowercase__ = field(default=lowerCamelCase , metadata={'help': 'Whether to use SortishSampler or not.'} ) lowercase__ = field( default=lowerCamelCase , metadata={'help': 'Whether to use generate to calculate generative metrics (ROUGE, BLEU).'} ) lowercase__ = field( default=lowerCamelCase , metadata={ 'help': ( 'The `max_length` to use on each evaluation loop when `predict_with_generate=True`. Will default ' 'to the `max_length` value of the model configuration.' ) } , ) lowercase__ = field( default=lowerCamelCase , metadata={ 'help': ( 'The `num_beams` to use on each evaluation loop when `predict_with_generate=True`. Will default ' 'to the `num_beams` value of the model configuration.' ) } , ) lowercase__ = field( default=lowerCamelCase , metadata={ 'help': 'Model id, file path or url pointing to a GenerationConfig json file, to use during prediction.' } , ) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = super().to_dict() for k, v in d.items(): if isinstance(__a , __a): _UpperCamelCase = v.to_dict() return d
78
1
"""simple docstring""" import sacrebleu as scb from packaging import version from sacrebleu import TER import datasets _a = """\ @inproceedings{snover-etal-2006-study, title = \"A Study of Translation Edit Rate with Targeted Human Annotation\", author = \"Snover, Matthew and Dorr, Bonnie and Schwartz, Rich and Micciulla, Linnea and Makhoul, John\", booktitle = \"Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers\", month = aug # \" 8-12\", year = \"2006\", address = \"Cambridge, Massachusetts, USA\", publisher = \"Association for Machine Translation in the Americas\", url = \"https://aclanthology.org/2006.amta-papers.25\", pages = \"223--231\", } @inproceedings{post-2018-call, title = \"A Call for Clarity in Reporting {BLEU} Scores\", author = \"Post, Matt\", booktitle = \"Proceedings of the Third Conference on Machine Translation: Research Papers\", month = oct, year = \"2018\", address = \"Belgium, Brussels\", publisher = \"Association for Computational Linguistics\", url = \"https://www.aclweb.org/anthology/W18-6319\", pages = \"186--191\", } """ _a = """\ TER (Translation Edit Rate, also called Translation Error Rate) is a metric to quantify the edit operations that a hypothesis requires to match a reference translation. We use the implementation that is already present in sacrebleu (https://github.com/mjpost/sacreBLEU#ter), which in turn is inspired by the TERCOM implementation, which can be found here: https://github.com/jhclark/tercom. The implementation here is slightly different from sacrebleu in terms of the required input format. The length of the references and hypotheses lists need to be the same, so you may need to transpose your references compared to sacrebleu's required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534 See the README.md file at https://github.com/mjpost/sacreBLEU#ter for more information. """ _a = """ Produces TER scores alongside the number of edits and reference length. Args: predictions (list of str): The system stream (a sequence of segments). references (list of list of str): A list of one or more reference streams (each a sequence of segments). normalized (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`. ignore_punct (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`. support_zh_ja_chars (boolean): If `True`, tokenization/normalization supports processing of Chinese characters, as well as Japanese Kanji, Hiragana, Katakana, and Phonetic Extensions of Katakana. Only applies if `normalized = True`. Defaults to `False`. case_sensitive (boolean): If `False`, makes all predictions and references lowercase to ignore differences in case. Defaults to `False`. Returns: 'score' (float): TER score (num_edits / sum_ref_lengths * 100) 'num_edits' (int): The cumulative number of edits 'ref_length' (float): The cumulative average reference length Examples: Example 1: >>> predictions = [\"does this sentence match??\", ... \"what about this sentence?\", ... \"What did the TER metric user say to the developer?\"] >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"], ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"], ... [\"Your jokes are...\", \"...TERrible\"]] >>> ter = datasets.load_metric(\"ter\") >>> results = ter.compute(predictions=predictions, ... references=references, ... case_sensitive=True) >>> print(results) {'score': 150.0, 'num_edits': 15, 'ref_length': 10.0} Example 2: >>> predictions = [\"does this sentence match??\", ... \"what about this sentence?\"] >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"], ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"]] >>> ter = datasets.load_metric(\"ter\") >>> results = ter.compute(predictions=predictions, ... references=references, ... case_sensitive=True) >>> print(results) {'score': 62.5, 'num_edits': 5, 'ref_length': 8.0} Example 3: >>> predictions = [\"does this sentence match??\", ... \"what about this sentence?\"] >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"], ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"]] >>> ter = datasets.load_metric(\"ter\") >>> results = ter.compute(predictions=predictions, ... references=references, ... normalized=True, ... case_sensitive=True) >>> print(results) {'score': 57.14285714285714, 'num_edits': 6, 'ref_length': 10.5} Example 4: >>> predictions = [\"does this sentence match??\", ... \"what about this sentence?\"] >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"], ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"]] >>> ter = datasets.load_metric(\"ter\") >>> results = ter.compute(predictions=predictions, ... references=references, ... ignore_punct=True, ... case_sensitive=False) >>> print(results) {'score': 0.0, 'num_edits': 0, 'ref_length': 8.0} Example 5: >>> predictions = [\"does this sentence match??\", ... \"what about this sentence?\", ... \"What did the TER metric user say to the developer?\"] >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"], ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"], ... [\"Your jokes are...\", \"...TERrible\"]] >>> ter = datasets.load_metric(\"ter\") >>> results = ter.compute(predictions=predictions, ... references=references, ... ignore_punct=True, ... case_sensitive=False) >>> print(results) {'score': 100.0, 'num_edits': 10, 'ref_length': 10.0} """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _UpperCAmelCase( datasets.Metric ): def UpperCAmelCase ( self) -> int: '''simple docstring''' if version.parse(scb.__version__) < version.parse('''1.4.12'''): raise ImportWarning( '''To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn\'t match this condition.\n''' '''You can install it with `pip install "sacrebleu>=1.4.12"`.''') return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='''http://www.cs.umd.edu/~snover/tercom/''' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence'''), '''references''': datasets.Sequence(datasets.Value('''string''' , id='''sequence''') , id='''references'''), }) , codebase_urls=['''https://github.com/mjpost/sacreBLEU#ter'''] , reference_urls=[ '''https://github.com/jhclark/tercom''', ] , ) def UpperCAmelCase ( self , __a , __a , __a = False , __a = False , __a = False , __a = False , ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = len(references[0]) if any(len(__a) != references_per_prediction for refs in references): raise ValueError('''Sacrebleu requires the same number of references for each prediction''') _UpperCamelCase = [[refs[i] for refs in references] for i in range(__a)] _UpperCamelCase = TER( normalized=__a , no_punct=__a , asian_support=__a , case_sensitive=__a , ) _UpperCamelCase = sb_ter.corpus_score(__a , __a) return {"score": output.score, "num_edits": output.num_edits, "ref_length": output.ref_length}
78
"""simple docstring""" import argparse import torch from transformers import BlenderbotConfig, BlenderbotForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) _a = [ ["""attention""", """attn"""], ["""encoder_attention""", """encoder_attn"""], ["""q_lin""", """q_proj"""], ["""k_lin""", """k_proj"""], ["""v_lin""", """v_proj"""], ["""out_lin""", """out_proj"""], ["""norm_embeddings""", """layernorm_embedding"""], ["""position_embeddings""", """embed_positions"""], ["""embeddings""", """embed_tokens"""], ["""ffn.lin""", """fc"""], ] def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" if k == "embeddings.weight": return "shared.weight" for parlai_name, hf_name in PATTERNS: _UpperCamelCase = k.replace(__snake_case, __snake_case ) if k.startswith('''encoder''' ): _UpperCamelCase = k.replace('''.attn''', '''.self_attn''' ) _UpperCamelCase = k.replace('''norm1''', '''self_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm2''', '''final_layer_norm''' ) elif k.startswith('''decoder''' ): _UpperCamelCase = k.replace('''norm1''', '''self_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm2''', '''encoder_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm3''', '''final_layer_norm''' ) return k def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" _UpperCamelCase = [ '''model.encoder.layernorm_embedding.weight''', '''model.encoder.layernorm_embedding.bias''', '''model.decoder.layernorm_embedding.weight''', '''model.decoder.layernorm_embedding.bias''', ] for k in keys: _UpperCamelCase = sd.pop(__snake_case ) _UpperCamelCase = k.replace('''layernorm_embedding''', '''layer_norm''' ) assert new_k not in sd _UpperCamelCase = v _a = ["""START"""] @torch.no_grad() def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> int: """simple docstring""" _UpperCamelCase = torch.load(__snake_case, map_location='''cpu''' ) _UpperCamelCase = model['''model'''] _UpperCamelCase = BlenderbotConfig.from_json_file(__snake_case ) _UpperCamelCase = BlenderbotForConditionalGeneration(__snake_case ) _UpperCamelCase = m.model.state_dict().keys() _UpperCamelCase = [] _UpperCamelCase = {} for k, v in sd.items(): if k in IGNORE_KEYS: continue _UpperCamelCase = rename_state_dict_key(__snake_case ) if new_k not in valid_keys: failures.append([k, new_k] ) else: _UpperCamelCase = v if cfg.normalize_before: # Blenderbot-3B checkpoints. Rename layernorm_embedding -> layer_norm rename_layernorm_keys(__snake_case ) m.model.load_state_dict(__snake_case, strict=__snake_case ) m.half() m.save_pretrained(__snake_case ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument("""--src_path""", type=str, help="""like blenderbot-model.bin""") parser.add_argument("""--save_dir""", default="""hf_blenderbot""", type=str, help="""Where to save converted model.""") parser.add_argument( """--hf_config_json""", default="""blenderbot-3b-config.json""", type=str, help="""Path to config to use""" ) _a = parser.parse_args() convert_parlai_checkpoint(args.src_path, args.save_dir, args.hf_config_json)
78
1
"""simple docstring""" import qiskit def lowerCamelCase__ ( __snake_case, __snake_case ) -> qiskit.result.counts.Counts: """simple docstring""" _UpperCamelCase = qiskit.Aer.get_backend('''aer_simulator''' ) # Create a Quantum Circuit acting on the q register _UpperCamelCase = qiskit.QuantumCircuit(__snake_case, __snake_case ) # Map the quantum measurement to the classical bits circuit.measure([0], [0] ) # Execute the circuit on the simulator _UpperCamelCase = qiskit.execute(__snake_case, __snake_case, shots=10_00 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(__snake_case ) if __name__ == "__main__": print(F"""Total count for various states are: {single_qubit_measure(1, 1)}""")
78
"""simple docstring""" import argparse import os.path as osp import re import torch from safetensors.torch import load_file, save_file # =================# # UNet Conversion # # =================# _a = [ # (stable-diffusion, HF Diffusers) ("""time_embed.0.weight""", """time_embedding.linear_1.weight"""), ("""time_embed.0.bias""", """time_embedding.linear_1.bias"""), ("""time_embed.2.weight""", """time_embedding.linear_2.weight"""), ("""time_embed.2.bias""", """time_embedding.linear_2.bias"""), ("""input_blocks.0.0.weight""", """conv_in.weight"""), ("""input_blocks.0.0.bias""", """conv_in.bias"""), ("""out.0.weight""", """conv_norm_out.weight"""), ("""out.0.bias""", """conv_norm_out.bias"""), ("""out.2.weight""", """conv_out.weight"""), ("""out.2.bias""", """conv_out.bias"""), ] _a = [ # (stable-diffusion, HF Diffusers) ("""in_layers.0""", """norm1"""), ("""in_layers.2""", """conv1"""), ("""out_layers.0""", """norm2"""), ("""out_layers.3""", """conv2"""), ("""emb_layers.1""", """time_emb_proj"""), ("""skip_connection""", """conv_shortcut"""), ] _a = [] # hardcoded number of downblocks and resnets/attentions... # would need smarter logic for other networks. for i in range(4): # loop over downblocks/upblocks for j in range(2): # loop over resnets/attentions for downblocks _a = F"""down_blocks.{i}.resnets.{j}.""" _a = F"""input_blocks.{3*i + j + 1}.0.""" unet_conversion_map_layer.append((sd_down_res_prefix, hf_down_res_prefix)) if i < 3: # no attention layers in down_blocks.3 _a = F"""down_blocks.{i}.attentions.{j}.""" _a = F"""input_blocks.{3*i + j + 1}.1.""" unet_conversion_map_layer.append((sd_down_atn_prefix, hf_down_atn_prefix)) for j in range(3): # loop over resnets/attentions for upblocks _a = F"""up_blocks.{i}.resnets.{j}.""" _a = F"""output_blocks.{3*i + j}.0.""" unet_conversion_map_layer.append((sd_up_res_prefix, hf_up_res_prefix)) if i > 0: # no attention layers in up_blocks.0 _a = F"""up_blocks.{i}.attentions.{j}.""" _a = F"""output_blocks.{3*i + j}.1.""" unet_conversion_map_layer.append((sd_up_atn_prefix, hf_up_atn_prefix)) if i < 3: # no downsample in down_blocks.3 _a = F"""down_blocks.{i}.downsamplers.0.conv.""" _a = F"""input_blocks.{3*(i+1)}.0.op.""" unet_conversion_map_layer.append((sd_downsample_prefix, hf_downsample_prefix)) # no upsample in up_blocks.3 _a = F"""up_blocks.{i}.upsamplers.0.""" _a = F"""output_blocks.{3*i + 2}.{1 if i == 0 else 2}.""" unet_conversion_map_layer.append((sd_upsample_prefix, hf_upsample_prefix)) _a = """mid_block.attentions.0.""" _a = """middle_block.1.""" unet_conversion_map_layer.append((sd_mid_atn_prefix, hf_mid_atn_prefix)) for j in range(2): _a = F"""mid_block.resnets.{j}.""" _a = F"""middle_block.{2*j}.""" unet_conversion_map_layer.append((sd_mid_res_prefix, hf_mid_res_prefix)) def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" _UpperCamelCase = {k: k for k in unet_state_dict.keys()} for sd_name, hf_name in unet_conversion_map: _UpperCamelCase = sd_name for k, v in mapping.items(): if "resnets" in k: for sd_part, hf_part in unet_conversion_map_resnet: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v for k, v in mapping.items(): for sd_part, hf_part in unet_conversion_map_layer: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v _UpperCamelCase = {v: unet_state_dict[k] for k, v in mapping.items()} return new_state_dict # ================# # VAE Conversion # # ================# _a = [ # (stable-diffusion, HF Diffusers) ("""nin_shortcut""", """conv_shortcut"""), ("""norm_out""", """conv_norm_out"""), ("""mid.attn_1.""", """mid_block.attentions.0."""), ] for i in range(4): # down_blocks have two resnets for j in range(2): _a = F"""encoder.down_blocks.{i}.resnets.{j}.""" _a = F"""encoder.down.{i}.block.{j}.""" vae_conversion_map.append((sd_down_prefix, hf_down_prefix)) if i < 3: _a = F"""down_blocks.{i}.downsamplers.0.""" _a = F"""down.{i}.downsample.""" vae_conversion_map.append((sd_downsample_prefix, hf_downsample_prefix)) _a = F"""up_blocks.{i}.upsamplers.0.""" _a = F"""up.{3-i}.upsample.""" vae_conversion_map.append((sd_upsample_prefix, hf_upsample_prefix)) # up_blocks have three resnets # also, up blocks in hf are numbered in reverse from sd for j in range(3): _a = F"""decoder.up_blocks.{i}.resnets.{j}.""" _a = F"""decoder.up.{3-i}.block.{j}.""" vae_conversion_map.append((sd_up_prefix, hf_up_prefix)) # this part accounts for mid blocks in both the encoder and the decoder for i in range(2): _a = F"""mid_block.resnets.{i}.""" _a = F"""mid.block_{i+1}.""" vae_conversion_map.append((sd_mid_res_prefix, hf_mid_res_prefix)) _a = [ # (stable-diffusion, HF Diffusers) ("""norm.""", """group_norm."""), ("""q.""", """query."""), ("""k.""", """key."""), ("""v.""", """value."""), ("""proj_out.""", """proj_attn."""), ] def lowerCamelCase__ ( __snake_case ) -> List[str]: """simple docstring""" return w.reshape(*w.shape, 1, 1 ) def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" _UpperCamelCase = {k: k for k in vae_state_dict.keys()} for k, v in mapping.items(): for sd_part, hf_part in vae_conversion_map: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v for k, v in mapping.items(): if "attentions" in k: for sd_part, hf_part in vae_conversion_map_attn: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v _UpperCamelCase = {v: vae_state_dict[k] for k, v in mapping.items()} _UpperCamelCase = ['''q''', '''k''', '''v''', '''proj_out'''] for k, v in new_state_dict.items(): for weight_name in weights_to_convert: if F'''mid.attn_1.{weight_name}.weight''' in k: print(F'''Reshaping {k} for SD format''' ) _UpperCamelCase = reshape_weight_for_sd(__snake_case ) return new_state_dict # =========================# # Text Encoder Conversion # # =========================# _a = [ # (stable-diffusion, HF Diffusers) ("""resblocks.""", """text_model.encoder.layers."""), ("""ln_1""", """layer_norm1"""), ("""ln_2""", """layer_norm2"""), (""".c_fc.""", """.fc1."""), (""".c_proj.""", """.fc2."""), (""".attn""", """.self_attn"""), ("""ln_final.""", """transformer.text_model.final_layer_norm."""), ("""token_embedding.weight""", """transformer.text_model.embeddings.token_embedding.weight"""), ("""positional_embedding""", """transformer.text_model.embeddings.position_embedding.weight"""), ] _a = {re.escape(x[1]): x[0] for x in textenc_conversion_lst} _a = re.compile("""|""".join(protected.keys())) # Ordering is from https://github.com/pytorch/pytorch/blob/master/test/cpp/api/modules.cpp _a = {"""q""": 0, """k""": 1, """v""": 2} def lowerCamelCase__ ( __snake_case ) -> Any: """simple docstring""" _UpperCamelCase = {} _UpperCamelCase = {} _UpperCamelCase = {} for k, v in text_enc_dict.items(): if ( k.endswith('''.self_attn.q_proj.weight''' ) or k.endswith('''.self_attn.k_proj.weight''' ) or k.endswith('''.self_attn.v_proj.weight''' ) ): _UpperCamelCase = k[: -len('''.q_proj.weight''' )] _UpperCamelCase = k[-len('''q_proj.weight''' )] if k_pre not in capture_qkv_weight: _UpperCamelCase = [None, None, None] _UpperCamelCase = v continue if ( k.endswith('''.self_attn.q_proj.bias''' ) or k.endswith('''.self_attn.k_proj.bias''' ) or k.endswith('''.self_attn.v_proj.bias''' ) ): _UpperCamelCase = k[: -len('''.q_proj.bias''' )] _UpperCamelCase = k[-len('''q_proj.bias''' )] if k_pre not in capture_qkv_bias: _UpperCamelCase = [None, None, None] _UpperCamelCase = v continue _UpperCamelCase = textenc_pattern.sub(lambda __snake_case : protected[re.escape(m.group(0 ) )], __snake_case ) _UpperCamelCase = v for k_pre, tensors in capture_qkv_weight.items(): if None in tensors: raise Exception('''CORRUPTED MODEL: one of the q-k-v values for the text encoder was missing''' ) _UpperCamelCase = textenc_pattern.sub(lambda __snake_case : protected[re.escape(m.group(0 ) )], __snake_case ) _UpperCamelCase = torch.cat(__snake_case ) for k_pre, tensors in capture_qkv_bias.items(): if None in tensors: raise Exception('''CORRUPTED MODEL: one of the q-k-v values for the text encoder was missing''' ) _UpperCamelCase = textenc_pattern.sub(lambda __snake_case : protected[re.escape(m.group(0 ) )], __snake_case ) _UpperCamelCase = torch.cat(__snake_case ) return new_state_dict def lowerCamelCase__ ( __snake_case ) -> Tuple: """simple docstring""" return text_enc_dict if __name__ == "__main__": _a = argparse.ArgumentParser() parser.add_argument("""--model_path""", default=None, type=str, required=True, help="""Path to the model to convert.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, required=True, help="""Path to the output model.""") parser.add_argument("""--half""", action="""store_true""", help="""Save weights in half precision.""") parser.add_argument( """--use_safetensors""", action="""store_true""", help="""Save weights use safetensors, default is ckpt.""" ) _a = parser.parse_args() assert args.model_path is not None, "Must provide a model path!" assert args.checkpoint_path is not None, "Must provide a checkpoint path!" # Path for safetensors _a = osp.join(args.model_path, """unet""", """diffusion_pytorch_model.safetensors""") _a = osp.join(args.model_path, """vae""", """diffusion_pytorch_model.safetensors""") _a = osp.join(args.model_path, """text_encoder""", """model.safetensors""") # Load models from safetensors if it exists, if it doesn't pytorch if osp.exists(unet_path): _a = load_file(unet_path, device="""cpu""") else: _a = osp.join(args.model_path, """unet""", """diffusion_pytorch_model.bin""") _a = torch.load(unet_path, map_location="""cpu""") if osp.exists(vae_path): _a = load_file(vae_path, device="""cpu""") else: _a = osp.join(args.model_path, """vae""", """diffusion_pytorch_model.bin""") _a = torch.load(vae_path, map_location="""cpu""") if osp.exists(text_enc_path): _a = load_file(text_enc_path, device="""cpu""") else: _a = osp.join(args.model_path, """text_encoder""", """pytorch_model.bin""") _a = torch.load(text_enc_path, map_location="""cpu""") # Convert the UNet model _a = convert_unet_state_dict(unet_state_dict) _a = {"""model.diffusion_model.""" + k: v for k, v in unet_state_dict.items()} # Convert the VAE model _a = convert_vae_state_dict(vae_state_dict) _a = {"""first_stage_model.""" + k: v for k, v in vae_state_dict.items()} # Easiest way to identify v2.0 model seems to be that the text encoder (OpenCLIP) is deeper _a = """text_model.encoder.layers.22.layer_norm2.bias""" in text_enc_dict if is_vaa_model: # Need to add the tag 'transformer' in advance so we can knock it out from the final layer-norm _a = {"""transformer.""" + k: v for k, v in text_enc_dict.items()} _a = convert_text_enc_state_dict_vaa(text_enc_dict) _a = {"""cond_stage_model.model.""" + k: v for k, v in text_enc_dict.items()} else: _a = convert_text_enc_state_dict(text_enc_dict) _a = {"""cond_stage_model.transformer.""" + k: v for k, v in text_enc_dict.items()} # Put together new checkpoint _a = {**unet_state_dict, **vae_state_dict, **text_enc_dict} if args.half: _a = {k: v.half() for k, v in state_dict.items()} if args.use_safetensors: save_file(state_dict, args.checkpoint_path) else: _a = {"""state_dict""": state_dict} torch.save(state_dict, args.checkpoint_path)
78
1
"""simple docstring""" from __future__ import annotations def lowerCamelCase__ ( __snake_case ) -> float: """simple docstring""" if not nums: raise ValueError('''List is empty''' ) return sum(__snake_case ) / len(__snake_case ) if __name__ == "__main__": import doctest doctest.testmod()
78
"""simple docstring""" import argparse import torch from transformers import OpenAIGPTConfig, OpenAIGPTModel, load_tf_weights_in_openai_gpt from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> Union[str, Any]: """simple docstring""" if openai_config_file == "": _UpperCamelCase = OpenAIGPTConfig() else: _UpperCamelCase = OpenAIGPTConfig.from_json_file(__snake_case ) _UpperCamelCase = OpenAIGPTModel(__snake_case ) # Load weights from numpy load_tf_weights_in_openai_gpt(__snake_case, __snake_case, __snake_case ) # Save pytorch-model _UpperCamelCase = pytorch_dump_folder_path + '''/''' + WEIGHTS_NAME _UpperCamelCase = pytorch_dump_folder_path + '''/''' + CONFIG_NAME print(F'''Save PyTorch model to {pytorch_weights_dump_path}''' ) torch.save(model.state_dict(), __snake_case ) print(F'''Save configuration file to {pytorch_config_dump_path}''' ) with open(__snake_case, '''w''', encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( """--openai_checkpoint_folder_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--openai_config_file""", default="""""", type=str, help=( """An optional config json file corresponding to the pre-trained OpenAI model. \n""" """This specifies the model architecture.""" ), ) _a = parser.parse_args() convert_openai_checkpoint_to_pytorch( args.openai_checkpoint_folder_path, args.openai_config_file, args.pytorch_dump_folder_path )
78
1
"""simple docstring""" from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar _a = TypeVar("""T""") class _UpperCAmelCase( Generic[T] ): def __init__( self , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = data _UpperCamelCase = None def __str__( self) -> str: '''simple docstring''' return F'''{self.data}''' class _UpperCAmelCase( Generic[T] ): def __init__( self) -> None: '''simple docstring''' _UpperCamelCase = None def __iter__( self) -> Iterator[T]: '''simple docstring''' _UpperCamelCase = self.top while node: yield node.data _UpperCamelCase = node.next def __str__( self) -> str: '''simple docstring''' return "->".join([str(__a) for item in self]) def __len__( self) -> int: '''simple docstring''' return len(tuple(iter(self))) def UpperCAmelCase ( self) -> bool: '''simple docstring''' return self.top is None def UpperCAmelCase ( self , __a) -> None: '''simple docstring''' _UpperCamelCase = Node(__a) if not self.is_empty(): _UpperCamelCase = self.top _UpperCamelCase = node def UpperCAmelCase ( self) -> T: '''simple docstring''' if self.is_empty(): raise IndexError('''pop from empty stack''') assert isinstance(self.top , __a) _UpperCamelCase = self.top _UpperCamelCase = self.top.next return pop_node.data def UpperCAmelCase ( self) -> T: '''simple docstring''' if self.is_empty(): raise IndexError('''peek from empty stack''') assert self.top is not None return self.top.data def UpperCAmelCase ( self) -> None: '''simple docstring''' _UpperCamelCase = None if __name__ == "__main__": from doctest import testmod testmod()
78
"""simple docstring""" from __future__ import annotations import unittest from transformers import AutoTokenizer, MBartConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFMBartForConditionalGeneration, TFMBartModel @require_tf class _UpperCAmelCase: lowercase__ = MBartConfig lowercase__ = {} lowercase__ = 'gelu' def __init__( self , __a , __a=13 , __a=7 , __a=True , __a=False , __a=99 , __a=32 , __a=2 , __a=4 , __a=37 , __a=0.1 , __a=0.1 , __a=20 , __a=2 , __a=1 , __a=0 , ) -> Any: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = seq_length _UpperCamelCase = is_training _UpperCamelCase = use_labels _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = eos_token_id _UpperCamelCase = pad_token_id _UpperCamelCase = bos_token_id def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size) _UpperCamelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size) , 1) _UpperCamelCase = tf.concat([input_ids, eos_tensor] , axis=1) _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) _UpperCamelCase = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) _UpperCamelCase = prepare_mbart_inputs_dict(__a , __a , __a) return config, inputs_dict def UpperCAmelCase ( self , __a , __a) -> Optional[int]: '''simple docstring''' _UpperCamelCase = TFMBartModel(config=__a).get_decoder() _UpperCamelCase = inputs_dict['''input_ids'''] _UpperCamelCase = input_ids[:1, :] _UpperCamelCase = inputs_dict['''attention_mask'''][:1, :] _UpperCamelCase = inputs_dict['''head_mask'''] _UpperCamelCase = 1 # first forward pass _UpperCamelCase = model(__a , attention_mask=__a , head_mask=__a , use_cache=__a) _UpperCamelCase , _UpperCamelCase = outputs.to_tuple() _UpperCamelCase = past_key_values[1] def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case, __snake_case=None, __snake_case=None, __snake_case=None, __snake_case=None, __snake_case=None, ) -> Optional[int]: """simple docstring""" if attention_mask is None: _UpperCamelCase = tf.cast(tf.math.not_equal(__snake_case, config.pad_token_id ), tf.inta ) if decoder_attention_mask is None: _UpperCamelCase = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape, dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:], config.pad_token_id ), tf.inta ), ], axis=-1, ) if head_mask is None: _UpperCamelCase = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _UpperCamelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: _UpperCamelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class _UpperCAmelCase( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase__ = (TFMBartForConditionalGeneration, TFMBartModel) if is_tf_available() else () lowercase__ = (TFMBartForConditionalGeneration,) if is_tf_available() else () lowercase__ = ( { 'conversational': TFMBartForConditionalGeneration, 'feature-extraction': TFMBartModel, 'summarization': TFMBartForConditionalGeneration, 'text2text-generation': TFMBartForConditionalGeneration, 'translation': TFMBartForConditionalGeneration, } if is_tf_available() else {} ) lowercase__ = True lowercase__ = False lowercase__ = False def UpperCAmelCase ( self , __a , __a , __a , __a , __a) -> Dict: '''simple docstring''' if pipeline_test_casse_name != "FeatureExtractionPipelineTests": # Exception encountered when calling layer '...' return True return False def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = TFMBartModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=__a) def UpperCAmelCase ( self) -> str: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__a) @require_sentencepiece @require_tokenizers @require_tf class _UpperCAmelCase( unittest.TestCase ): lowercase__ = [ ' UN Chief Says There Is No Military Solution in Syria', ] lowercase__ = [ 'Şeful ONU declară că nu există o soluţie militară în Siria', ] lowercase__ = 'facebook/mbart-large-en-ro' @cached_property def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' return AutoTokenizer.from_pretrained(self.model_name) @cached_property def UpperCAmelCase ( self) -> str: '''simple docstring''' _UpperCamelCase = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name) return model def UpperCAmelCase ( self , **__a) -> List[str]: '''simple docstring''' _UpperCamelCase = self.translate_src_text(**__a) self.assertListEqual(self.expected_text , __a) def UpperCAmelCase ( self , **__a) -> Dict: '''simple docstring''' _UpperCamelCase = self.tokenizer(self.src_text , **__a , return_tensors='''tf''') _UpperCamelCase = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2) _UpperCamelCase = self.tokenizer.batch_decode(__a , skip_special_tokens=__a) return generated_words @slow def UpperCAmelCase ( self) -> Any: '''simple docstring''' self._assert_generated_batch_equal_expected()
78
1
"""simple docstring""" from __future__ import annotations def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case, __snake_case ) -> None: """simple docstring""" if (direction == 1 and array[indexa] > array[indexa]) or ( direction == 0 and array[indexa] < array[indexa] ): _UpperCamelCase , _UpperCamelCase = array[indexa], array[indexa] def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case, __snake_case ) -> None: """simple docstring""" if length > 1: _UpperCamelCase = int(length / 2 ) for i in range(__snake_case, low + middle ): comp_and_swap(__snake_case, __snake_case, i + middle, __snake_case ) bitonic_merge(__snake_case, __snake_case, __snake_case, __snake_case ) bitonic_merge(__snake_case, low + middle, __snake_case, __snake_case ) def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case, __snake_case ) -> None: """simple docstring""" if length > 1: _UpperCamelCase = int(length / 2 ) bitonic_sort(__snake_case, __snake_case, __snake_case, 1 ) bitonic_sort(__snake_case, low + middle, __snake_case, 0 ) bitonic_merge(__snake_case, __snake_case, __snake_case, __snake_case ) if __name__ == "__main__": _a = input("""Enter numbers separated by a comma:\n""").strip() _a = [int(item.strip()) for item in user_input.split(""",""")] bitonic_sort(unsorted, 0, len(unsorted), 1) print("""\nSorted array in ascending order is: """, end="""""") print(*unsorted, sep=""", """) bitonic_merge(unsorted, 0, len(unsorted), 0) print("""Sorted array in descending order is: """, end="""""") print(*unsorted, sep=""", """)
78
"""simple docstring""" from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _a = logging.get_logger(__name__) class _UpperCAmelCase( lowerCamelCase ): lowercase__ = ['pixel_values'] def __init__( self , __a = True , __a = 1 / 2_55 , __a = True , __a = 8 , **__a , ) -> None: '''simple docstring''' super().__init__(**__a) _UpperCamelCase = do_rescale _UpperCamelCase = rescale_factor _UpperCamelCase = do_pad _UpperCamelCase = pad_size def UpperCAmelCase ( self , __a , __a , __a = None , **__a) -> np.ndarray: '''simple docstring''' return rescale(__a , scale=__a , data_format=__a , **__a) def UpperCAmelCase ( self , __a , __a , __a = None) -> List[Any]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = get_image_size(__a) _UpperCamelCase = (old_height // size + 1) * size - old_height _UpperCamelCase = (old_width // size + 1) * size - old_width return pad(__a , ((0, pad_height), (0, pad_width)) , mode='''symmetric''' , data_format=__a) def UpperCAmelCase ( self , __a , __a = None , __a = None , __a = None , __a = None , __a = None , __a = ChannelDimension.FIRST , **__a , ) -> Tuple: '''simple docstring''' _UpperCamelCase = do_rescale if do_rescale is not None else self.do_rescale _UpperCamelCase = rescale_factor if rescale_factor is not None else self.rescale_factor _UpperCamelCase = do_pad if do_pad is not None else self.do_pad _UpperCamelCase = pad_size if pad_size is not None else self.pad_size _UpperCamelCase = make_list_of_images(__a) if not valid_images(__a): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''') if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''') # All transformations expect numpy arrays. _UpperCamelCase = [to_numpy_array(__a) for image in images] if do_rescale: _UpperCamelCase = [self.rescale(image=__a , scale=__a) for image in images] if do_pad: _UpperCamelCase = [self.pad(__a , size=__a) for image in images] _UpperCamelCase = [to_channel_dimension_format(__a , __a) for image in images] _UpperCamelCase = {'''pixel_values''': images} return BatchFeature(data=__a , tensor_type=__a)
78
1
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_rembert import RemBertTokenizer else: _a = None _a = logging.get_logger(__name__) _a = {"""vocab_file""": """sentencepiece.model""", """tokenizer_file""": """tokenizer.json"""} _a = { """vocab_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/sentencepiece.model""", }, """tokenizer_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/tokenizer.json""", }, } _a = { """google/rembert""": 256, } _a = """▁""" class _UpperCAmelCase( lowerCamelCase ): lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = RemBertTokenizer def __init__( self , __a=None , __a=None , __a=True , __a=True , __a=False , __a="[CLS]" , __a="[SEP]" , __a="<unk>" , __a="[SEP]" , __a="<pad>" , __a="[CLS]" , __a="[MASK]" , **__a , ) -> Optional[Any]: '''simple docstring''' # Mask token behave like a normal word, i.e. include the space before it _UpperCamelCase = AddedToken(__a , lstrip=__a , rstrip=__a) if isinstance(__a , __a) else mask_token super().__init__( __a , tokenizer_file=__a , do_lower_case=__a , remove_space=__a , keep_accents=__a , bos_token=__a , eos_token=__a , unk_token=__a , sep_token=__a , pad_token=__a , cls_token=__a , mask_token=__a , **__a , ) _UpperCamelCase = do_lower_case _UpperCamelCase = remove_space _UpperCamelCase = keep_accents _UpperCamelCase = vocab_file _UpperCamelCase = False if not self.vocab_file else True def UpperCAmelCase ( self , __a , __a = None) -> List[int]: '''simple docstring''' _UpperCamelCase = [self.sep_token_id] _UpperCamelCase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def UpperCAmelCase ( self , __a , __a = None , __a = False) -> List[int]: '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''') return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(__a)) + [1] + ([0] * len(__a)) + [1] return [1] + ([0] * len(__a)) + [1] def UpperCAmelCase ( self , __a , __a = None) -> List[int]: '''simple docstring''' _UpperCamelCase = [self.sep_token_id] _UpperCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def UpperCAmelCase ( self , __a , __a = None) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(__a): logger.error('''Vocabulary path ({}) should be a directory'''.format(__a)) return _UpperCamelCase = os.path.join( __a , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file''']) if os.path.abspath(self.vocab_file) != os.path.abspath(__a): copyfile(self.vocab_file , __a) return (out_vocab_file,)
78
"""simple docstring""" from importlib import import_module from .logging import get_logger _a = get_logger(__name__) class _UpperCAmelCase: def __init__( self , __a , __a=None) -> Dict: '''simple docstring''' _UpperCamelCase = attrs or [] if module is not None: for key in module.__dict__: if key in attrs or not key.startswith('''__'''): setattr(self , __a , getattr(__a , __a)) _UpperCamelCase = module._original_module if isinstance(__a , _PatchedModuleObj) else module class _UpperCAmelCase: lowercase__ = [] def __init__( self , __a , __a , __a , __a=None) -> List[str]: '''simple docstring''' _UpperCamelCase = obj _UpperCamelCase = target _UpperCamelCase = new _UpperCamelCase = target.split('''.''')[0] _UpperCamelCase = {} _UpperCamelCase = attrs or [] def __enter__( self) -> int: '''simple docstring''' *_UpperCamelCase , _UpperCamelCase = self.target.split('''.''') # Patch modules: # it's used to patch attributes of submodules like "os.path.join"; # in this case we need to patch "os" and "os.path" for i in range(len(__a)): try: _UpperCamelCase = import_module('''.'''.join(submodules[: i + 1])) except ModuleNotFoundError: continue # We iterate over all the globals in self.obj in case we find "os" or "os.path" for attr in self.obj.__dir__(): _UpperCamelCase = getattr(self.obj , __a) # We don't check for the name of the global, but rather if its value *is* "os" or "os.path". # This allows to patch renamed modules like "from os import path as ospath". if obj_attr is submodule or ( (isinstance(__a , _PatchedModuleObj) and obj_attr._original_module is submodule) ): _UpperCamelCase = obj_attr # patch at top level setattr(self.obj , __a , _PatchedModuleObj(__a , attrs=self.attrs)) _UpperCamelCase = getattr(self.obj , __a) # construct lower levels patches for key in submodules[i + 1 :]: setattr(__a , __a , _PatchedModuleObj(getattr(__a , __a , __a) , attrs=self.attrs)) _UpperCamelCase = getattr(__a , __a) # finally set the target attribute setattr(__a , __a , self.new) # Patch attribute itself: # it's used for builtins like "open", # and also to patch "os.path.join" we may also need to patch "join" # itself if it was imported as "from os.path import join". if submodules: # if it's an attribute of a submodule like "os.path.join" try: _UpperCamelCase = getattr(import_module('''.'''.join(__a)) , __a) except (AttributeError, ModuleNotFoundError): return # We iterate over all the globals in self.obj in case we find "os.path.join" for attr in self.obj.__dir__(): # We don't check for the name of the global, but rather if its value *is* "os.path.join". # This allows to patch renamed attributes like "from os.path import join as pjoin". if getattr(self.obj , __a) is attr_value: _UpperCamelCase = getattr(self.obj , __a) setattr(self.obj , __a , self.new) elif target_attr in globals()["__builtins__"]: # if it'a s builtin like "open" _UpperCamelCase = globals()['''__builtins__'''][target_attr] setattr(self.obj , __a , self.new) else: raise RuntimeError(F'''Tried to patch attribute {target_attr} instead of a submodule.''') def __exit__( self , *__a) -> Tuple: '''simple docstring''' for attr in list(self.original): setattr(self.obj , __a , self.original.pop(__a)) def UpperCAmelCase ( self) -> Dict: '''simple docstring''' self.__enter__() self._active_patches.append(self) def UpperCAmelCase ( self) -> str: '''simple docstring''' try: self._active_patches.remove(self) except ValueError: # If the patch hasn't been started this will fail return None return self.__exit__()
78
1
"""simple docstring""" from collections import namedtuple import requests from lxml import html # type: ignore _a = namedtuple("""covid_data""", """cases deaths recovered""") def lowerCamelCase__ ( __snake_case = "https://www.worldometers.info/coronavirus/" ) -> covid_data: """simple docstring""" _UpperCamelCase = '''//div[@class = "maincounter-number"]/span/text()''' return covid_data(*html.fromstring(requests.get(__snake_case ).content ).xpath(__snake_case ) ) _a = """Total COVID-19 cases in the world: {} Total deaths due to COVID-19 in the world: {} Total COVID-19 patients recovered in the world: {}""" print(fmt.format(*covid_stats()))
78
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_flax_available, is_torch_available, is_transformers_available, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .multicontrolnet import MultiControlNetModel from .pipeline_controlnet import StableDiffusionControlNetPipeline from .pipeline_controlnet_imgaimg import StableDiffusionControlNetImgaImgPipeline from .pipeline_controlnet_inpaint import StableDiffusionControlNetInpaintPipeline if is_transformers_available() and is_flax_available(): from .pipeline_flax_controlnet import FlaxStableDiffusionControlNetPipeline
78
1
"""simple docstring""" import os from typing import BinaryIO, Optional, Union import numpy as np import pyarrow.parquet as pq from .. import Audio, Dataset, Features, Image, NamedSplit, Value, config from ..features.features import FeatureType, _visit from ..formatting import query_table from ..packaged_modules import _PACKAGED_DATASETS_MODULES from ..packaged_modules.parquet.parquet import Parquet from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" _UpperCamelCase = np.inf def set_batch_size(__snake_case ) -> None: nonlocal batch_size if isinstance(__snake_case, __snake_case ): _UpperCamelCase = min(__snake_case, config.PARQUET_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS ) elif isinstance(__snake_case, __snake_case ): _UpperCamelCase = min(__snake_case, config.PARQUET_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS ) elif isinstance(__snake_case, __snake_case ) and feature.dtype == "binary": _UpperCamelCase = min(__snake_case, config.PARQUET_ROW_GROUP_SIZE_FOR_BINARY_DATASETS ) _visit(__snake_case, __snake_case ) return None if batch_size is np.inf else batch_size class _UpperCAmelCase( lowerCamelCase ): def __init__( self , __a , __a = None , __a = None , __a = None , __a = False , __a = False , __a = None , **__a , ) -> Dict: '''simple docstring''' super().__init__( __a , split=__a , features=__a , cache_dir=__a , keep_in_memory=__a , streaming=__a , num_proc=__a , **__a , ) _UpperCamelCase = path_or_paths if isinstance(__a , __a) else {self.split: path_or_paths} _UpperCamelCase = _PACKAGED_DATASETS_MODULES['''parquet'''][1] _UpperCamelCase = Parquet( cache_dir=__a , data_files=__a , features=__a , hash=__a , **__a , ) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' # Build iterable dataset if self.streaming: _UpperCamelCase = self.builder.as_streaming_dataset(split=self.split) # Build regular (map-style) dataset else: _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None self.builder.download_and_prepare( download_config=__a , download_mode=__a , verification_mode=__a , base_path=__a , num_proc=self.num_proc , ) _UpperCamelCase = self.builder.as_dataset( split=self.split , verification_mode=__a , in_memory=self.keep_in_memory) return dataset class _UpperCAmelCase: def __init__( self , __a , __a , __a = None , **__a , ) -> Dict: '''simple docstring''' _UpperCamelCase = dataset _UpperCamelCase = path_or_buf _UpperCamelCase = batch_size or get_writer_batch_size(dataset.features) _UpperCamelCase = parquet_writer_kwargs def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = self.batch_size if self.batch_size else config.DEFAULT_MAX_BATCH_SIZE if isinstance(self.path_or_buf , (str, bytes, os.PathLike)): with open(self.path_or_buf , '''wb+''') as buffer: _UpperCamelCase = self._write(file_obj=__a , batch_size=__a , **self.parquet_writer_kwargs) else: _UpperCamelCase = self._write(file_obj=self.path_or_buf , batch_size=__a , **self.parquet_writer_kwargs) return written def UpperCAmelCase ( self , __a , __a , **__a) -> int: '''simple docstring''' _UpperCamelCase = 0 _UpperCamelCase = parquet_writer_kwargs.pop('''path_or_buf''' , __a) _UpperCamelCase = self.dataset.features.arrow_schema _UpperCamelCase = pq.ParquetWriter(__a , schema=__a , **__a) for offset in logging.tqdm( range(0 , len(self.dataset) , __a) , unit='''ba''' , disable=not logging.is_progress_bar_enabled() , desc='''Creating parquet from Arrow format''' , ): _UpperCamelCase = query_table( table=self.dataset._data , key=slice(__a , offset + batch_size) , indices=self.dataset._indices if self.dataset._indices is not None else None , ) writer.write_table(__a) written += batch.nbytes writer.close() return written
78
"""simple docstring""" from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging _a = logging.get_logger(__name__) _a = { """EleutherAI/gpt-neo-1.3B""": """https://huggingface.co/EleutherAI/gpt-neo-1.3B/resolve/main/config.json""", # See all GPTNeo models at https://huggingface.co/models?filter=gpt_neo } class _UpperCAmelCase( lowerCamelCase ): lowercase__ = 'gpt_neo' lowercase__ = ['past_key_values'] lowercase__ = {'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers'} def __init__( self , __a=5_02_57 , __a=20_48 , __a=20_48 , __a=24 , __a=[[["global", "local"], 12]] , __a=16 , __a=None , __a=2_56 , __a="gelu_new" , __a=0.0 , __a=0.0 , __a=0.0 , __a=0.1 , __a=1e-5 , __a=0.02 , __a=True , __a=5_02_56 , __a=5_02_56 , **__a , ) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = vocab_size _UpperCamelCase = max_position_embeddings _UpperCamelCase = hidden_size _UpperCamelCase = num_layers _UpperCamelCase = num_heads _UpperCamelCase = intermediate_size _UpperCamelCase = window_size _UpperCamelCase = activation_function _UpperCamelCase = resid_dropout _UpperCamelCase = embed_dropout _UpperCamelCase = attention_dropout _UpperCamelCase = classifier_dropout _UpperCamelCase = layer_norm_epsilon _UpperCamelCase = initializer_range _UpperCamelCase = use_cache _UpperCamelCase = bos_token_id _UpperCamelCase = eos_token_id _UpperCamelCase = attention_types _UpperCamelCase = self.expand_attention_types_params(__a) if len(self.attention_layers) != self.num_layers: raise ValueError( '''Configuration for convolutional module is incorrect. ''' '''It is required that `len(config.attention_layers)` == `config.num_layers` ''' F'''but is `len(config.attention_layers) = {len(self.attention_layers)}`, ''' F'''`config.num_layers = {self.num_layers}`. ''' '''`config.attention_layers` is prepared using `config.attention_types`. ''' '''Please verify the value of `config.attention_types` argument.''') super().__init__(bos_token_id=__a , eos_token_id=__a , **__a) @staticmethod def UpperCAmelCase ( __a) -> int: '''simple docstring''' _UpperCamelCase = [] for item in attention_types: for _ in range(item[1]): attentions.extend(item[0]) return attentions def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case, __snake_case ) -> str: """simple docstring""" import torch _UpperCamelCase = input.size() _UpperCamelCase = len(__snake_case ) _UpperCamelCase = shape[dimension] _UpperCamelCase = torch.arange(0, __snake_case, __snake_case ) _UpperCamelCase = torch.div(sizedim - size, __snake_case, rounding_mode='''floor''' ) + 1 _UpperCamelCase = torch.arange(__snake_case ) + low_indices[:min_length][:, None] _UpperCamelCase = [slice(__snake_case )] * rank _UpperCamelCase = indices _UpperCamelCase = input[s] _UpperCamelCase = list(range(0, rank + 1 ) ) perm.append(perm.pop(dimension + 1 ) ) return sliced.permute(__snake_case ) def lowerCamelCase__ ( __snake_case, __snake_case ) -> str: """simple docstring""" import torch _UpperCamelCase = torch.arange(1, __snake_case ) _UpperCamelCase = torch.remainder(__snake_case, __snake_case ) _UpperCamelCase = remainders == 0 _UpperCamelCase = candidates[divisor_indices] _UpperCamelCase = torch.max(__snake_case ) return largest_divisor, torch.div(__snake_case, __snake_case, rounding_mode='''floor''' ) class _UpperCAmelCase( lowerCamelCase ): @property def UpperCAmelCase ( self) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' _UpperCamelCase = OrderedDict({'''input_ids''': {0: '''batch''', 1: '''sequence'''}}) if self.use_past: self.fill_with_past_key_values_(__a , direction='''inputs''') _UpperCamelCase = {0: '''batch''', 1: '''past_sequence + sequence'''} else: _UpperCamelCase = {0: '''batch''', 1: '''sequence'''} return common_inputs @property def UpperCAmelCase ( self) -> int: '''simple docstring''' return self._config.num_heads def UpperCAmelCase ( self , __a , __a = -1 , __a = -1 , __a = False , __a = None , ) -> Mapping[str, Any]: '''simple docstring''' _UpperCamelCase = super(__a , self).generate_dummy_inputs( __a , batch_size=__a , seq_length=__a , is_pair=__a , framework=__a) # We need to order the input in the way they appears in the forward() _UpperCamelCase = OrderedDict({'''input_ids''': common_inputs['''input_ids''']}) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''') else: import torch _UpperCamelCase , _UpperCamelCase = common_inputs['''input_ids'''].shape # Not using the same length for past_key_values _UpperCamelCase = seqlen + 2 _UpperCamelCase = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) _UpperCamelCase = [ (torch.zeros(__a), torch.zeros(__a)) for _ in range(self.num_layers) ] _UpperCamelCase = common_inputs['''attention_mask'''] if self.use_past: _UpperCamelCase = ordered_inputs['''attention_mask'''].dtype _UpperCamelCase = torch.cat( [ordered_inputs['''attention_mask'''], torch.ones(__a , __a , dtype=__a)] , dim=1) return ordered_inputs @property def UpperCAmelCase ( self) -> int: '''simple docstring''' return 13
78
1
"""simple docstring""" import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate # and perform gradient accumulation # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## _a = 16 _a = 32 def lowerCamelCase__ ( __snake_case, __snake_case = 16 ) -> Optional[int]: """simple docstring""" _UpperCamelCase = AutoTokenizer.from_pretrained('''bert-base-cased''' ) _UpperCamelCase = load_dataset('''glue''', '''mrpc''' ) def tokenize_function(__snake_case ): # max_length=None => use the model max length (it's actually the default) _UpperCamelCase = tokenizer(examples['''sentence1'''], examples['''sentence2'''], truncation=__snake_case, max_length=__snake_case ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): _UpperCamelCase = datasets.map( __snake_case, batched=__snake_case, remove_columns=['''idx''', '''sentence1''', '''sentence2'''], ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _UpperCamelCase = tokenized_datasets.rename_column('''label''', '''labels''' ) def collate_fn(__snake_case ): # On TPU it's best to pad everything to the same length or training will be very slow. _UpperCamelCase = 1_28 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": _UpperCamelCase = 16 elif accelerator.mixed_precision != "no": _UpperCamelCase = 8 else: _UpperCamelCase = None return tokenizer.pad( __snake_case, padding='''longest''', max_length=__snake_case, pad_to_multiple_of=__snake_case, return_tensors='''pt''', ) # Instantiate dataloaders. _UpperCamelCase = DataLoader( tokenized_datasets['''train'''], shuffle=__snake_case, collate_fn=__snake_case, batch_size=__snake_case ) _UpperCamelCase = DataLoader( tokenized_datasets['''validation'''], shuffle=__snake_case, collate_fn=__snake_case, batch_size=__snake_case ) return train_dataloader, eval_dataloader # For testing only if os.environ.get("""TESTING_MOCKED_DATALOADERS""", None) == "1": from accelerate.test_utils.training import mocked_dataloaders _a = mocked_dataloaders # noqa: F811 def lowerCamelCase__ ( __snake_case, __snake_case ) -> int: """simple docstring""" if os.environ.get('''TESTING_MOCKED_DATALOADERS''', __snake_case ) == "1": _UpperCamelCase = 2 # New Code # _UpperCamelCase = int(args.gradient_accumulation_steps ) # Initialize accelerator _UpperCamelCase = Accelerator( cpu=args.cpu, mixed_precision=args.mixed_precision, gradient_accumulation_steps=__snake_case ) if accelerator.distributed_type == DistributedType.TPU and gradient_accumulation_steps > 1: raise NotImplementedError( '''Gradient accumulation on TPUs is currently not supported. Pass `gradient_accumulation_steps=1`''' ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _UpperCamelCase = config['''lr'''] _UpperCamelCase = int(config['''num_epochs'''] ) _UpperCamelCase = int(config['''seed'''] ) _UpperCamelCase = int(config['''batch_size'''] ) _UpperCamelCase = evaluate.load('''glue''', '''mrpc''' ) set_seed(__snake_case ) _UpperCamelCase , _UpperCamelCase = get_dataloaders(__snake_case, __snake_case ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _UpperCamelCase = AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''', return_dict=__snake_case ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _UpperCamelCase = model.to(accelerator.device ) # Instantiate optimizer _UpperCamelCase = AdamW(params=model.parameters(), lr=__snake_case ) # Instantiate scheduler _UpperCamelCase = get_linear_schedule_with_warmup( optimizer=__snake_case, num_warmup_steps=1_00, num_training_steps=(len(__snake_case ) * num_epochs), ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = accelerator.prepare( __snake_case, __snake_case, __snake_case, __snake_case, __snake_case ) # Now we train the model for epoch in range(__snake_case ): model.train() for step, batch in enumerate(__snake_case ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) # New code # # We use the new `accumulate` context manager to perform gradient accumulation # We also currently do not support TPUs nor advise it as bugs were found on the XLA side when running our tests. with accelerator.accumulate(__snake_case ): _UpperCamelCase = model(**__snake_case ) _UpperCamelCase = output.loss accelerator.backward(__snake_case ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__snake_case ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _UpperCamelCase = model(**__snake_case ) _UpperCamelCase = outputs.logits.argmax(dim=-1 ) _UpperCamelCase , _UpperCamelCase = accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=__snake_case, references=__snake_case, ) _UpperCamelCase = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F'''epoch {epoch}:''', __snake_case ) def lowerCamelCase__ ( ) -> str: """simple docstring""" _UpperCamelCase = argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''', type=__snake_case, default=__snake_case, choices=['''no''', '''fp16''', '''bf16''', '''fp8'''], help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''', ) # New Code # parser.add_argument( '''--gradient_accumulation_steps''', type=__snake_case, default=1, help='''The number of minibatches to be ran before gradients are accumulated.''', ) parser.add_argument('''--cpu''', action='''store_true''', help='''If passed, will train on the CPU.''' ) _UpperCamelCase = parser.parse_args() _UpperCamelCase = {'''lr''': 2e-5, '''num_epochs''': 3, '''seed''': 42, '''batch_size''': 16} training_function(__snake_case, __snake_case ) if __name__ == "__main__": main()
78
"""simple docstring""" import sys from collections import defaultdict class _UpperCAmelCase: def __init__( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = [] def UpperCAmelCase ( self , __a) -> Optional[Any]: '''simple docstring''' return self.node_position[vertex] def UpperCAmelCase ( self , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = pos def UpperCAmelCase ( self , __a , __a , __a , __a) -> Tuple: '''simple docstring''' if start > size // 2 - 1: return else: if 2 * start + 2 >= size: _UpperCamelCase = 2 * start + 1 else: if heap[2 * start + 1] < heap[2 * start + 2]: _UpperCamelCase = 2 * start + 1 else: _UpperCamelCase = 2 * start + 2 if heap[smallest_child] < heap[start]: _UpperCamelCase , _UpperCamelCase = heap[smallest_child], positions[smallest_child] _UpperCamelCase , _UpperCamelCase = ( heap[start], positions[start], ) _UpperCamelCase , _UpperCamelCase = temp, tempa _UpperCamelCase = self.get_position(positions[smallest_child]) self.set_position( positions[smallest_child] , self.get_position(positions[start])) self.set_position(positions[start] , __a) self.top_to_bottom(__a , __a , __a , __a) def UpperCAmelCase ( self , __a , __a , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = position[index] while index != 0: _UpperCamelCase = int((index - 2) / 2) if index % 2 == 0 else int((index - 1) / 2) if val < heap[parent]: _UpperCamelCase = heap[parent] _UpperCamelCase = position[parent] self.set_position(position[parent] , __a) else: _UpperCamelCase = val _UpperCamelCase = temp self.set_position(__a , __a) break _UpperCamelCase = parent else: _UpperCamelCase = val _UpperCamelCase = temp self.set_position(__a , 0) def UpperCAmelCase ( self , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = len(__a) // 2 - 1 for i in range(__a , -1 , -1): self.top_to_bottom(__a , __a , len(__a) , __a) def UpperCAmelCase ( self , __a , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = positions[0] _UpperCamelCase = sys.maxsize self.top_to_bottom(__a , 0 , len(__a) , __a) return temp def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" _UpperCamelCase = Heap() _UpperCamelCase = [0] * len(__snake_case ) _UpperCamelCase = [-1] * len(__snake_case ) # Neighboring Tree Vertex of selected vertex # Minimum Distance of explored vertex with neighboring vertex of partial tree # formed in graph _UpperCamelCase = [] # Heap of Distance of vertices from their neighboring vertex _UpperCamelCase = [] for vertex in range(len(__snake_case ) ): distance_tv.append(sys.maxsize ) positions.append(__snake_case ) heap.node_position.append(__snake_case ) _UpperCamelCase = [] _UpperCamelCase = 1 _UpperCamelCase = sys.maxsize for neighbor, distance in adjacency_list[0]: _UpperCamelCase = 0 _UpperCamelCase = distance heap.heapify(__snake_case, __snake_case ) for _ in range(1, len(__snake_case ) ): _UpperCamelCase = heap.delete_minimum(__snake_case, __snake_case ) if visited[vertex] == 0: tree_edges.append((nbr_tv[vertex], vertex) ) _UpperCamelCase = 1 for neighbor, distance in adjacency_list[vertex]: if ( visited[neighbor] == 0 and distance < distance_tv[heap.get_position(__snake_case )] ): _UpperCamelCase = distance heap.bottom_to_top( __snake_case, heap.get_position(__snake_case ), __snake_case, __snake_case ) _UpperCamelCase = vertex return tree_edges if __name__ == "__main__": # pragma: no cover # < --------- Prims Algorithm --------- > _a = int(input("""Enter number of edges: """).strip()) _a = defaultdict(list) for _ in range(edges_number): _a = [int(x) for x in input().strip().split()] adjacency_list[edge[0]].append([edge[1], edge[2]]) adjacency_list[edge[1]].append([edge[0], edge[2]]) print(prisms_algorithm(adjacency_list))
78
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available _a = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ["""BartphoTokenizer"""] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bartpho import BartphoTokenizer else: import sys _a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
78
"""simple docstring""" import json import sys def lowerCamelCase__ ( __snake_case, __snake_case ) -> Union[str, Any]: """simple docstring""" with open(__snake_case, encoding='''utf-8''' ) as f: _UpperCamelCase = json.load(__snake_case ) _UpperCamelCase = ['''<details>''', '''<summary>Show updated benchmarks!</summary>''', ''' '''] for benchmark_name in sorted(__snake_case ): _UpperCamelCase = results[benchmark_name] _UpperCamelCase = benchmark_name.split('''/''' )[-1] output_md.append(F'''### Benchmark: {benchmark_file_name}''' ) _UpperCamelCase = '''| metric |''' _UpperCamelCase = '''|--------|''' _UpperCamelCase = '''| new / old (diff) |''' for metric_name in sorted(__snake_case ): _UpperCamelCase = benchmark_res[metric_name] _UpperCamelCase = metric_vals['''new'''] _UpperCamelCase = metric_vals.get('''old''', __snake_case ) _UpperCamelCase = metric_vals.get('''diff''', __snake_case ) _UpperCamelCase = F''' {new_val:f}''' if isinstance(__snake_case, (int, float) ) else '''None''' if old_val is not None: val_str += F''' / {old_val:f}''' if isinstance(__snake_case, (int, float) ) else "None" if dif_val is not None: val_str += F''' ({dif_val:f})''' if isinstance(__snake_case, (int, float) ) else "None" title += " " + metric_name + " |" lines += "---|" value += val_str + " |" output_md += [title, lines, value, " "] output_md.append('''</details>''' ) with open(__snake_case, '''w''', encoding='''utf-8''' ) as f: f.writelines('''\n'''.join(__snake_case ) ) if __name__ == "__main__": _a = sys.argv[1] _a = sys.argv[2] format_json_to_md(input_json_file, output_md_file)
78
1
"""simple docstring""" import unittest from transformers import AlbertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForPreTraining, AlbertForQuestionAnswering, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertModel, ) from transformers.models.albert.modeling_albert import ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST class _UpperCAmelCase: def __init__( self , __a , __a=13 , __a=7 , __a=True , __a=True , __a=True , __a=True , __a=99 , __a=16 , __a=36 , __a=6 , __a=6 , __a=6 , __a=37 , __a="gelu" , __a=0.1 , __a=0.1 , __a=5_12 , __a=16 , __a=2 , __a=0.02 , __a=3 , __a=4 , __a=None , ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = seq_length _UpperCamelCase = is_training _UpperCamelCase = use_input_mask _UpperCamelCase = use_token_type_ids _UpperCamelCase = use_labels _UpperCamelCase = vocab_size _UpperCamelCase = embedding_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_hidden_groups _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = type_vocab_size _UpperCamelCase = type_sequence_label_size _UpperCamelCase = initializer_range _UpperCamelCase = num_labels _UpperCamelCase = num_choices _UpperCamelCase = scope def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) _UpperCamelCase = None if self.use_input_mask: _UpperCamelCase = random_attention_mask([self.batch_size, self.seq_length]) _UpperCamelCase = None if self.use_token_type_ids: _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size) _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) _UpperCamelCase = ids_tensor([self.batch_size] , self.num_choices) _UpperCamelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase ( self) -> Dict: '''simple docstring''' return AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , num_hidden_groups=self.num_hidden_groups , ) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Optional[int]: '''simple docstring''' _UpperCamelCase = AlbertModel(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , token_type_ids=__a) _UpperCamelCase = model(__a , token_type_ids=__a) _UpperCamelCase = model(__a) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = AlbertForPreTraining(config=__a) model.to(__a) model.eval() _UpperCamelCase = model( __a , attention_mask=__a , token_type_ids=__a , labels=__a , sentence_order_label=__a , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) self.parent.assertEqual(result.sop_logits.shape , (self.batch_size, config.num_labels)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Optional[int]: '''simple docstring''' _UpperCamelCase = AlbertForMaskedLM(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Any: '''simple docstring''' _UpperCamelCase = AlbertForQuestionAnswering(config=__a) model.to(__a) model.eval() _UpperCamelCase = model( __a , attention_mask=__a , token_type_ids=__a , start_positions=__a , end_positions=__a , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> List[Any]: '''simple docstring''' _UpperCamelCase = self.num_labels _UpperCamelCase = AlbertForSequenceClassification(__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> List[str]: '''simple docstring''' _UpperCamelCase = self.num_labels _UpperCamelCase = AlbertForTokenClassification(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.num_choices _UpperCamelCase = AlbertForMultipleChoice(config=__a) model.to(__a) model.eval() _UpperCamelCase = input_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() _UpperCamelCase = token_type_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() _UpperCamelCase = input_mask.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() _UpperCamelCase = model( __a , attention_mask=__a , token_type_ids=__a , labels=__a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices)) def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.prepare_config_and_inputs() ( ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ) = config_and_inputs _UpperCamelCase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class _UpperCAmelCase( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase__ = ( ( AlbertModel, AlbertForPreTraining, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertForQuestionAnswering, ) if is_torch_available() else () ) lowercase__ = ( { 'feature-extraction': AlbertModel, 'fill-mask': AlbertForMaskedLM, 'question-answering': AlbertForQuestionAnswering, 'text-classification': AlbertForSequenceClassification, 'token-classification': AlbertForTokenClassification, 'zero-shot': AlbertForSequenceClassification, } if is_torch_available() else {} ) lowercase__ = True def UpperCAmelCase ( self , __a , __a , __a=False) -> List[str]: '''simple docstring''' _UpperCamelCase = super()._prepare_for_class(__a , __a , return_labels=__a) if return_labels: if model_class in get_values(__a): _UpperCamelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=__a) _UpperCamelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__a) return inputs_dict def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = AlbertModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=__a , hidden_size=37) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*__a) def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__a) def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__a) def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__a) def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__a) def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _UpperCamelCase = type self.model_tester.create_and_check_model(*__a) @slow def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' for model_name in ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase = AlbertModel.from_pretrained(__a) self.assertIsNotNone(__a) @require_torch class _UpperCAmelCase( unittest.TestCase ): @slow def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = AlbertModel.from_pretrained('''albert-base-v2''') _UpperCamelCase = torch.tensor([[0, 3_45, 2_32, 3_28, 7_40, 1_40, 16_95, 69, 60_78, 15_88, 2]]) _UpperCamelCase = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]) with torch.no_grad(): _UpperCamelCase = model(__a , attention_mask=__a)[0] _UpperCamelCase = torch.Size((1, 11, 7_68)) self.assertEqual(output.shape , __a) _UpperCamelCase = torch.tensor( [[[-0.6513, 1.5035, -0.2766], [-0.6515, 1.5046, -0.2780], [-0.6512, 1.5049, -0.2784]]]) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , __a , atol=1e-4))
78
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTImageProcessor, ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) def lowerCamelCase__ ( __snake_case, __snake_case=False ) -> Tuple: """simple docstring""" _UpperCamelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'''blocks.{i}.norm1.weight''', F'''vit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((F'''blocks.{i}.norm1.bias''', F'''vit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((F'''blocks.{i}.attn.proj.weight''', F'''vit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.attn.proj.bias''', F'''vit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((F'''blocks.{i}.norm2.weight''', F'''vit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((F'''blocks.{i}.norm2.bias''', F'''vit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.weight''', F'''vit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.bias''', F'''vit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.weight''', F'''vit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.bias''', F'''vit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ('''cls_token''', '''vit.embeddings.cls_token'''), ('''patch_embed.proj.weight''', '''vit.embeddings.patch_embeddings.projection.weight'''), ('''patch_embed.proj.bias''', '''vit.embeddings.patch_embeddings.projection.bias'''), ('''pos_embed''', '''vit.embeddings.position_embeddings'''), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ('''pre_logits.fc.weight''', '''pooler.dense.weight'''), ('''pre_logits.fc.bias''', '''pooler.dense.bias'''), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _UpperCamelCase = [(pair[0], pair[1][4:]) if pair[1].startswith('''vit''' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('''norm.weight''', '''vit.layernorm.weight'''), ('''norm.bias''', '''vit.layernorm.bias'''), ('''head.weight''', '''classifier.weight'''), ('''head.bias''', '''classifier.bias'''), ] ) return rename_keys def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case=False ) -> str: """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: _UpperCamelCase = '''''' else: _UpperCamelCase = '''vit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _UpperCamelCase = state_dict.pop(F'''blocks.{i}.attn.qkv.weight''' ) _UpperCamelCase = state_dict.pop(F'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict _UpperCamelCase = in_proj_weight[ : config.hidden_size, : ] _UpperCamelCase = in_proj_bias[: config.hidden_size] _UpperCamelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _UpperCamelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _UpperCamelCase = in_proj_weight[ -config.hidden_size :, : ] _UpperCamelCase = in_proj_bias[-config.hidden_size :] def lowerCamelCase__ ( __snake_case ) -> Dict: """simple docstring""" _UpperCamelCase = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(__snake_case, __snake_case ) def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> Any: """simple docstring""" _UpperCamelCase = dct.pop(__snake_case ) _UpperCamelCase = val def lowerCamelCase__ ( ) -> Dict: """simple docstring""" _UpperCamelCase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _UpperCamelCase = Image.open(requests.get(__snake_case, stream=__snake_case ).raw ) return im @torch.no_grad() def lowerCamelCase__ ( __snake_case, __snake_case ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = ViTConfig() _UpperCamelCase = False # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size if vit_name[-5:] == "in21k": _UpperCamelCase = True _UpperCamelCase = int(vit_name[-12:-10] ) _UpperCamelCase = int(vit_name[-9:-6] ) else: _UpperCamelCase = 10_00 _UpperCamelCase = '''huggingface/label-files''' _UpperCamelCase = '''imagenet-1k-id2label.json''' _UpperCamelCase = json.load(open(hf_hub_download(__snake_case, __snake_case, repo_type='''dataset''' ), '''r''' ) ) _UpperCamelCase = {int(__snake_case ): v for k, v in idalabel.items()} _UpperCamelCase = idalabel _UpperCamelCase = {v: k for k, v in idalabel.items()} _UpperCamelCase = int(vit_name[-6:-4] ) _UpperCamelCase = int(vit_name[-3:] ) # size of the architecture if "deit" in vit_name: if vit_name[9:].startswith('''tiny''' ): _UpperCamelCase = 1_92 _UpperCamelCase = 7_68 _UpperCamelCase = 12 _UpperCamelCase = 3 elif vit_name[9:].startswith('''small''' ): _UpperCamelCase = 3_84 _UpperCamelCase = 15_36 _UpperCamelCase = 12 _UpperCamelCase = 6 else: pass else: if vit_name[4:].startswith('''small''' ): _UpperCamelCase = 7_68 _UpperCamelCase = 23_04 _UpperCamelCase = 8 _UpperCamelCase = 8 elif vit_name[4:].startswith('''base''' ): pass elif vit_name[4:].startswith('''large''' ): _UpperCamelCase = 10_24 _UpperCamelCase = 40_96 _UpperCamelCase = 24 _UpperCamelCase = 16 elif vit_name[4:].startswith('''huge''' ): _UpperCamelCase = 12_80 _UpperCamelCase = 51_20 _UpperCamelCase = 32 _UpperCamelCase = 16 # load original model from timm _UpperCamelCase = timm.create_model(__snake_case, pretrained=__snake_case ) timm_model.eval() # load state_dict of original model, remove and rename some keys _UpperCamelCase = timm_model.state_dict() if base_model: remove_classification_head_(__snake_case ) _UpperCamelCase = create_rename_keys(__snake_case, __snake_case ) for src, dest in rename_keys: rename_key(__snake_case, __snake_case, __snake_case ) read_in_q_k_v(__snake_case, __snake_case, __snake_case ) # load HuggingFace model if vit_name[-5:] == "in21k": _UpperCamelCase = ViTModel(__snake_case ).eval() else: _UpperCamelCase = ViTForImageClassification(__snake_case ).eval() model.load_state_dict(__snake_case ) # Check outputs on an image, prepared by ViTImageProcessor/DeiTImageProcessor if "deit" in vit_name: _UpperCamelCase = DeiTImageProcessor(size=config.image_size ) else: _UpperCamelCase = ViTImageProcessor(size=config.image_size ) _UpperCamelCase = image_processor(images=prepare_img(), return_tensors='''pt''' ) _UpperCamelCase = encoding['''pixel_values'''] _UpperCamelCase = model(__snake_case ) if base_model: _UpperCamelCase = timm_model.forward_features(__snake_case ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(__snake_case, outputs.pooler_output, atol=1e-3 ) else: _UpperCamelCase = timm_model(__snake_case ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__snake_case, outputs.logits, atol=1e-3 ) Path(__snake_case ).mkdir(exist_ok=__snake_case ) print(F'''Saving model {vit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(__snake_case ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(__snake_case ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( """--vit_name""", default="""vit_base_patch16_224""", type=str, help="""Name of the ViT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) _a = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path)
78
1
"""simple docstring""" import time from contextlib import contextmanager from pathlib import Path import pytest import requests from huggingface_hub.hf_api import HfApi, HfFolder _a = """__DUMMY_TRANSFORMERS_USER__""" _a = """Dummy User""" _a = """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""" _a = """https://hub-ci.huggingface.co""" _a = CI_HUB_ENDPOINT + """/datasets/{repo_id}/resolve/{revision}/{path}""" _a = CI_HUB_ENDPOINT + """/{repo_id}/resolve/{revision}/{filename}""" _a = Path("""~/.huggingface/hub_ci_token""").expanduser() @pytest.fixture def lowerCamelCase__ ( __snake_case ) -> Dict: """simple docstring""" monkeypatch.setattr( '''huggingface_hub.file_download.HUGGINGFACE_CO_URL_TEMPLATE''', __snake_case ) @pytest.fixture def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" monkeypatch.setattr('''datasets.config.HF_ENDPOINT''', __snake_case ) monkeypatch.setattr('''datasets.config.HUB_DATASETS_URL''', __snake_case ) @pytest.fixture def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" monkeypatch.setattr('''huggingface_hub.hf_api.HfFolder.path_token''', __snake_case ) @pytest.fixture def lowerCamelCase__ ( __snake_case, __snake_case ) -> Any: """simple docstring""" HfFolder.save_token(__snake_case ) yield HfFolder.delete_token() @pytest.fixture(scope='''session''' ) def lowerCamelCase__ ( ) -> Dict: """simple docstring""" return HfApi(endpoint=__snake_case ) @pytest.fixture(scope='''session''' ) def lowerCamelCase__ ( __snake_case ) -> Dict: """simple docstring""" _UpperCamelCase = HfFolder.get_token() HfFolder.save_token(__snake_case ) yield CI_HUB_USER_TOKEN if previous_token is not None: HfFolder.save_token(__snake_case ) @pytest.fixture def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" def _cleanup_repo(__snake_case ): hf_api.delete_repo(__snake_case, token=__snake_case, repo_type='''dataset''' ) return _cleanup_repo @pytest.fixture def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" @contextmanager def _temporary_repo(__snake_case ): try: yield repo_id finally: cleanup_repo(__snake_case ) return _temporary_repo @pytest.fixture(scope='''session''' ) def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = F'''repo_txt_data-{int(time.time() * 10e3 )}''' _UpperCamelCase = F'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(__snake_case, token=__snake_case, repo_type='''dataset''', private=__snake_case ) hf_api.upload_file( token=__snake_case, path_or_fileobj=str(__snake_case ), path_in_repo='''data/text_data.txt''', repo_id=__snake_case, repo_type='''dataset''', ) yield repo_id try: hf_api.delete_repo(__snake_case, token=__snake_case, repo_type='''dataset''' ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> Optional[int]: """simple docstring""" return hf_private_dataset_repo_txt_data_ @pytest.fixture(scope='''session''' ) def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> List[Any]: """simple docstring""" _UpperCamelCase = F'''repo_zipped_txt_data-{int(time.time() * 10e3 )}''' _UpperCamelCase = F'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(__snake_case, token=__snake_case, repo_type='''dataset''', private=__snake_case ) hf_api.upload_file( token=__snake_case, path_or_fileobj=str(__snake_case ), path_in_repo='''data.zip''', repo_id=__snake_case, repo_type='''dataset''', ) yield repo_id try: hf_api.delete_repo(__snake_case, token=__snake_case, repo_type='''dataset''' ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> Tuple: """simple docstring""" return hf_private_dataset_repo_zipped_txt_data_ @pytest.fixture(scope='''session''' ) def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> Tuple: """simple docstring""" _UpperCamelCase = F'''repo_zipped_img_data-{int(time.time() * 10e3 )}''' _UpperCamelCase = F'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(__snake_case, token=__snake_case, repo_type='''dataset''', private=__snake_case ) hf_api.upload_file( token=__snake_case, path_or_fileobj=str(__snake_case ), path_in_repo='''data.zip''', repo_id=__snake_case, repo_type='''dataset''', ) yield repo_id try: hf_api.delete_repo(__snake_case, token=__snake_case, repo_type='''dataset''' ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> Tuple: """simple docstring""" return hf_private_dataset_repo_zipped_img_data_
78
"""simple docstring""" import unittest from transformers import AlbertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForPreTraining, AlbertForQuestionAnswering, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertModel, ) from transformers.models.albert.modeling_albert import ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST class _UpperCAmelCase: def __init__( self , __a , __a=13 , __a=7 , __a=True , __a=True , __a=True , __a=True , __a=99 , __a=16 , __a=36 , __a=6 , __a=6 , __a=6 , __a=37 , __a="gelu" , __a=0.1 , __a=0.1 , __a=5_12 , __a=16 , __a=2 , __a=0.02 , __a=3 , __a=4 , __a=None , ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = seq_length _UpperCamelCase = is_training _UpperCamelCase = use_input_mask _UpperCamelCase = use_token_type_ids _UpperCamelCase = use_labels _UpperCamelCase = vocab_size _UpperCamelCase = embedding_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_hidden_groups _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = type_vocab_size _UpperCamelCase = type_sequence_label_size _UpperCamelCase = initializer_range _UpperCamelCase = num_labels _UpperCamelCase = num_choices _UpperCamelCase = scope def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) _UpperCamelCase = None if self.use_input_mask: _UpperCamelCase = random_attention_mask([self.batch_size, self.seq_length]) _UpperCamelCase = None if self.use_token_type_ids: _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size) _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) _UpperCamelCase = ids_tensor([self.batch_size] , self.num_choices) _UpperCamelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase ( self) -> Dict: '''simple docstring''' return AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , num_hidden_groups=self.num_hidden_groups , ) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Optional[int]: '''simple docstring''' _UpperCamelCase = AlbertModel(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , token_type_ids=__a) _UpperCamelCase = model(__a , token_type_ids=__a) _UpperCamelCase = model(__a) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = AlbertForPreTraining(config=__a) model.to(__a) model.eval() _UpperCamelCase = model( __a , attention_mask=__a , token_type_ids=__a , labels=__a , sentence_order_label=__a , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) self.parent.assertEqual(result.sop_logits.shape , (self.batch_size, config.num_labels)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Optional[int]: '''simple docstring''' _UpperCamelCase = AlbertForMaskedLM(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Any: '''simple docstring''' _UpperCamelCase = AlbertForQuestionAnswering(config=__a) model.to(__a) model.eval() _UpperCamelCase = model( __a , attention_mask=__a , token_type_ids=__a , start_positions=__a , end_positions=__a , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> List[Any]: '''simple docstring''' _UpperCamelCase = self.num_labels _UpperCamelCase = AlbertForSequenceClassification(__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> List[str]: '''simple docstring''' _UpperCamelCase = self.num_labels _UpperCamelCase = AlbertForTokenClassification(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.num_choices _UpperCamelCase = AlbertForMultipleChoice(config=__a) model.to(__a) model.eval() _UpperCamelCase = input_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() _UpperCamelCase = token_type_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() _UpperCamelCase = input_mask.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() _UpperCamelCase = model( __a , attention_mask=__a , token_type_ids=__a , labels=__a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices)) def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.prepare_config_and_inputs() ( ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ) = config_and_inputs _UpperCamelCase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class _UpperCAmelCase( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase__ = ( ( AlbertModel, AlbertForPreTraining, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertForQuestionAnswering, ) if is_torch_available() else () ) lowercase__ = ( { 'feature-extraction': AlbertModel, 'fill-mask': AlbertForMaskedLM, 'question-answering': AlbertForQuestionAnswering, 'text-classification': AlbertForSequenceClassification, 'token-classification': AlbertForTokenClassification, 'zero-shot': AlbertForSequenceClassification, } if is_torch_available() else {} ) lowercase__ = True def UpperCAmelCase ( self , __a , __a , __a=False) -> List[str]: '''simple docstring''' _UpperCamelCase = super()._prepare_for_class(__a , __a , return_labels=__a) if return_labels: if model_class in get_values(__a): _UpperCamelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=__a) _UpperCamelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__a) return inputs_dict def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = AlbertModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=__a , hidden_size=37) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*__a) def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__a) def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__a) def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__a) def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__a) def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _UpperCamelCase = type self.model_tester.create_and_check_model(*__a) @slow def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' for model_name in ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase = AlbertModel.from_pretrained(__a) self.assertIsNotNone(__a) @require_torch class _UpperCAmelCase( unittest.TestCase ): @slow def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = AlbertModel.from_pretrained('''albert-base-v2''') _UpperCamelCase = torch.tensor([[0, 3_45, 2_32, 3_28, 7_40, 1_40, 16_95, 69, 60_78, 15_88, 2]]) _UpperCamelCase = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]) with torch.no_grad(): _UpperCamelCase = model(__a , attention_mask=__a)[0] _UpperCamelCase = torch.Size((1, 11, 7_68)) self.assertEqual(output.shape , __a) _UpperCamelCase = torch.tensor( [[[-0.6513, 1.5035, -0.2766], [-0.6515, 1.5046, -0.2780], [-0.6512, 1.5049, -0.2784]]]) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , __a , atol=1e-4))
78
1
"""simple docstring""" import argparse import json import os import fairseq import torch from torch import nn from transformers import ( SpeechaTextaConfig, SpeechaTextaForCausalLM, SpeechaTextaTokenizer, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() _a = logging.get_logger(__name__) _a = { """post_extract_proj""": """feature_projection.projection""", """encoder.pos_conv.0""": """encoder.pos_conv_embed.conv""", """self_attn.k_proj""": """encoder.layers.*.attention.k_proj""", """self_attn.v_proj""": """encoder.layers.*.attention.v_proj""", """self_attn.q_proj""": """encoder.layers.*.attention.q_proj""", """self_attn.out_proj""": """encoder.layers.*.attention.out_proj""", """self_attn_layer_norm""": """encoder.layers.*.layer_norm""", """fc1""": """encoder.layers.*.feed_forward.intermediate_dense""", """fc2""": """encoder.layers.*.feed_forward.output_dense""", """final_layer_norm""": """encoder.layers.*.final_layer_norm""", """encoder.layer_norm""": """encoder.layer_norm""", """w2v_model.layer_norm""": """feature_projection.layer_norm""", """quantizer.weight_proj""": """quantizer.weight_proj""", """quantizer.vars""": """quantizer.codevectors""", """project_q""": """project_q""", """final_proj""": """project_hid""", """w2v_encoder.proj""": """lm_head""", """mask_emb""": """masked_spec_embed""", } _a = [ """lm_head""", """quantizer.weight_proj""", """quantizer.codevectors""", """project_q""", """project_hid""", ] def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case, __snake_case, __snake_case ) -> Tuple: """simple docstring""" for attribute in key.split('''.''' ): _UpperCamelCase = getattr(__snake_case, __snake_case ) if weight_type is not None: _UpperCamelCase = getattr(__snake_case, __snake_case ).shape else: _UpperCamelCase = hf_pointer.shape assert hf_shape == value.shape, ( F'''Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be''' F''' {value.shape} for {full_name}''' ) if weight_type == "weight": _UpperCamelCase = value elif weight_type == "weight_g": _UpperCamelCase = value elif weight_type == "weight_v": _UpperCamelCase = value elif weight_type == "bias": _UpperCamelCase = value else: _UpperCamelCase = value logger.info(F'''{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.''' ) def lowerCamelCase__ ( __snake_case, __snake_case ) -> List[str]: """simple docstring""" _UpperCamelCase = [] _UpperCamelCase = fairseq_model.state_dict() _UpperCamelCase = hf_model.feature_extractor # if encoder has different dim to decoder -> use proj_weight _UpperCamelCase = None for name, value in fairseq_dict.items(): _UpperCamelCase = False if "conv_layers" in name: load_conv_layer( __snake_case, __snake_case, __snake_case, __snake_case, hf_model.config.feat_extract_norm == '''group''', ) _UpperCamelCase = True elif name.split('''.''' )[0] == "proj": _UpperCamelCase = fairseq_model.proj _UpperCamelCase = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: _UpperCamelCase = True if "*" in mapped_key: _UpperCamelCase = name.split(__snake_case )[0].split('''.''' )[-2] _UpperCamelCase = mapped_key.replace('''*''', __snake_case ) if "weight_g" in name: _UpperCamelCase = '''weight_g''' elif "weight_v" in name: _UpperCamelCase = '''weight_v''' elif "bias" in name: _UpperCamelCase = '''bias''' elif "weight" in name: _UpperCamelCase = '''weight''' else: _UpperCamelCase = None set_recursively(__snake_case, __snake_case, __snake_case, __snake_case, __snake_case ) continue if not is_used: unused_weights.append(__snake_case ) logger.warning(F'''Unused weights: {unused_weights}''' ) return proj_weight def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case, __snake_case, __snake_case ) -> int: """simple docstring""" _UpperCamelCase = full_name.split('''conv_layers.''' )[-1] _UpperCamelCase = name.split('''.''' ) _UpperCamelCase = int(items[0] ) _UpperCamelCase = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.''' ) _UpperCamelCase = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.''' ) _UpperCamelCase = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'''{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was''' " found." ) _UpperCamelCase = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.''' ) _UpperCamelCase = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) else: unused_weights.append(__snake_case ) def lowerCamelCase__ ( __snake_case ) -> Any: """simple docstring""" _UpperCamelCase , _UpperCamelCase = emb.weight.shape _UpperCamelCase = nn.Linear(__snake_case, __snake_case, bias=__snake_case ) _UpperCamelCase = emb.weight.data return lin_layer def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" with open(__snake_case, '''r''', encoding='''utf-8''' ) as f: _UpperCamelCase = f.readlines() _UpperCamelCase = [line.split(''' ''' )[0] for line in lines] _UpperCamelCase = len(__snake_case ) _UpperCamelCase = { '''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3, } vocab_dict.update(dict(zip(__snake_case, range(4, num_words + 4 ) ) ) ) return vocab_dict @torch.no_grad() def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case, __snake_case, __snake_case, __snake_case, __snake_case, ) -> List[Any]: """simple docstring""" _UpperCamelCase = WavaVecaConfig.from_pretrained(__snake_case ) _UpperCamelCase = SpeechaTextaConfig.from_pretrained( __snake_case, vocab_size=__snake_case, decoder_layers=__snake_case, do_stable_layer_norm=__snake_case ) _UpperCamelCase = WavaVecaFeatureExtractor( feature_size=1, sampling_rate=1_60_00, padding_value=0, do_normalize=__snake_case, return_attention_mask=__snake_case, ) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path], arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) _UpperCamelCase = model[0].eval() # set weights for wav2vec2 encoder _UpperCamelCase = WavaVecaModel(__snake_case ) _UpperCamelCase = recursively_load_weights_wavaveca(model.encoder, __snake_case ) _UpperCamelCase = SpeechaTextaForCausalLM(__snake_case ) _UpperCamelCase , _UpperCamelCase = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict(), strict=__snake_case ) # set output linear layer unexpected_keys.remove('''embed_out''' ) _UpperCamelCase = nn.Parameter(model.decoder.embed_out.detach() ) # layer norm is init to identity matrix so leaving it is fine logger.warning(F'''The following keys are missing when loading the decoder weights: {missing_keys}''' ) logger.warning(F'''The following keys are unexpected when loading the decoder weights: {unexpected_keys}''' ) _UpperCamelCase = SpeechEncoderDecoderModel(encoder=__snake_case, decoder=__snake_case ) _UpperCamelCase = False # add projection layer _UpperCamelCase = nn.Parameter(projection_layer.weight ) _UpperCamelCase = nn.Parameter(projection_layer.bias ) _UpperCamelCase = create_vocab_dict(__snake_case ) with open(os.path.join(__snake_case, '''vocab.json''' ), '''w''' ) as fp: json.dump(__snake_case, __snake_case ) _UpperCamelCase = SpeechaTextaTokenizer(os.path.join(__snake_case, '''vocab.json''' ) ) tokenizer.save_pretrained(__snake_case ) _UpperCamelCase = hf_wavavec.config.to_dict() _UpperCamelCase = tokenizer.pad_token_id _UpperCamelCase = tokenizer.bos_token_id _UpperCamelCase = tokenizer.eos_token_id _UpperCamelCase = '''speech_to_text_2''' _UpperCamelCase = '''wav2vec2''' _UpperCamelCase = SpeechEncoderDecoderConfig.from_dict(__snake_case ) hf_wavavec.save_pretrained(__snake_case ) feature_extractor.save_pretrained(__snake_case ) if __name__ == "__main__": _a = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--dict_path""", default=None, type=str, help="""Path to dict of fine-tuned model""") parser.add_argument( """--encoder_config_path""", default="""facebook/wav2vec2-large-lv60""", type=str, help="""Path to hf encoder wav2vec2 checkpoint config""", ) parser.add_argument( """--decoder_config_path""", default="""facebook/s2t-small-mustc-en-fr-st""", type=str, help="""Path to hf decoder s2t checkpoint config""", ) parser.add_argument("""--vocab_size""", default=1_0224, type=int, help="""Vocab size of decoder""") parser.add_argument("""--num_decoder_layers""", default=7, type=int, help="""Number of decoder layers""") _a = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, vocab_size=args.vocab_size, num_decoder_layers=args.num_decoder_layers, )
78
"""simple docstring""" import os from typing import BinaryIO, Optional, Union import numpy as np import pyarrow.parquet as pq from .. import Audio, Dataset, Features, Image, NamedSplit, Value, config from ..features.features import FeatureType, _visit from ..formatting import query_table from ..packaged_modules import _PACKAGED_DATASETS_MODULES from ..packaged_modules.parquet.parquet import Parquet from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" _UpperCamelCase = np.inf def set_batch_size(__snake_case ) -> None: nonlocal batch_size if isinstance(__snake_case, __snake_case ): _UpperCamelCase = min(__snake_case, config.PARQUET_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS ) elif isinstance(__snake_case, __snake_case ): _UpperCamelCase = min(__snake_case, config.PARQUET_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS ) elif isinstance(__snake_case, __snake_case ) and feature.dtype == "binary": _UpperCamelCase = min(__snake_case, config.PARQUET_ROW_GROUP_SIZE_FOR_BINARY_DATASETS ) _visit(__snake_case, __snake_case ) return None if batch_size is np.inf else batch_size class _UpperCAmelCase( lowerCamelCase ): def __init__( self , __a , __a = None , __a = None , __a = None , __a = False , __a = False , __a = None , **__a , ) -> Dict: '''simple docstring''' super().__init__( __a , split=__a , features=__a , cache_dir=__a , keep_in_memory=__a , streaming=__a , num_proc=__a , **__a , ) _UpperCamelCase = path_or_paths if isinstance(__a , __a) else {self.split: path_or_paths} _UpperCamelCase = _PACKAGED_DATASETS_MODULES['''parquet'''][1] _UpperCamelCase = Parquet( cache_dir=__a , data_files=__a , features=__a , hash=__a , **__a , ) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' # Build iterable dataset if self.streaming: _UpperCamelCase = self.builder.as_streaming_dataset(split=self.split) # Build regular (map-style) dataset else: _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None self.builder.download_and_prepare( download_config=__a , download_mode=__a , verification_mode=__a , base_path=__a , num_proc=self.num_proc , ) _UpperCamelCase = self.builder.as_dataset( split=self.split , verification_mode=__a , in_memory=self.keep_in_memory) return dataset class _UpperCAmelCase: def __init__( self , __a , __a , __a = None , **__a , ) -> Dict: '''simple docstring''' _UpperCamelCase = dataset _UpperCamelCase = path_or_buf _UpperCamelCase = batch_size or get_writer_batch_size(dataset.features) _UpperCamelCase = parquet_writer_kwargs def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = self.batch_size if self.batch_size else config.DEFAULT_MAX_BATCH_SIZE if isinstance(self.path_or_buf , (str, bytes, os.PathLike)): with open(self.path_or_buf , '''wb+''') as buffer: _UpperCamelCase = self._write(file_obj=__a , batch_size=__a , **self.parquet_writer_kwargs) else: _UpperCamelCase = self._write(file_obj=self.path_or_buf , batch_size=__a , **self.parquet_writer_kwargs) return written def UpperCAmelCase ( self , __a , __a , **__a) -> int: '''simple docstring''' _UpperCamelCase = 0 _UpperCamelCase = parquet_writer_kwargs.pop('''path_or_buf''' , __a) _UpperCamelCase = self.dataset.features.arrow_schema _UpperCamelCase = pq.ParquetWriter(__a , schema=__a , **__a) for offset in logging.tqdm( range(0 , len(self.dataset) , __a) , unit='''ba''' , disable=not logging.is_progress_bar_enabled() , desc='''Creating parquet from Arrow format''' , ): _UpperCamelCase = query_table( table=self.dataset._data , key=slice(__a , offset + batch_size) , indices=self.dataset._indices if self.dataset._indices is not None else None , ) writer.write_table(__a) written += batch.nbytes writer.close() return written
78
1
"""simple docstring""" import json import multiprocessing import os import re from collections import defaultdict import torch from accelerate import Accelerator from accelerate.utils import set_seed from arguments import HumanEvalArguments from datasets import load_dataset, load_metric from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from tqdm import tqdm import transformers from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, StoppingCriteria, StoppingCriteriaList _a = ["""\nclass""", """\ndef""", """\n#""", """\n@""", """\nprint""", """\nif"""] class _UpperCAmelCase( lowerCamelCase ): def __init__( self , __a , __a , __a=None , __a=1) -> Dict: '''simple docstring''' _UpperCamelCase = tokenizer _UpperCamelCase = dataset _UpperCamelCase = len(__a) if n_tasks is None else n_tasks _UpperCamelCase = n_copies def __iter__( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = [] for task in range(self.n_tasks): # without strip, the model generate commented codes ... prompts.append(self.tokenizer.eos_token + self.dataset[task]['''prompt'''].strip()) _UpperCamelCase = self.tokenizer(__a , padding=__a , return_tensors='''pt''') for task in range(self.n_tasks): for _ in range(self.n_copies): yield { "ids": outputs.input_ids[task], "task_id": task, "input_len": outputs.attention_mask[task].sum(), } class _UpperCAmelCase( lowerCamelCase ): def __init__( self , __a , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = start_length _UpperCamelCase = eof_strings _UpperCamelCase = tokenizer def __call__( self , __a , __a , **__a) -> List[Any]: '''simple docstring''' _UpperCamelCase = self.tokenizer.batch_decode(input_ids[:, self.start_length :]) _UpperCamelCase = [] for decoded_generation in decoded_generations: done.append(any(stop_string in decoded_generation for stop_string in self.eof_strings)) return all(__a) def lowerCamelCase__ ( __snake_case ) -> Dict: """simple docstring""" _UpperCamelCase = re.split('''(%s)''' % '''|'''.join(__snake_case ), __snake_case ) # last string should be "" return "".join(string_list[:-2] ) def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case, __snake_case, __snake_case, __snake_case=20, **__snake_case ) -> Any: """simple docstring""" _UpperCamelCase = defaultdict(__snake_case ) # dict of list of generated tokens for step, batch in tqdm(enumerate(__snake_case ) ): with torch.no_grad(): _UpperCamelCase = batch['''ids'''].shape[-1] _UpperCamelCase = accelerator.unwrap_model(__snake_case ).generate( input_ids=batch['''ids'''][:, : batch['''input_len''']], num_return_sequences=__snake_case, **__snake_case ) # each task is generated batch_size times _UpperCamelCase = batch['''task_id'''].repeat(__snake_case ) _UpperCamelCase = accelerator.pad_across_processes( __snake_case, dim=1, pad_index=tokenizer.pad_token_id ) _UpperCamelCase , _UpperCamelCase = accelerator.gather((generated_tokens, generated_tasks) ) _UpperCamelCase = generated_tokens.cpu().numpy() _UpperCamelCase = generated_tasks.cpu().numpy() for task, generated_tokens in zip(__snake_case, __snake_case ): gen_token_dict[task].append(__snake_case ) _UpperCamelCase = [[] for _ in range(__snake_case )] for task, generated_tokens in gen_token_dict.items(): for s in generated_tokens: _UpperCamelCase = tokenizer.decode(__snake_case, skip_special_tokens=__snake_case, clean_up_tokenization_spaces=__snake_case ) code_gens[task].append(remove_last_block(__snake_case ) ) return code_gens def lowerCamelCase__ ( ) -> Optional[Any]: """simple docstring""" _UpperCamelCase = HfArgumentParser(__snake_case ) _UpperCamelCase = parser.parse_args() transformers.logging.set_verbosity_error() # enables code execution in code_eval metric _UpperCamelCase = args.HF_ALLOW_CODE_EVAL # make sure tokenizer plays nice with multiprocessing _UpperCamelCase = '''false''' if args.num_workers is None: _UpperCamelCase = multiprocessing.cpu_count() # Use dataset load to feed to accelerate _UpperCamelCase = Accelerator() set_seed(args.seed, device_specific=__snake_case ) # Load model and tokenizer _UpperCamelCase = AutoTokenizer.from_pretrained(args.model_ckpt ) _UpperCamelCase = tokenizer.eos_token _UpperCamelCase = AutoModelForCausalLM.from_pretrained(args.model_ckpt ) # Generation settings _UpperCamelCase = { '''do_sample''': args.do_sample, '''temperature''': args.temperature, '''max_new_tokens''': args.max_new_tokens, '''top_p''': args.top_p, '''top_k''': args.top_k, '''stopping_criteria''': StoppingCriteriaList([EndOfFunctionCriteria(0, __snake_case, __snake_case )] ), } # Load evaluation dataset and metric _UpperCamelCase = load_dataset('''openai_humaneval''' ) _UpperCamelCase = load_metric('''code_eval''' ) _UpperCamelCase = args.num_tasks if args.num_tasks is not None else len(human_eval['''test'''] ) _UpperCamelCase = args.n_samples // args.batch_size _UpperCamelCase = TokenizedDataset(__snake_case, human_eval['''test'''], n_copies=__snake_case, n_tasks=__snake_case ) # do not confuse args.batch_size, which is actually the num_return_sequences _UpperCamelCase = DataLoader(__snake_case, batch_size=1 ) # Run a quick test to see if code evaluation is enabled try: _UpperCamelCase = code_eval_metric.compute(references=[''''''], predictions=[['''''']] ) except ValueError as exception: print( '''Code evaluation not enabled. Read the warning below carefully and then use `--HF_ALLOW_CODE_EVAL="1"`''' ''' flag to enable code evaluation.''' ) raise exception _UpperCamelCase , _UpperCamelCase = accelerator.prepare(__snake_case, __snake_case ) _UpperCamelCase = complete_code( __snake_case, __snake_case, __snake_case, __snake_case, n_tasks=__snake_case, batch_size=args.batch_size, **__snake_case, ) if accelerator.is_main_process: _UpperCamelCase = [] for task in tqdm(range(__snake_case ) ): _UpperCamelCase = human_eval['''test'''][task]['''test'''] _UpperCamelCase = F'''check({human_eval["test"][task]["entry_point"]})''' references.append('''\n''' + test_func + '''\n''' + entry_point ) # Evaluate completions with "code_eval" metric _UpperCamelCase , _UpperCamelCase = code_eval_metric.compute( references=__snake_case, predictions=__snake_case, num_workers=args.num_workers ) print(F'''Results: {pass_at_k}''' ) # Save results to json file with open(args.output_file, '''w''' ) as fp: json.dump(__snake_case, __snake_case ) # For some reason the folliwng seems to be necessary sometimes for code_eval to work nice with multiprocessing # https://stackoverflow.com/questions/60804599/python-multiprocessing-keeps-spawning-the-whole-script if __name__ == "__main__": main()
78
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class _UpperCAmelCase( unittest.TestCase ): def __init__( self , __a , __a=7 , __a=3 , __a=18 , __a=30 , __a=4_00 , __a=True , __a=None , __a=True , __a=None , __a=True , ) -> int: '''simple docstring''' _UpperCamelCase = size if size is not None else {'''shortest_edge''': 20} _UpperCamelCase = crop_size if crop_size is not None else {'''height''': 18, '''width''': 18} _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = num_channels _UpperCamelCase = image_size _UpperCamelCase = min_resolution _UpperCamelCase = max_resolution _UpperCamelCase = do_resize _UpperCamelCase = size _UpperCamelCase = do_center_crop _UpperCamelCase = crop_size _UpperCamelCase = do_flip_channel_order def UpperCAmelCase ( self) -> str: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_flip_channel_order": self.do_flip_channel_order, } @require_torch @require_vision class _UpperCAmelCase( lowerCamelCase , unittest.TestCase ): lowercase__ = MobileViTImageProcessor if is_vision_available() else None def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = MobileViTImageProcessingTester(self) @property def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(__a , '''do_resize''')) self.assertTrue(hasattr(__a , '''size''')) self.assertTrue(hasattr(__a , '''do_center_crop''')) self.assertTrue(hasattr(__a , '''center_crop''')) self.assertTrue(hasattr(__a , '''do_flip_channel_order''')) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size , {'''shortest_edge''': 20}) self.assertEqual(image_processor.crop_size , {'''height''': 18, '''width''': 18}) _UpperCamelCase = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84) self.assertEqual(image_processor.size , {'''shortest_edge''': 42}) self.assertEqual(image_processor.crop_size , {'''height''': 84, '''width''': 84}) def UpperCAmelCase ( self) -> Dict: '''simple docstring''' pass def UpperCAmelCase ( self) -> str: '''simple docstring''' # Initialize image_processing _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random PIL images _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__a) for image in image_inputs: self.assertIsInstance(__a , Image.Image) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(__a , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' # Initialize image_processing _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__a , numpify=__a) for image in image_inputs: self.assertIsInstance(__a , np.ndarray) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(__a , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def UpperCAmelCase ( self) -> int: '''simple docstring''' # Initialize image_processing _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__a , torchify=__a) for image in image_inputs: self.assertIsInstance(__a , torch.Tensor) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(__a , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
78
1
"""simple docstring""" def lowerCamelCase__ ( __snake_case ) -> Any: """simple docstring""" _UpperCamelCase = len(__snake_case ) for i in range(length - 1 ): _UpperCamelCase = i for k in range(i + 1, __snake_case ): if collection[k] < collection[least]: _UpperCamelCase = k if least != i: _UpperCamelCase , _UpperCamelCase = (collection[i], collection[least]) return collection if __name__ == "__main__": _a = input("""Enter numbers separated by a comma:\n""").strip() _a = [int(item) for item in user_input.split(""",""")] print(selection_sort(unsorted))
78
"""simple docstring""" import warnings from typing import List import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import is_flax_available, is_tf_available, is_torch_available class _UpperCAmelCase( lowerCamelCase ): lowercase__ = ['image_processor', 'tokenizer'] lowercase__ = 'OwlViTImageProcessor' lowercase__ = ('CLIPTokenizer', 'CLIPTokenizerFast') def __init__( self , __a=None , __a=None , **__a) -> List[Any]: '''simple docstring''' _UpperCamelCase = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , __a , ) _UpperCamelCase = kwargs.pop('''feature_extractor''') _UpperCamelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''') if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''') super().__init__(__a , __a) def __call__( self , __a=None , __a=None , __a=None , __a="max_length" , __a="np" , **__a) -> List[str]: '''simple docstring''' if text is None and query_images is None and images is None: raise ValueError( '''You have to specify at least one text or query image or image. All three cannot be none.''') if text is not None: if isinstance(__a , __a) or (isinstance(__a , __a) and not isinstance(text[0] , __a)): _UpperCamelCase = [self.tokenizer(__a , padding=__a , return_tensors=__a , **__a)] elif isinstance(__a , __a) and isinstance(text[0] , __a): _UpperCamelCase = [] # Maximum number of queries across batch _UpperCamelCase = max([len(__a) for t in text]) # Pad all batch samples to max number of text queries for t in text: if len(__a) != max_num_queries: _UpperCamelCase = t + [''' '''] * (max_num_queries - len(__a)) _UpperCamelCase = self.tokenizer(__a , padding=__a , return_tensors=__a , **__a) encodings.append(__a) else: raise TypeError('''Input text should be a string, a list of strings or a nested list of strings''') if return_tensors == "np": _UpperCamelCase = np.concatenate([encoding['''input_ids'''] for encoding in encodings] , axis=0) _UpperCamelCase = np.concatenate([encoding['''attention_mask'''] for encoding in encodings] , axis=0) elif return_tensors == "jax" and is_flax_available(): import jax.numpy as jnp _UpperCamelCase = jnp.concatenate([encoding['''input_ids'''] for encoding in encodings] , axis=0) _UpperCamelCase = jnp.concatenate([encoding['''attention_mask'''] for encoding in encodings] , axis=0) elif return_tensors == "pt" and is_torch_available(): import torch _UpperCamelCase = torch.cat([encoding['''input_ids'''] for encoding in encodings] , dim=0) _UpperCamelCase = torch.cat([encoding['''attention_mask'''] for encoding in encodings] , dim=0) elif return_tensors == "tf" and is_tf_available(): import tensorflow as tf _UpperCamelCase = tf.stack([encoding['''input_ids'''] for encoding in encodings] , axis=0) _UpperCamelCase = tf.stack([encoding['''attention_mask'''] for encoding in encodings] , axis=0) else: raise ValueError('''Target return tensor type could not be returned''') _UpperCamelCase = BatchEncoding() _UpperCamelCase = input_ids _UpperCamelCase = attention_mask if query_images is not None: _UpperCamelCase = BatchEncoding() _UpperCamelCase = self.image_processor( __a , return_tensors=__a , **__a).pixel_values _UpperCamelCase = query_pixel_values if images is not None: _UpperCamelCase = self.image_processor(__a , return_tensors=__a , **__a) if text is not None and images is not None: _UpperCamelCase = image_features.pixel_values return encoding elif query_images is not None and images is not None: _UpperCamelCase = image_features.pixel_values return encoding elif text is not None or query_images is not None: return encoding else: return BatchEncoding(data=dict(**__a) , tensor_type=__a) def UpperCAmelCase ( self , *__a , **__a) -> str: '''simple docstring''' return self.image_processor.post_process(*__a , **__a) def UpperCAmelCase ( self , *__a , **__a) -> Dict: '''simple docstring''' return self.image_processor.post_process_object_detection(*__a , **__a) def UpperCAmelCase ( self , *__a , **__a) -> Optional[Any]: '''simple docstring''' return self.image_processor.post_process_image_guided_detection(*__a , **__a) def UpperCAmelCase ( self , *__a , **__a) -> Optional[int]: '''simple docstring''' return self.tokenizer.batch_decode(*__a , **__a) def UpperCAmelCase ( self , *__a , **__a) -> Optional[Any]: '''simple docstring''' return self.tokenizer.decode(*__a , **__a) @property def UpperCAmelCase ( self) -> str: '''simple docstring''' warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , __a , ) return self.image_processor_class @property def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , __a , ) return self.image_processor
78
1
"""simple docstring""" import collections import importlib.util import os import re from pathlib import Path _a = """src/transformers""" # Matches is_xxx_available() _a = re.compile(R"""is\_([a-z_]*)_available()""") # Catches a one-line _import_struct = {xxx} _a = re.compile(R"""^_import_structure\s+=\s+\{([^\}]+)\}""") # Catches a line with a key-values pattern: "bla": ["foo", "bar"] _a = re.compile(R"""\s+\"\S*\":\s+\[([^\]]*)\]""") # Catches a line if not is_foo_available _a = re.compile(R"""^\s*if\s+not\s+is\_[a-z_]*\_available\(\)""") # Catches a line _import_struct["bla"].append("foo") _a = re.compile(R"""^\s*_import_structure\[\"\S*\"\]\.append\(\"(\S*)\"\)""") # Catches a line _import_struct["bla"].extend(["foo", "bar"]) or _import_struct["bla"] = ["foo", "bar"] _a = re.compile(R"""^\s*_import_structure\[\S*\](?:\.extend\(|\s*=\s+)\[([^\]]*)\]""") # Catches a line with an object between quotes and a comma: "MyModel", _a = re.compile("""^\s+\"([^\"]+)\",""") # Catches a line with objects between brackets only: ["foo", "bar"], _a = re.compile("""^\s+\[([^\]]+)\]""") # Catches a line with from foo import bar, bla, boo _a = re.compile(R"""\s+from\s+\S*\s+import\s+([^\(\s].*)\n""") # Catches a line with try: _a = re.compile(R"""^\s*try:""") # Catches a line with else: _a = re.compile(R"""^\s*else:""") def lowerCamelCase__ ( __snake_case ) -> List[str]: """simple docstring""" if _re_test_backend.search(__snake_case ) is None: return None _UpperCamelCase = [b[0] for b in _re_backend.findall(__snake_case )] backends.sort() return "_and_".join(__snake_case ) def lowerCamelCase__ ( __snake_case ) -> Any: """simple docstring""" with open(__snake_case, '''r''', encoding='''utf-8''', newline='''\n''' ) as f: _UpperCamelCase = f.readlines() _UpperCamelCase = 0 while line_index < len(__snake_case ) and not lines[line_index].startswith('''_import_structure = {''' ): line_index += 1 # If this is a traditional init, just return. if line_index >= len(__snake_case ): return None # First grab the objects without a specific backend in _import_structure _UpperCamelCase = [] while not lines[line_index].startswith('''if TYPE_CHECKING''' ) and find_backend(lines[line_index] ) is None: _UpperCamelCase = lines[line_index] # If we have everything on a single line, let's deal with it. if _re_one_line_import_struct.search(__snake_case ): _UpperCamelCase = _re_one_line_import_struct.search(__snake_case ).groups()[0] _UpperCamelCase = re.findall('''\[([^\]]+)\]''', __snake_case ) for imp in imports: objects.extend([obj[1:-1] for obj in imp.split(''', ''' )] ) line_index += 1 continue _UpperCamelCase = _re_import_struct_key_value.search(__snake_case ) if single_line_import_search is not None: _UpperCamelCase = [obj[1:-1] for obj in single_line_import_search.groups()[0].split(''', ''' ) if len(__snake_case ) > 0] objects.extend(__snake_case ) elif line.startswith(''' ''' * 8 + '''"''' ): objects.append(line[9:-3] ) line_index += 1 _UpperCamelCase = {'''none''': objects} # Let's continue with backend-specific objects in _import_structure while not lines[line_index].startswith('''if TYPE_CHECKING''' ): # If the line is an if not is_backend_available, we grab all objects associated. _UpperCamelCase = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: _UpperCamelCase = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 _UpperCamelCase = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(''' ''' * 4 ): _UpperCamelCase = lines[line_index] if _re_import_struct_add_one.search(__snake_case ) is not None: objects.append(_re_import_struct_add_one.search(__snake_case ).groups()[0] ) elif _re_import_struct_add_many.search(__snake_case ) is not None: _UpperCamelCase = _re_import_struct_add_many.search(__snake_case ).groups()[0].split(''', ''' ) _UpperCamelCase = [obj[1:-1] for obj in imports if len(__snake_case ) > 0] objects.extend(__snake_case ) elif _re_between_brackets.search(__snake_case ) is not None: _UpperCamelCase = _re_between_brackets.search(__snake_case ).groups()[0].split(''', ''' ) _UpperCamelCase = [obj[1:-1] for obj in imports if len(__snake_case ) > 0] objects.extend(__snake_case ) elif _re_quote_object.search(__snake_case ) is not None: objects.append(_re_quote_object.search(__snake_case ).groups()[0] ) elif line.startswith(''' ''' * 8 + '''"''' ): objects.append(line[9:-3] ) elif line.startswith(''' ''' * 12 + '''"''' ): objects.append(line[13:-3] ) line_index += 1 _UpperCamelCase = objects else: line_index += 1 # At this stage we are in the TYPE_CHECKING part, first grab the objects without a specific backend _UpperCamelCase = [] while ( line_index < len(__snake_case ) and find_backend(lines[line_index] ) is None and not lines[line_index].startswith('''else''' ) ): _UpperCamelCase = lines[line_index] _UpperCamelCase = _re_import.search(__snake_case ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(''', ''' ) ) elif line.startswith(''' ''' * 8 ): objects.append(line[8:-2] ) line_index += 1 _UpperCamelCase = {'''none''': objects} # Let's continue with backend-specific objects while line_index < len(__snake_case ): # If the line is an if is_backend_available, we grab all objects associated. _UpperCamelCase = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: _UpperCamelCase = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 _UpperCamelCase = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(''' ''' * 8 ): _UpperCamelCase = lines[line_index] _UpperCamelCase = _re_import.search(__snake_case ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(''', ''' ) ) elif line.startswith(''' ''' * 12 ): objects.append(line[12:-2] ) line_index += 1 _UpperCamelCase = objects else: line_index += 1 return import_dict_objects, type_hint_objects def lowerCamelCase__ ( __snake_case, __snake_case ) -> Tuple: """simple docstring""" def find_duplicates(__snake_case ): return [k for k, v in collections.Counter(__snake_case ).items() if v > 1] if list(import_dict_objects.keys() ) != list(type_hint_objects.keys() ): return ["Both sides of the init do not have the same backends!"] _UpperCamelCase = [] for key in import_dict_objects.keys(): _UpperCamelCase = find_duplicates(import_dict_objects[key] ) if duplicate_imports: errors.append(F'''Duplicate _import_structure definitions for: {duplicate_imports}''' ) _UpperCamelCase = find_duplicates(type_hint_objects[key] ) if duplicate_type_hints: errors.append(F'''Duplicate TYPE_CHECKING objects for: {duplicate_type_hints}''' ) if sorted(set(import_dict_objects[key] ) ) != sorted(set(type_hint_objects[key] ) ): _UpperCamelCase = '''base imports''' if key == '''none''' else F'''{key} backend''' errors.append(F'''Differences for {name}:''' ) for a in type_hint_objects[key]: if a not in import_dict_objects[key]: errors.append(F''' {a} in TYPE_HINT but not in _import_structure.''' ) for a in import_dict_objects[key]: if a not in type_hint_objects[key]: errors.append(F''' {a} in _import_structure but not in TYPE_HINT.''' ) return errors def lowerCamelCase__ ( ) -> Tuple: """simple docstring""" _UpperCamelCase = [] for root, _, files in os.walk(__snake_case ): if "__init__.py" in files: _UpperCamelCase = os.path.join(__snake_case, '''__init__.py''' ) _UpperCamelCase = parse_init(__snake_case ) if objects is not None: _UpperCamelCase = analyze_results(*__snake_case ) if len(__snake_case ) > 0: _UpperCamelCase = F'''Problem in {fname}, both halves do not define the same objects.\n{errors[0]}''' failures.append('''\n'''.join(__snake_case ) ) if len(__snake_case ) > 0: raise ValueError('''\n\n'''.join(__snake_case ) ) def lowerCamelCase__ ( ) -> Dict: """simple docstring""" _UpperCamelCase = [] for path, directories, files in os.walk(__snake_case ): for folder in directories: # Ignore private modules if folder.startswith('''_''' ): directories.remove(__snake_case ) continue # Ignore leftovers from branches (empty folders apart from pycache) if len(list((Path(__snake_case ) / folder).glob('''*.py''' ) ) ) == 0: continue _UpperCamelCase = str((Path(__snake_case ) / folder).relative_to(__snake_case ) ) _UpperCamelCase = short_path.replace(os.path.sep, '''.''' ) submodules.append(__snake_case ) for fname in files: if fname == "__init__.py": continue _UpperCamelCase = str((Path(__snake_case ) / fname).relative_to(__snake_case ) ) _UpperCamelCase = short_path.replace('''.py''', '''''' ).replace(os.path.sep, '''.''' ) if len(submodule.split('''.''' ) ) == 1: submodules.append(__snake_case ) return submodules _a = [ """convert_pytorch_checkpoint_to_tf2""", """modeling_flax_pytorch_utils""", ] def lowerCamelCase__ ( ) -> str: """simple docstring""" _UpperCamelCase = importlib.util.spec_from_file_location( '''transformers''', os.path.join(__snake_case, '''__init__.py''' ), submodule_search_locations=[PATH_TO_TRANSFORMERS], ) _UpperCamelCase = spec.loader.load_module() _UpperCamelCase = [ module for module in get_transformers_submodules() if module not in IGNORE_SUBMODULES and module not in transformers._import_structure.keys() ] if len(__snake_case ) > 0: _UpperCamelCase = '''\n'''.join(F'''- {module}''' for module in module_not_registered ) raise ValueError( '''The following submodules are not properly registered in the main init of Transformers:\n''' F'''{list_of_modules}\n''' '''Make sure they appear somewhere in the keys of `_import_structure` with an empty list as value.''' ) if __name__ == "__main__": check_all_inits() check_submodules()
78
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) _a = { """configuration_perceiver""": ["""PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """PerceiverConfig""", """PerceiverOnnxConfig"""], """tokenization_perceiver""": ["""PerceiverTokenizer"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ["""PerceiverFeatureExtractor"""] _a = ["""PerceiverImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ """PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST""", """PerceiverForImageClassificationConvProcessing""", """PerceiverForImageClassificationFourier""", """PerceiverForImageClassificationLearned""", """PerceiverForMaskedLM""", """PerceiverForMultimodalAutoencoding""", """PerceiverForOpticalFlow""", """PerceiverForSequenceClassification""", """PerceiverLayer""", """PerceiverModel""", """PerceiverPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys _a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
78
1
"""simple docstring""" from __future__ import annotations from collections import namedtuple from dataclasses import dataclass @dataclass class _UpperCAmelCase: lowercase__ = 42 lowercase__ = None lowercase__ = None _a = namedtuple("""CoinsDistribResult""", """moves excess""") def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" if root is None: return 0 # Validation def count_nodes(__snake_case ) -> int: if node is None: return 0 return count_nodes(node.left ) + count_nodes(node.right ) + 1 def count_coins(__snake_case ) -> int: if node is None: return 0 return count_coins(node.left ) + count_coins(node.right ) + node.data if count_nodes(__snake_case ) != count_coins(__snake_case ): raise ValueError('''The nodes number should be same as the number of coins''' ) # Main calculation def get_distrib(__snake_case ) -> CoinsDistribResult: if node is None: return CoinsDistribResult(0, 1 ) _UpperCamelCase , _UpperCamelCase = get_distrib(node.left ) _UpperCamelCase , _UpperCamelCase = get_distrib(node.right ) _UpperCamelCase = 1 - left_distrib_excess _UpperCamelCase = 1 - right_distrib_excess _UpperCamelCase = ( left_distrib_moves + right_distrib_moves + abs(__snake_case ) + abs(__snake_case ) ) _UpperCamelCase = node.data - coins_to_left - coins_to_right return CoinsDistribResult(__snake_case, __snake_case ) return get_distrib(__snake_case )[0] if __name__ == "__main__": import doctest doctest.testmod()
78
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class _UpperCAmelCase: def __init__( self , __a , __a=13 , __a=2 , __a=24 , __a=16 , __a=True , __a=True , __a=32 , __a=5 , __a=4 , __a=37 , __a="gelu" , __a=0.1 , __a=0.1 , __a=10 , __a=0.02 , __a=None , __a=2 , __a=2 , ) -> List[str]: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = patch_size _UpperCamelCase = max_length _UpperCamelCase = num_mel_bins _UpperCamelCase = is_training _UpperCamelCase = use_labels _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = type_sequence_label_size _UpperCamelCase = initializer_range _UpperCamelCase = scope _UpperCamelCase = frequency_stride _UpperCamelCase = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) _UpperCamelCase = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 _UpperCamelCase = (self.max_length - self.patch_size) // self.time_stride + 1 _UpperCamelCase = frequency_out_dimension * time_out_dimension _UpperCamelCase = num_patches + 2 def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins]) _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size) _UpperCamelCase = self.get_config() return config, input_values, labels def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__a , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def UpperCAmelCase ( self , __a , __a , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = ASTModel(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.prepare_config_and_inputs() ( ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ) = config_and_inputs _UpperCamelCase = {'''input_values''': input_values} return config, inputs_dict @require_torch class _UpperCAmelCase( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase__ = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) lowercase__ = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) lowercase__ = False lowercase__ = False lowercase__ = False lowercase__ = False def UpperCAmelCase ( self , __a , __a , __a , __a , __a) -> Optional[Any]: '''simple docstring''' if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = ASTModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=__a , has_text_modality=__a , hidden_size=37) def UpperCAmelCase ( self) -> int: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='''AST does not use inputs_embeds''') def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' pass def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = model_class(__a) self.assertIsInstance(model.get_input_embeddings() , (nn.Module)) _UpperCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__a , nn.Linear)) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = model_class(__a) _UpperCamelCase = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCamelCase = [*signature.parameters.keys()] _UpperCamelCase = ['''input_values'''] self.assertListEqual(arg_names[:1] , __a) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a) @slow def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase = ASTModel.from_pretrained(__a) self.assertIsNotNone(__a) def lowerCamelCase__ ( ) -> List[str]: """simple docstring""" _UpperCamelCase = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''', filename='''sample_audio.flac''', repo_type='''dataset''' ) _UpperCamelCase , _UpperCamelCase = torchaudio.load(__snake_case ) return audio, sampling_rate @require_torch @require_torchaudio class _UpperCAmelCase( unittest.TestCase ): @cached_property def UpperCAmelCase ( self) -> Dict: '''simple docstring''' return ( ASTFeatureExtractor.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''') if is_torchaudio_available() else None ) @slow def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.default_feature_extractor _UpperCamelCase = ASTForAudioClassification.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''').to(__a) _UpperCamelCase = self.default_feature_extractor _UpperCamelCase , _UpperCamelCase = prepare_audio() _UpperCamelCase = audio.squeeze().numpy() _UpperCamelCase = feature_extractor(__a , sampling_rate=__a , return_tensors='''pt''').to(__a) # forward pass with torch.no_grad(): _UpperCamelCase = model(**__a) # verify the logits _UpperCamelCase = torch.Size((1, 5_27)) self.assertEqual(outputs.logits.shape , __a) _UpperCamelCase = torch.tensor([-0.8760, -7.0042, -8.6602]).to(__a) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __a , atol=1e-4))
78
1
"""simple docstring""" from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer _a = logging.get_logger(__name__) _a = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } _a = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } _a = { """facebook/blenderbot_small-90M""": 512, } class _UpperCAmelCase( lowerCamelCase ): lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = BlenderbotSmallTokenizer def __init__( self , __a=None , __a=None , __a="<|endoftext|>" , __a="<|endoftext|>" , __a="<|endoftext|>" , __a=False , __a=True , **__a , ) -> int: '''simple docstring''' super().__init__( ByteLevelBPETokenizer( vocab=__a , merges=__a , add_prefix_space=__a , trim_offsets=__a , ) , bos_token=__a , eos_token=__a , unk_token=__a , **__a , ) _UpperCamelCase = add_prefix_space def UpperCAmelCase ( self , __a , __a=None) -> Tuple: '''simple docstring''' _UpperCamelCase = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def UpperCAmelCase ( self , __a , __a = None) -> List[int]: '''simple docstring''' _UpperCamelCase = [self.sep_token_id] _UpperCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0]
78
"""simple docstring""" def lowerCamelCase__ ( ) -> list[list[int]]: """simple docstring""" return [list(range(10_00 - i, -10_00 - i, -1 ) ) for i in range(10_00 )] _a = generate_large_matrix() _a = ( [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]], [[3, 2], [1, 0]], [[7, 7, 6]], [[7, 7, 6], [-1, -2, -3]], grid, ) def lowerCamelCase__ ( __snake_case ) -> None: """simple docstring""" assert all(row == sorted(__snake_case, reverse=__snake_case ) for row in grid ) assert all(list(__snake_case ) == sorted(__snake_case, reverse=__snake_case ) for col in zip(*__snake_case ) ) def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" _UpperCamelCase = 0 _UpperCamelCase = len(__snake_case ) - 1 # Edge cases such as no values or all numbers are negative. if not array or array[0] < 0: return 0 while right + 1 > left: _UpperCamelCase = (left + right) // 2 _UpperCamelCase = array[mid] # Num must be negative and the index must be greater than or equal to 0. if num < 0 and array[mid - 1] >= 0: return mid if num >= 0: _UpperCamelCase = mid + 1 else: _UpperCamelCase = mid - 1 # No negative numbers so return the last index of the array + 1 which is the length. return len(__snake_case ) def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" _UpperCamelCase = 0 _UpperCamelCase = len(grid[0] ) for i in range(len(__snake_case ) ): _UpperCamelCase = find_negative_index(grid[i][:bound] ) total += bound return (len(__snake_case ) * len(grid[0] )) - total def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" return len([number for row in grid for number in row if number < 0] ) def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" _UpperCamelCase = 0 for row in grid: for i, number in enumerate(__snake_case ): if number < 0: total += len(__snake_case ) - i break return total def lowerCamelCase__ ( ) -> None: """simple docstring""" from timeit import timeit print('''Running benchmarks''' ) _UpperCamelCase = ( '''from __main__ import count_negatives_binary_search, ''' '''count_negatives_brute_force, count_negatives_brute_force_with_break, grid''' ) for func in ( "count_negatives_binary_search", # took 0.7727 seconds "count_negatives_brute_force_with_break", # took 4.6505 seconds "count_negatives_brute_force", # took 12.8160 seconds ): _UpperCamelCase = timeit(F'''{func}(grid=grid)''', setup=__snake_case, number=5_00 ) print(F'''{func}() took {time:0.4f} seconds''' ) if __name__ == "__main__": import doctest doctest.testmod() benchmark()
78
1
"""simple docstring""" import re import time from typing import Optional import IPython.display as disp from ..trainer_callback import TrainerCallback from ..trainer_utils import IntervalStrategy, has_length def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" _UpperCamelCase = int(__snake_case ) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = t // 36_00, (t // 60) % 60, t % 60 return F'''{h}:{m:02d}:{s:02d}''' if h != 0 else F'''{m:02d}:{s:02d}''' def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case, __snake_case, __snake_case=3_00 ) -> Dict: """simple docstring""" return F''' <div> {prefix} <progress value=\'{value}\' max=\'{total}\' style=\'width:{width}px; height:20px; vertical-align: middle;\'></progress> {label} </div> ''' def lowerCamelCase__ ( __snake_case ) -> Any: """simple docstring""" _UpperCamelCase = '''<table border="1" class="dataframe">\n''' html_code += """ <thead>\n <tr style="text-align: left;">\n""" for i in items[0]: html_code += F''' <th>{i}</th>\n''' html_code += " </tr>\n </thead>\n <tbody>\n" for line in items[1:]: html_code += " <tr>\n" for elt in line: _UpperCamelCase = F'''{elt:.6f}''' if isinstance(__snake_case, __snake_case ) else str(__snake_case ) html_code += F''' <td>{elt}</td>\n''' html_code += " </tr>\n" html_code += " </tbody>\n</table><p>" return html_code class _UpperCAmelCase: lowercase__ = 5 lowercase__ = 0.2 def __init__( self , __a , __a = None , __a = True , __a = None , __a = 3_00 , ) -> List[Any]: '''simple docstring''' _UpperCamelCase = total _UpperCamelCase = '''''' if prefix is None else prefix _UpperCamelCase = leave _UpperCamelCase = parent _UpperCamelCase = width _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None def UpperCAmelCase ( self , __a , __a = False , __a = None) -> int: '''simple docstring''' _UpperCamelCase = value if comment is not None: _UpperCamelCase = comment if self.last_value is None: _UpperCamelCase = _UpperCamelCase = time.time() _UpperCamelCase = _UpperCamelCase = value _UpperCamelCase = _UpperCamelCase = None _UpperCamelCase = self.warmup _UpperCamelCase = 1 self.update_bar(__a) elif value <= self.last_value and not force_update: return elif force_update or self.first_calls > 0 or value >= min(self.last_value + self.wait_for , self.total): if self.first_calls > 0: self.first_calls -= 1 _UpperCamelCase = time.time() _UpperCamelCase = current_time - self.start_time # We could have value = self.start_value if the update is called twixe with the same start value. if value > self.start_value: _UpperCamelCase = self.elapsed_time / (value - self.start_value) else: _UpperCamelCase = None if value >= self.total: _UpperCamelCase = self.total _UpperCamelCase = None if not self.leave: self.close() elif self.average_time_per_item is not None: _UpperCamelCase = self.average_time_per_item * (self.total - value) self.update_bar(__a) _UpperCamelCase = value _UpperCamelCase = current_time if self.average_time_per_item is None: _UpperCamelCase = 1 else: _UpperCamelCase = max(int(self.update_every / self.average_time_per_item) , 1) def UpperCAmelCase ( self , __a , __a=None) -> str: '''simple docstring''' _UpperCamelCase = ''' ''' * (len(str(self.total)) - len(str(__a))) + str(__a) if self.elapsed_time is None: _UpperCamelCase = F'''[{spaced_value}/{self.total} : < :''' elif self.predicted_remaining is None: _UpperCamelCase = F'''[{spaced_value}/{self.total} {format_time(self.elapsed_time)}''' else: _UpperCamelCase = ( F'''[{spaced_value}/{self.total} {format_time(self.elapsed_time)} <''' F''' {format_time(self.predicted_remaining)}''' ) self.label += F''', {1/self.average_time_per_item:.2f} it/s''' self.label += "]" if self.comment is None or len(self.comment) == 0 else F''', {self.comment}]''' self.display() def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = html_progress_bar(self.value , self.total , self.prefix , self.label , self.width) if self.parent is not None: # If this is a child bar, the parent will take care of the display. self.parent.display() return if self.output is None: _UpperCamelCase = disp.display(disp.HTML(self.html_code) , display_id=__a) else: self.output.update(disp.HTML(self.html_code)) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' if self.parent is None and self.output is not None: self.output.update(disp.HTML('''''')) class _UpperCAmelCase( lowerCamelCase ): def __init__( self , __a , __a=None) -> Dict: '''simple docstring''' super().__init__(__a) _UpperCamelCase = None if column_names is None else [column_names] _UpperCamelCase = None def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = html_progress_bar(self.value , self.total , self.prefix , self.label , self.width) if self.inner_table is not None: self.html_code += text_to_html_table(self.inner_table) if self.child_bar is not None: self.html_code += self.child_bar.html_code if self.output is None: _UpperCamelCase = disp.display(disp.HTML(self.html_code) , display_id=__a) else: self.output.update(disp.HTML(self.html_code)) def UpperCAmelCase ( self , __a) -> Optional[Any]: '''simple docstring''' if self.inner_table is None: _UpperCamelCase = [list(values.keys()), list(values.values())] else: _UpperCamelCase = self.inner_table[0] if len(self.inner_table) == 1: # We give a chance to update the column names at the first iteration for key in values.keys(): if key not in columns: columns.append(__a) _UpperCamelCase = columns self.inner_table.append([values[c] for c in columns]) def UpperCAmelCase ( self , __a , __a=None , __a=3_00) -> List[str]: '''simple docstring''' _UpperCamelCase = NotebookProgressBar(__a , prefix=__a , parent=self , width=__a) return self.child_bar def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = None self.display() class _UpperCAmelCase( lowerCamelCase ): def __init__( self) -> str: '''simple docstring''' _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = False def UpperCAmelCase ( self , __a , __a , __a , **__a) -> int: '''simple docstring''' _UpperCamelCase = '''Epoch''' if args.evaluation_strategy == IntervalStrategy.EPOCH else '''Step''' _UpperCamelCase = 0 _UpperCamelCase = 0 _UpperCamelCase = [self.first_column] + ['''Training Loss'''] if args.evaluation_strategy != IntervalStrategy.NO: column_names.append('''Validation Loss''') _UpperCamelCase = NotebookTrainingTracker(state.max_steps , __a) def UpperCAmelCase ( self , __a , __a , __a , **__a) -> int: '''simple docstring''' _UpperCamelCase = int(state.epoch) if int(state.epoch) == state.epoch else F'''{state.epoch:.2f}''' self.training_tracker.update( state.global_step + 1 , comment=F'''Epoch {epoch}/{state.num_train_epochs}''' , force_update=self._force_next_update , ) _UpperCamelCase = False def UpperCAmelCase ( self , __a , __a , __a , __a=None , **__a) -> List[str]: '''simple docstring''' if not has_length(__a): return if self.prediction_bar is None: if self.training_tracker is not None: _UpperCamelCase = self.training_tracker.add_child(len(__a)) else: _UpperCamelCase = NotebookProgressBar(len(__a)) self.prediction_bar.update(1) else: self.prediction_bar.update(self.prediction_bar.value + 1) def UpperCAmelCase ( self , __a , __a , __a , **__a) -> int: '''simple docstring''' if self.prediction_bar is not None: self.prediction_bar.close() _UpperCamelCase = None def UpperCAmelCase ( self , __a , __a , __a , __a=None , **__a) -> Dict: '''simple docstring''' # Only for when there is no evaluation if args.evaluation_strategy == IntervalStrategy.NO and "loss" in logs: _UpperCamelCase = {'''Training Loss''': logs['''loss''']} # First column is necessarily Step sine we're not in epoch eval strategy _UpperCamelCase = state.global_step self.training_tracker.write_line(__a) def UpperCAmelCase ( self , __a , __a , __a , __a=None , **__a) -> Tuple: '''simple docstring''' if self.training_tracker is not None: _UpperCamelCase = {'''Training Loss''': '''No log''', '''Validation Loss''': '''No log'''} for log in reversed(state.log_history): if "loss" in log: _UpperCamelCase = log['''loss'''] break if self.first_column == "Epoch": _UpperCamelCase = int(state.epoch) else: _UpperCamelCase = state.global_step _UpperCamelCase = '''eval''' for k in metrics: if k.endswith('''_loss'''): _UpperCamelCase = re.sub(R'''\_loss$''' , '''''' , __a) _UpperCamelCase = metrics.pop('''total_flos''' , __a) _UpperCamelCase = metrics.pop('''epoch''' , __a) _UpperCamelCase = metrics.pop(F'''{metric_key_prefix}_runtime''' , __a) _UpperCamelCase = metrics.pop(F'''{metric_key_prefix}_samples_per_second''' , __a) _UpperCamelCase = metrics.pop(F'''{metric_key_prefix}_steps_per_second''' , __a) _UpperCamelCase = metrics.pop(F'''{metric_key_prefix}_jit_compilation_time''' , __a) for k, v in metrics.items(): if k == F'''{metric_key_prefix}_loss''': _UpperCamelCase = v else: _UpperCamelCase = k.split('''_''') _UpperCamelCase = ''' '''.join([part.capitalize() for part in splits[1:]]) _UpperCamelCase = v self.training_tracker.write_line(__a) self.training_tracker.remove_child() _UpperCamelCase = None # Evaluation takes a long time so we should force the next update. _UpperCamelCase = True def UpperCAmelCase ( self , __a , __a , __a , **__a) -> Dict: '''simple docstring''' self.training_tracker.update( state.global_step , comment=F'''Epoch {int(state.epoch)}/{state.num_train_epochs}''' , force_update=__a) _UpperCamelCase = None
78
"""simple docstring""" import copy import re class _UpperCAmelCase: lowercase__ = 'hp' lowercase__ = {} lowercase__ = None @classmethod def UpperCAmelCase ( cls , __a , __a) -> Dict: '''simple docstring''' _UpperCamelCase = prefix _UpperCamelCase = defaults cls.build_naming_info() @staticmethod def UpperCAmelCase ( __a , __a) -> Union[str, Any]: '''simple docstring''' if len(__a) == 0: return "" _UpperCamelCase = None if any(char.isdigit() for char in word): raise Exception(F'''Parameters should not contain numbers: \'{word}\' contains a number''') if word in info["short_word"]: return info["short_word"][word] for prefix_len in range(1 , len(__a) + 1): _UpperCamelCase = word[:prefix_len] if prefix in info["reverse_short_word"]: continue else: _UpperCamelCase = prefix break if short_word is None: # Paranoid fallback def int_to_alphabetic(__a): _UpperCamelCase = '''''' while integer != 0: _UpperCamelCase = chr(ord('''A''') + integer % 10) + s integer //= 10 return s _UpperCamelCase = 0 while True: _UpperCamelCase = word + '''#''' + int_to_alphabetic(__a) if sword in info["reverse_short_word"]: continue else: _UpperCamelCase = sword break _UpperCamelCase = short_word _UpperCamelCase = word return short_word @staticmethod def UpperCAmelCase ( __a , __a) -> Any: '''simple docstring''' _UpperCamelCase = param_name.split('''_''') _UpperCamelCase = [TrialShortNamer.shortname_for_word(__a , __a) for word in words] # We try to create a separatorless short name, but if there is a collision we have to fallback # to a separated short name _UpperCamelCase = ['''''', '''_'''] for separator in separators: _UpperCamelCase = separator.join(__a) if shortname not in info["reverse_short_param"]: _UpperCamelCase = shortname _UpperCamelCase = param_name return shortname return param_name @staticmethod def UpperCAmelCase ( __a , __a) -> List[str]: '''simple docstring''' _UpperCamelCase = TrialShortNamer.shortname_for_key(__a , __a) _UpperCamelCase = short_name _UpperCamelCase = param_name @classmethod def UpperCAmelCase ( cls) -> Any: '''simple docstring''' if cls.NAMING_INFO is not None: return _UpperCamelCase = { '''short_word''': {}, '''reverse_short_word''': {}, '''short_param''': {}, '''reverse_short_param''': {}, } _UpperCamelCase = list(cls.DEFAULTS.keys()) for k in field_keys: cls.add_new_param_name(__a , __a) _UpperCamelCase = info @classmethod def UpperCAmelCase ( cls , __a) -> Optional[Any]: '''simple docstring''' cls.build_naming_info() assert cls.PREFIX is not None _UpperCamelCase = [copy.copy(cls.PREFIX)] for k, v in params.items(): if k not in cls.DEFAULTS: raise Exception(F'''You should provide a default value for the param name {k} with value {v}''') if v == cls.DEFAULTS[k]: # The default value is not added to the name continue _UpperCamelCase = cls.NAMING_INFO['''short_param'''][k] if isinstance(__a , __a): _UpperCamelCase = 1 if v else 0 _UpperCamelCase = '''''' if isinstance(__a , (int, float)) else '''-''' _UpperCamelCase = F'''{key}{sep}{v}''' name.append(__a) return "_".join(__a) @classmethod def UpperCAmelCase ( cls , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = repr[len(cls.PREFIX) + 1 :] if repr == "": _UpperCamelCase = [] else: _UpperCamelCase = repr.split('''_''') _UpperCamelCase = {} for value in values: if "-" in value: _UpperCamelCase , _UpperCamelCase = value.split('''-''') else: _UpperCamelCase = re.sub('''[0-9.]''' , '''''' , __a) _UpperCamelCase = float(re.sub('''[^0-9.]''' , '''''' , __a)) _UpperCamelCase = cls.NAMING_INFO['''reverse_short_param'''][p_k] _UpperCamelCase = p_v for k in cls.DEFAULTS: if k not in parameters: _UpperCamelCase = cls.DEFAULTS[k] return parameters
78
1
"""simple docstring""" from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _a = logging.get_logger(__name__) class _UpperCAmelCase( lowerCamelCase ): lowercase__ = ['pixel_values'] def __init__( self , __a = True , __a = 1 / 2_55 , __a = True , __a = 8 , **__a , ) -> None: '''simple docstring''' super().__init__(**__a) _UpperCamelCase = do_rescale _UpperCamelCase = rescale_factor _UpperCamelCase = do_pad _UpperCamelCase = pad_size def UpperCAmelCase ( self , __a , __a , __a = None , **__a) -> np.ndarray: '''simple docstring''' return rescale(__a , scale=__a , data_format=__a , **__a) def UpperCAmelCase ( self , __a , __a , __a = None) -> List[Any]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = get_image_size(__a) _UpperCamelCase = (old_height // size + 1) * size - old_height _UpperCamelCase = (old_width // size + 1) * size - old_width return pad(__a , ((0, pad_height), (0, pad_width)) , mode='''symmetric''' , data_format=__a) def UpperCAmelCase ( self , __a , __a = None , __a = None , __a = None , __a = None , __a = None , __a = ChannelDimension.FIRST , **__a , ) -> Tuple: '''simple docstring''' _UpperCamelCase = do_rescale if do_rescale is not None else self.do_rescale _UpperCamelCase = rescale_factor if rescale_factor is not None else self.rescale_factor _UpperCamelCase = do_pad if do_pad is not None else self.do_pad _UpperCamelCase = pad_size if pad_size is not None else self.pad_size _UpperCamelCase = make_list_of_images(__a) if not valid_images(__a): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''') if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''') # All transformations expect numpy arrays. _UpperCamelCase = [to_numpy_array(__a) for image in images] if do_rescale: _UpperCamelCase = [self.rescale(image=__a , scale=__a) for image in images] if do_pad: _UpperCamelCase = [self.pad(__a , size=__a) for image in images] _UpperCamelCase = [to_channel_dimension_format(__a , __a) for image in images] _UpperCamelCase = {'''pixel_values''': images} return BatchFeature(data=__a , tensor_type=__a)
78
"""simple docstring""" import os import time import pytest from datasets.utils.filelock import FileLock, Timeout def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = FileLock(str(tmpdir / '''foo.lock''' ) ) _UpperCamelCase = FileLock(str(tmpdir / '''foo.lock''' ) ) _UpperCamelCase = 0.01 with locka.acquire(): with pytest.raises(__snake_case ): _UpperCamelCase = time.time() locka.acquire(__snake_case ) assert time.time() - _start > timeout def lowerCamelCase__ ( __snake_case ) -> List[Any]: """simple docstring""" _UpperCamelCase = '''a''' * 10_00 + '''.lock''' _UpperCamelCase = FileLock(str(tmpdir / filename ) ) assert locka._lock_file.endswith('''.lock''' ) assert not locka._lock_file.endswith(__snake_case ) assert len(os.path.basename(locka._lock_file ) ) <= 2_55 _UpperCamelCase = FileLock(tmpdir / filename ) with locka.acquire(): with pytest.raises(__snake_case ): locka.acquire(0 )
78
1
"""simple docstring""" import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, UNetaDConditionModel, VideoToVideoSDPipeline, ) from diffusers.utils import floats_tensor, is_xformers_available, skip_mps from diffusers.utils.testing_utils import enable_full_determinism, slow, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class _UpperCAmelCase( lowerCamelCase , unittest.TestCase ): lowercase__ = VideoToVideoSDPipeline lowercase__ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS.union({'video'} ) - {'image', 'width', 'height'} lowercase__ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'video'} ) - {'image'} lowercase__ = PipelineTesterMixin.required_optional_params - {'latents'} lowercase__ = False # No `output_type`. lowercase__ = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'return_dict', 'callback', 'callback_steps', ] ) def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' torch.manual_seed(0) _UpperCamelCase = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''DownBlock3D''') , up_block_types=('''UpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''') , cross_attention_dim=32 , attention_head_dim=4 , ) _UpperCamelCase = DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule='''scaled_linear''' , clip_sample=__a , set_alpha_to_one=__a , ) torch.manual_seed(0) _UpperCamelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0) _UpperCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act='''gelu''' , projection_dim=5_12 , ) _UpperCamelCase = CLIPTextModel(__a) _UpperCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') _UpperCamelCase = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, } return components def UpperCAmelCase ( self , __a , __a=0) -> Optional[int]: '''simple docstring''' # 3 frames _UpperCamelCase = floats_tensor((1, 3, 3, 32, 32) , rng=random.Random(__a)).to(__a) if str(__a).startswith('''mps'''): _UpperCamelCase = torch.manual_seed(__a) else: _UpperCamelCase = torch.Generator(device=__a).manual_seed(__a) _UpperCamelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''video''': video, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''pt''', } return inputs def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator _UpperCamelCase = self.get_dummy_components() _UpperCamelCase = VideoToVideoSDPipeline(**__a) _UpperCamelCase = sd_pipe.to(__a) sd_pipe.set_progress_bar_config(disable=__a) _UpperCamelCase = self.get_dummy_inputs(__a) _UpperCamelCase = '''np''' _UpperCamelCase = sd_pipe(**__a).frames _UpperCamelCase = frames[0][-3:, -3:, -1] assert frames[0].shape == (32, 32, 3) _UpperCamelCase = np.array([1_06, 1_17, 1_13, 1_74, 1_37, 1_12, 1_48, 1_51, 1_31]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=__a , expected_max_diff=5e-3) @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''') def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' pass @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''') def UpperCAmelCase ( self) -> int: '''simple docstring''' pass @unittest.skip(reason='''`num_images_per_prompt` argument is not supported for this pipeline.''') def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' pass def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' return super().test_progress_bar() @slow @skip_mps class _UpperCAmelCase( unittest.TestCase ): def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = VideoToVideoSDPipeline.from_pretrained('''cerspense/zeroscope_v2_XL''' , torch_dtype=torch.floataa) pipe.enable_model_cpu_offload() # 10 frames _UpperCamelCase = torch.Generator(device='''cpu''').manual_seed(0) _UpperCamelCase = torch.randn((1, 10, 3, 10_24, 5_76) , generator=__a) _UpperCamelCase = video.to('''cuda''') _UpperCamelCase = '''Spiderman is surfing''' _UpperCamelCase = pipe(__a , video=__a , generator=__a , num_inference_steps=3 , output_type='''pt''').frames _UpperCamelCase = np.array([-1.045_8984, -1.127_9297, -0.966_3086, -0.9150_3906, -0.7509_7656]) assert np.abs(video_frames.cpu().numpy()[0, 0, 0, 0, -5:] - expected_array).sum() < 1e-2
78
"""simple docstring""" from math import sqrt def lowerCamelCase__ ( __snake_case ) -> bool: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and ( number >= 0 ), "'number' must been an int and positive" _UpperCamelCase = True # 0 and 1 are none primes. if number <= 1: _UpperCamelCase = False for divisor in range(2, int(round(sqrt(__snake_case ) ) ) + 1 ): # if 'number' divisible by 'divisor' then sets 'status' # of false and break up the loop. if number % divisor == 0: _UpperCamelCase = False break # precondition assert isinstance(__snake_case, __snake_case ), "'status' must been from type bool" return status def lowerCamelCase__ ( __snake_case ) -> Dict: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n > 2), "'N' must been an int and > 2" # beginList: contains all natural numbers from 2 up to N _UpperCamelCase = list(range(2, n + 1 ) ) _UpperCamelCase = [] # this list will be returns. # actual sieve of erathostenes for i in range(len(__snake_case ) ): for j in range(i + 1, len(__snake_case ) ): if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0): _UpperCamelCase = 0 # filters actual prime numbers. _UpperCamelCase = [x for x in begin_list if x != 0] # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type list" return ans def lowerCamelCase__ ( __snake_case ) -> List[str]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n > 2), "'N' must been an int and > 2" _UpperCamelCase = [] # iterates over all numbers between 2 up to N+1 # if a number is prime then appends to list 'ans' for number in range(2, n + 1 ): if is_prime(__snake_case ): ans.append(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type list" return ans def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and number >= 0, "'number' must been an int and >= 0" _UpperCamelCase = [] # this list will be returns of the function. # potential prime number factors. _UpperCamelCase = 2 _UpperCamelCase = number if number == 0 or number == 1: ans.append(__snake_case ) # if 'number' not prime then builds the prime factorization of 'number' elif not is_prime(__snake_case ): while quotient != 1: if is_prime(__snake_case ) and (quotient % factor == 0): ans.append(__snake_case ) quotient /= factor else: factor += 1 else: ans.append(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type list" return ans def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and ( number >= 0 ), "'number' bust been an int and >= 0" _UpperCamelCase = 0 # prime factorization of 'number' _UpperCamelCase = prime_factorization(__snake_case ) _UpperCamelCase = max(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type int" return ans def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and ( number >= 0 ), "'number' bust been an int and >= 0" _UpperCamelCase = 0 # prime factorization of 'number' _UpperCamelCase = prime_factorization(__snake_case ) _UpperCamelCase = min(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type int" return ans def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ), "'number' must been an int" assert isinstance(number % 2 == 0, __snake_case ), "compare bust been from type bool" return number % 2 == 0 def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ), "'number' must been an int" assert isinstance(number % 2 != 0, __snake_case ), "compare bust been from type bool" return number % 2 != 0 def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" assert ( isinstance(__snake_case, __snake_case ) and (number > 2) and is_even(__snake_case ) ), "'number' must been an int, even and > 2" _UpperCamelCase = [] # this list will returned # creates a list of prime numbers between 2 up to 'number' _UpperCamelCase = get_prime_numbers(__snake_case ) _UpperCamelCase = len(__snake_case ) # run variable for while-loops. _UpperCamelCase = 0 _UpperCamelCase = None # exit variable. for break up the loops _UpperCamelCase = True while i < len_pn and loop: _UpperCamelCase = i + 1 while j < len_pn and loop: if prime_numbers[i] + prime_numbers[j] == number: _UpperCamelCase = False ans.append(prime_numbers[i] ) ans.append(prime_numbers[j] ) j += 1 i += 1 # precondition assert ( isinstance(__snake_case, __snake_case ) and (len(__snake_case ) == 2) and (ans[0] + ans[1] == number) and is_prime(ans[0] ) and is_prime(ans[1] ) ), "'ans' must contains two primes. And sum of elements must been eq 'number'" return ans def lowerCamelCase__ ( __snake_case, __snake_case ) -> str: """simple docstring""" assert ( isinstance(__snake_case, __snake_case ) and isinstance(__snake_case, __snake_case ) and (numbera >= 0) and (numbera >= 0) ), "'number1' and 'number2' must been positive integer." _UpperCamelCase = 0 while numbera != 0: _UpperCamelCase = numbera % numbera _UpperCamelCase = numbera _UpperCamelCase = rest # precondition assert isinstance(__snake_case, __snake_case ) and ( numbera >= 0 ), "'number' must been from type int and positive" return numbera def lowerCamelCase__ ( __snake_case, __snake_case ) -> List[str]: """simple docstring""" assert ( isinstance(__snake_case, __snake_case ) and isinstance(__snake_case, __snake_case ) and (numbera >= 1) and (numbera >= 1) ), "'number1' and 'number2' must been positive integer." _UpperCamelCase = 1 # actual answer that will be return. # for kgV (x,1) if numbera > 1 and numbera > 1: # builds the prime factorization of 'number1' and 'number2' _UpperCamelCase = prime_factorization(__snake_case ) _UpperCamelCase = prime_factorization(__snake_case ) elif numbera == 1 or numbera == 1: _UpperCamelCase = [] _UpperCamelCase = [] _UpperCamelCase = max(__snake_case, __snake_case ) _UpperCamelCase = 0 _UpperCamelCase = 0 _UpperCamelCase = [] # captured numbers int both 'primeFac1' and 'primeFac2' # iterates through primeFac1 for n in prime_fac_a: if n not in done: if n in prime_fac_a: _UpperCamelCase = prime_fac_a.count(__snake_case ) _UpperCamelCase = prime_fac_a.count(__snake_case ) for _ in range(max(__snake_case, __snake_case ) ): ans *= n else: _UpperCamelCase = prime_fac_a.count(__snake_case ) for _ in range(__snake_case ): ans *= n done.append(__snake_case ) # iterates through primeFac2 for n in prime_fac_a: if n not in done: _UpperCamelCase = prime_fac_a.count(__snake_case ) for _ in range(__snake_case ): ans *= n done.append(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ) and ( ans >= 0 ), "'ans' must been from type int and positive" return ans def lowerCamelCase__ ( __snake_case ) -> Tuple: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n >= 0), "'number' must been a positive int" _UpperCamelCase = 0 _UpperCamelCase = 2 # this variable holds the answer while index < n: index += 1 ans += 1 # counts to the next number # if ans not prime then # runs to the next prime number. while not is_prime(__snake_case ): ans += 1 # precondition assert isinstance(__snake_case, __snake_case ) and is_prime( __snake_case ), "'ans' must been a prime number and from type int" return ans def lowerCamelCase__ ( __snake_case, __snake_case ) -> Tuple: """simple docstring""" assert ( is_prime(__snake_case ) and is_prime(__snake_case ) and (p_number_a < p_number_a) ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'" _UpperCamelCase = p_number_a + 1 # jump to the next number _UpperCamelCase = [] # this list will be returns. # if number is not prime then # fetch the next prime number. while not is_prime(__snake_case ): number += 1 while number < p_number_a: ans.append(__snake_case ) number += 1 # fetch the next prime number. while not is_prime(__snake_case ): number += 1 # precondition assert ( isinstance(__snake_case, __snake_case ) and ans[0] != p_number_a and ans[len(__snake_case ) - 1] != p_number_a ), "'ans' must been a list without the arguments" # 'ans' contains not 'pNumber1' and 'pNumber2' ! return ans def lowerCamelCase__ ( __snake_case ) -> List[Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n >= 1), "'n' must been int and >= 1" _UpperCamelCase = [] # will be returned. for divisor in range(1, n + 1 ): if n % divisor == 0: ans.append(__snake_case ) # precondition assert ans[0] == 1 and ans[len(__snake_case ) - 1] == n, "Error in function getDivisiors(...)" return ans def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and ( number > 1 ), "'number' must been an int and >= 1" _UpperCamelCase = get_divisors(__snake_case ) # precondition assert ( isinstance(__snake_case, __snake_case ) and (divisors[0] == 1) and (divisors[len(__snake_case ) - 1] == number) ), "Error in help-function getDivisiors(...)" # summed all divisors up to 'number' (exclusive), hence [:-1] return sum(divisors[:-1] ) == number def lowerCamelCase__ ( __snake_case, __snake_case ) -> Optional[Any]: """simple docstring""" assert ( isinstance(__snake_case, __snake_case ) and isinstance(__snake_case, __snake_case ) and (denominator != 0) ), "The arguments must been from type int and 'denominator' != 0" # build the greatest common divisor of numerator and denominator. _UpperCamelCase = gcd(abs(__snake_case ), abs(__snake_case ) ) # precondition assert ( isinstance(__snake_case, __snake_case ) and (numerator % gcd_of_fraction == 0) and (denominator % gcd_of_fraction == 0) ), "Error in function gcd(...,...)" return (numerator // gcd_of_fraction, denominator // gcd_of_fraction) def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n >= 0), "'n' must been a int and >= 0" _UpperCamelCase = 1 # this will be return. for factor in range(1, n + 1 ): ans *= factor return ans def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n >= 0), "'n' must been an int and >= 0" _UpperCamelCase = 0 _UpperCamelCase = 1 _UpperCamelCase = 1 # this will be return for _ in range(n - 1 ): _UpperCamelCase = ans ans += fiba _UpperCamelCase = tmp return ans
78
1
"""simple docstring""" def lowerCamelCase__ ( __snake_case, __snake_case ) -> Dict: """simple docstring""" if b == 0: return 1 if (b % 2) == 0: return actual_power(__snake_case, int(b / 2 ) ) * actual_power(__snake_case, int(b / 2 ) ) else: return a * actual_power(__snake_case, int(b / 2 ) ) * actual_power(__snake_case, int(b / 2 ) ) def lowerCamelCase__ ( __snake_case, __snake_case ) -> float: """simple docstring""" if b < 0: return 1 / actual_power(__snake_case, __snake_case ) return actual_power(__snake_case, __snake_case ) if __name__ == "__main__": print(power(-2, -3))
78
"""simple docstring""" import logging from dataclasses import dataclass, field from pathlib import Path from typing import Optional, Union from .generation.configuration_utils import GenerationConfig from .training_args import TrainingArguments from .utils import add_start_docstrings _a = logging.getLogger(__name__) @dataclass @add_start_docstrings(TrainingArguments.__doc__ ) class _UpperCAmelCase( lowerCamelCase ): lowercase__ = field(default=lowerCamelCase , metadata={'help': 'Whether to use SortishSampler or not.'} ) lowercase__ = field( default=lowerCamelCase , metadata={'help': 'Whether to use generate to calculate generative metrics (ROUGE, BLEU).'} ) lowercase__ = field( default=lowerCamelCase , metadata={ 'help': ( 'The `max_length` to use on each evaluation loop when `predict_with_generate=True`. Will default ' 'to the `max_length` value of the model configuration.' ) } , ) lowercase__ = field( default=lowerCamelCase , metadata={ 'help': ( 'The `num_beams` to use on each evaluation loop when `predict_with_generate=True`. Will default ' 'to the `num_beams` value of the model configuration.' ) } , ) lowercase__ = field( default=lowerCamelCase , metadata={ 'help': 'Model id, file path or url pointing to a GenerationConfig json file, to use during prediction.' } , ) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = super().to_dict() for k, v in d.items(): if isinstance(__a , __a): _UpperCamelCase = v.to_dict() return d
78
1
"""simple docstring""" from __future__ import annotations import unittest from transformers import BlenderbotConfig, BlenderbotTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotForConditionalGeneration, TFBlenderbotModel @require_tf class _UpperCAmelCase: lowercase__ = BlenderbotConfig lowercase__ = {} lowercase__ = 'gelu' def __init__( self , __a , __a=13 , __a=7 , __a=True , __a=False , __a=99 , __a=32 , __a=2 , __a=4 , __a=37 , __a=0.1 , __a=0.1 , __a=20 , __a=2 , __a=1 , __a=0 , ) -> Dict: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = seq_length _UpperCamelCase = is_training _UpperCamelCase = use_labels _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = eos_token_id _UpperCamelCase = pad_token_id _UpperCamelCase = bos_token_id def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size) _UpperCamelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size) , 1) _UpperCamelCase = tf.concat([input_ids, eos_tensor] , axis=1) _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) _UpperCamelCase = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) _UpperCamelCase = prepare_blenderbot_inputs_dict(__a , __a , __a) return config, inputs_dict def UpperCAmelCase ( self , __a , __a) -> Dict: '''simple docstring''' _UpperCamelCase = TFBlenderbotModel(config=__a).get_decoder() _UpperCamelCase = inputs_dict['''input_ids'''] _UpperCamelCase = input_ids[:1, :] _UpperCamelCase = inputs_dict['''attention_mask'''][:1, :] _UpperCamelCase = inputs_dict['''head_mask'''] _UpperCamelCase = 1 # first forward pass _UpperCamelCase = model(__a , attention_mask=__a , head_mask=__a , use_cache=__a) _UpperCamelCase , _UpperCamelCase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids _UpperCamelCase = ids_tensor((self.batch_size, 3) , config.vocab_size) _UpperCamelCase = tf.cast(ids_tensor((self.batch_size, 3) , 2) , tf.inta) # append to next input_ids and _UpperCamelCase = tf.concat([input_ids, next_tokens] , axis=-1) _UpperCamelCase = tf.concat([attention_mask, next_attn_mask] , axis=-1) _UpperCamelCase = model(__a , attention_mask=__a)[0] _UpperCamelCase = model(__a , attention_mask=__a , past_key_values=__a)[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1]) # select random slice _UpperCamelCase = int(ids_tensor((1,) , output_from_past.shape[-1])) _UpperCamelCase = output_from_no_past[:, -3:, random_slice_idx] _UpperCamelCase = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__a , __a , rtol=1e-3) def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case, __snake_case=None, __snake_case=None, __snake_case=None, __snake_case=None, __snake_case=None, ) -> Union[str, Any]: """simple docstring""" if attention_mask is None: _UpperCamelCase = tf.cast(tf.math.not_equal(__snake_case, config.pad_token_id ), tf.inta ) if decoder_attention_mask is None: _UpperCamelCase = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape, dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:], config.pad_token_id ), tf.inta ), ], axis=-1, ) if head_mask is None: _UpperCamelCase = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _UpperCamelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: _UpperCamelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class _UpperCAmelCase( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase__ = (TFBlenderbotForConditionalGeneration, TFBlenderbotModel) if is_tf_available() else () lowercase__ = (TFBlenderbotForConditionalGeneration,) if is_tf_available() else () lowercase__ = ( { 'conversational': TFBlenderbotForConditionalGeneration, 'feature-extraction': TFBlenderbotModel, 'summarization': TFBlenderbotForConditionalGeneration, 'text2text-generation': TFBlenderbotForConditionalGeneration, 'translation': TFBlenderbotForConditionalGeneration, } if is_tf_available() else {} ) lowercase__ = True lowercase__ = False lowercase__ = False def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = TFBlenderbotModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=__a) def UpperCAmelCase ( self) -> int: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__a) @require_tokenizers @require_tf class _UpperCAmelCase( unittest.TestCase ): lowercase__ = ['My friends are cool but they eat too many carbs.'] lowercase__ = 'facebook/blenderbot-400M-distill' @cached_property def UpperCAmelCase ( self) -> Dict: '''simple docstring''' return BlenderbotTokenizer.from_pretrained(self.model_name) @cached_property def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name) return model @slow def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.tokenizer(self.src_text , return_tensors='''tf''') _UpperCamelCase = self.model.generate( model_inputs.input_ids , ) _UpperCamelCase = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=__a)[0] assert ( generated_words == " That's unfortunate. Are they trying to lose weight or are they just trying to be healthier?" )
78
"""simple docstring""" import argparse import torch from transformers import BlenderbotConfig, BlenderbotForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) _a = [ ["""attention""", """attn"""], ["""encoder_attention""", """encoder_attn"""], ["""q_lin""", """q_proj"""], ["""k_lin""", """k_proj"""], ["""v_lin""", """v_proj"""], ["""out_lin""", """out_proj"""], ["""norm_embeddings""", """layernorm_embedding"""], ["""position_embeddings""", """embed_positions"""], ["""embeddings""", """embed_tokens"""], ["""ffn.lin""", """fc"""], ] def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" if k == "embeddings.weight": return "shared.weight" for parlai_name, hf_name in PATTERNS: _UpperCamelCase = k.replace(__snake_case, __snake_case ) if k.startswith('''encoder''' ): _UpperCamelCase = k.replace('''.attn''', '''.self_attn''' ) _UpperCamelCase = k.replace('''norm1''', '''self_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm2''', '''final_layer_norm''' ) elif k.startswith('''decoder''' ): _UpperCamelCase = k.replace('''norm1''', '''self_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm2''', '''encoder_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm3''', '''final_layer_norm''' ) return k def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" _UpperCamelCase = [ '''model.encoder.layernorm_embedding.weight''', '''model.encoder.layernorm_embedding.bias''', '''model.decoder.layernorm_embedding.weight''', '''model.decoder.layernorm_embedding.bias''', ] for k in keys: _UpperCamelCase = sd.pop(__snake_case ) _UpperCamelCase = k.replace('''layernorm_embedding''', '''layer_norm''' ) assert new_k not in sd _UpperCamelCase = v _a = ["""START"""] @torch.no_grad() def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> int: """simple docstring""" _UpperCamelCase = torch.load(__snake_case, map_location='''cpu''' ) _UpperCamelCase = model['''model'''] _UpperCamelCase = BlenderbotConfig.from_json_file(__snake_case ) _UpperCamelCase = BlenderbotForConditionalGeneration(__snake_case ) _UpperCamelCase = m.model.state_dict().keys() _UpperCamelCase = [] _UpperCamelCase = {} for k, v in sd.items(): if k in IGNORE_KEYS: continue _UpperCamelCase = rename_state_dict_key(__snake_case ) if new_k not in valid_keys: failures.append([k, new_k] ) else: _UpperCamelCase = v if cfg.normalize_before: # Blenderbot-3B checkpoints. Rename layernorm_embedding -> layer_norm rename_layernorm_keys(__snake_case ) m.model.load_state_dict(__snake_case, strict=__snake_case ) m.half() m.save_pretrained(__snake_case ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument("""--src_path""", type=str, help="""like blenderbot-model.bin""") parser.add_argument("""--save_dir""", default="""hf_blenderbot""", type=str, help="""Where to save converted model.""") parser.add_argument( """--hf_config_json""", default="""blenderbot-3b-config.json""", type=str, help="""Path to config to use""" ) _a = parser.parse_args() convert_parlai_checkpoint(args.src_path, args.save_dir, args.hf_config_json)
78
1
"""simple docstring""" from collections import deque from .hash_table import HashTable class _UpperCAmelCase( lowerCamelCase ): def __init__( self , *__a , **__a) -> Dict: '''simple docstring''' super().__init__(*__a , **__a) def UpperCAmelCase ( self , __a , __a) -> str: '''simple docstring''' _UpperCamelCase = deque([]) if self.values[key] is None else self.values[key] self.values[key].appendleft(__a) _UpperCamelCase = self.values[key] def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' return ( sum(self.charge_factor - len(__a) for slot in self.values) / self.size_table * self.charge_factor ) def UpperCAmelCase ( self , __a , __a=None) -> List[str]: '''simple docstring''' if not ( len(self.values[key]) == self.charge_factor and self.values.count(__a) == 0 ): return key return super()._collision_resolution(__a , __a)
78
"""simple docstring""" import argparse import os.path as osp import re import torch from safetensors.torch import load_file, save_file # =================# # UNet Conversion # # =================# _a = [ # (stable-diffusion, HF Diffusers) ("""time_embed.0.weight""", """time_embedding.linear_1.weight"""), ("""time_embed.0.bias""", """time_embedding.linear_1.bias"""), ("""time_embed.2.weight""", """time_embedding.linear_2.weight"""), ("""time_embed.2.bias""", """time_embedding.linear_2.bias"""), ("""input_blocks.0.0.weight""", """conv_in.weight"""), ("""input_blocks.0.0.bias""", """conv_in.bias"""), ("""out.0.weight""", """conv_norm_out.weight"""), ("""out.0.bias""", """conv_norm_out.bias"""), ("""out.2.weight""", """conv_out.weight"""), ("""out.2.bias""", """conv_out.bias"""), ] _a = [ # (stable-diffusion, HF Diffusers) ("""in_layers.0""", """norm1"""), ("""in_layers.2""", """conv1"""), ("""out_layers.0""", """norm2"""), ("""out_layers.3""", """conv2"""), ("""emb_layers.1""", """time_emb_proj"""), ("""skip_connection""", """conv_shortcut"""), ] _a = [] # hardcoded number of downblocks and resnets/attentions... # would need smarter logic for other networks. for i in range(4): # loop over downblocks/upblocks for j in range(2): # loop over resnets/attentions for downblocks _a = F"""down_blocks.{i}.resnets.{j}.""" _a = F"""input_blocks.{3*i + j + 1}.0.""" unet_conversion_map_layer.append((sd_down_res_prefix, hf_down_res_prefix)) if i < 3: # no attention layers in down_blocks.3 _a = F"""down_blocks.{i}.attentions.{j}.""" _a = F"""input_blocks.{3*i + j + 1}.1.""" unet_conversion_map_layer.append((sd_down_atn_prefix, hf_down_atn_prefix)) for j in range(3): # loop over resnets/attentions for upblocks _a = F"""up_blocks.{i}.resnets.{j}.""" _a = F"""output_blocks.{3*i + j}.0.""" unet_conversion_map_layer.append((sd_up_res_prefix, hf_up_res_prefix)) if i > 0: # no attention layers in up_blocks.0 _a = F"""up_blocks.{i}.attentions.{j}.""" _a = F"""output_blocks.{3*i + j}.1.""" unet_conversion_map_layer.append((sd_up_atn_prefix, hf_up_atn_prefix)) if i < 3: # no downsample in down_blocks.3 _a = F"""down_blocks.{i}.downsamplers.0.conv.""" _a = F"""input_blocks.{3*(i+1)}.0.op.""" unet_conversion_map_layer.append((sd_downsample_prefix, hf_downsample_prefix)) # no upsample in up_blocks.3 _a = F"""up_blocks.{i}.upsamplers.0.""" _a = F"""output_blocks.{3*i + 2}.{1 if i == 0 else 2}.""" unet_conversion_map_layer.append((sd_upsample_prefix, hf_upsample_prefix)) _a = """mid_block.attentions.0.""" _a = """middle_block.1.""" unet_conversion_map_layer.append((sd_mid_atn_prefix, hf_mid_atn_prefix)) for j in range(2): _a = F"""mid_block.resnets.{j}.""" _a = F"""middle_block.{2*j}.""" unet_conversion_map_layer.append((sd_mid_res_prefix, hf_mid_res_prefix)) def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" _UpperCamelCase = {k: k for k in unet_state_dict.keys()} for sd_name, hf_name in unet_conversion_map: _UpperCamelCase = sd_name for k, v in mapping.items(): if "resnets" in k: for sd_part, hf_part in unet_conversion_map_resnet: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v for k, v in mapping.items(): for sd_part, hf_part in unet_conversion_map_layer: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v _UpperCamelCase = {v: unet_state_dict[k] for k, v in mapping.items()} return new_state_dict # ================# # VAE Conversion # # ================# _a = [ # (stable-diffusion, HF Diffusers) ("""nin_shortcut""", """conv_shortcut"""), ("""norm_out""", """conv_norm_out"""), ("""mid.attn_1.""", """mid_block.attentions.0."""), ] for i in range(4): # down_blocks have two resnets for j in range(2): _a = F"""encoder.down_blocks.{i}.resnets.{j}.""" _a = F"""encoder.down.{i}.block.{j}.""" vae_conversion_map.append((sd_down_prefix, hf_down_prefix)) if i < 3: _a = F"""down_blocks.{i}.downsamplers.0.""" _a = F"""down.{i}.downsample.""" vae_conversion_map.append((sd_downsample_prefix, hf_downsample_prefix)) _a = F"""up_blocks.{i}.upsamplers.0.""" _a = F"""up.{3-i}.upsample.""" vae_conversion_map.append((sd_upsample_prefix, hf_upsample_prefix)) # up_blocks have three resnets # also, up blocks in hf are numbered in reverse from sd for j in range(3): _a = F"""decoder.up_blocks.{i}.resnets.{j}.""" _a = F"""decoder.up.{3-i}.block.{j}.""" vae_conversion_map.append((sd_up_prefix, hf_up_prefix)) # this part accounts for mid blocks in both the encoder and the decoder for i in range(2): _a = F"""mid_block.resnets.{i}.""" _a = F"""mid.block_{i+1}.""" vae_conversion_map.append((sd_mid_res_prefix, hf_mid_res_prefix)) _a = [ # (stable-diffusion, HF Diffusers) ("""norm.""", """group_norm."""), ("""q.""", """query."""), ("""k.""", """key."""), ("""v.""", """value."""), ("""proj_out.""", """proj_attn."""), ] def lowerCamelCase__ ( __snake_case ) -> List[str]: """simple docstring""" return w.reshape(*w.shape, 1, 1 ) def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" _UpperCamelCase = {k: k for k in vae_state_dict.keys()} for k, v in mapping.items(): for sd_part, hf_part in vae_conversion_map: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v for k, v in mapping.items(): if "attentions" in k: for sd_part, hf_part in vae_conversion_map_attn: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v _UpperCamelCase = {v: vae_state_dict[k] for k, v in mapping.items()} _UpperCamelCase = ['''q''', '''k''', '''v''', '''proj_out'''] for k, v in new_state_dict.items(): for weight_name in weights_to_convert: if F'''mid.attn_1.{weight_name}.weight''' in k: print(F'''Reshaping {k} for SD format''' ) _UpperCamelCase = reshape_weight_for_sd(__snake_case ) return new_state_dict # =========================# # Text Encoder Conversion # # =========================# _a = [ # (stable-diffusion, HF Diffusers) ("""resblocks.""", """text_model.encoder.layers."""), ("""ln_1""", """layer_norm1"""), ("""ln_2""", """layer_norm2"""), (""".c_fc.""", """.fc1."""), (""".c_proj.""", """.fc2."""), (""".attn""", """.self_attn"""), ("""ln_final.""", """transformer.text_model.final_layer_norm."""), ("""token_embedding.weight""", """transformer.text_model.embeddings.token_embedding.weight"""), ("""positional_embedding""", """transformer.text_model.embeddings.position_embedding.weight"""), ] _a = {re.escape(x[1]): x[0] for x in textenc_conversion_lst} _a = re.compile("""|""".join(protected.keys())) # Ordering is from https://github.com/pytorch/pytorch/blob/master/test/cpp/api/modules.cpp _a = {"""q""": 0, """k""": 1, """v""": 2} def lowerCamelCase__ ( __snake_case ) -> Any: """simple docstring""" _UpperCamelCase = {} _UpperCamelCase = {} _UpperCamelCase = {} for k, v in text_enc_dict.items(): if ( k.endswith('''.self_attn.q_proj.weight''' ) or k.endswith('''.self_attn.k_proj.weight''' ) or k.endswith('''.self_attn.v_proj.weight''' ) ): _UpperCamelCase = k[: -len('''.q_proj.weight''' )] _UpperCamelCase = k[-len('''q_proj.weight''' )] if k_pre not in capture_qkv_weight: _UpperCamelCase = [None, None, None] _UpperCamelCase = v continue if ( k.endswith('''.self_attn.q_proj.bias''' ) or k.endswith('''.self_attn.k_proj.bias''' ) or k.endswith('''.self_attn.v_proj.bias''' ) ): _UpperCamelCase = k[: -len('''.q_proj.bias''' )] _UpperCamelCase = k[-len('''q_proj.bias''' )] if k_pre not in capture_qkv_bias: _UpperCamelCase = [None, None, None] _UpperCamelCase = v continue _UpperCamelCase = textenc_pattern.sub(lambda __snake_case : protected[re.escape(m.group(0 ) )], __snake_case ) _UpperCamelCase = v for k_pre, tensors in capture_qkv_weight.items(): if None in tensors: raise Exception('''CORRUPTED MODEL: one of the q-k-v values for the text encoder was missing''' ) _UpperCamelCase = textenc_pattern.sub(lambda __snake_case : protected[re.escape(m.group(0 ) )], __snake_case ) _UpperCamelCase = torch.cat(__snake_case ) for k_pre, tensors in capture_qkv_bias.items(): if None in tensors: raise Exception('''CORRUPTED MODEL: one of the q-k-v values for the text encoder was missing''' ) _UpperCamelCase = textenc_pattern.sub(lambda __snake_case : protected[re.escape(m.group(0 ) )], __snake_case ) _UpperCamelCase = torch.cat(__snake_case ) return new_state_dict def lowerCamelCase__ ( __snake_case ) -> Tuple: """simple docstring""" return text_enc_dict if __name__ == "__main__": _a = argparse.ArgumentParser() parser.add_argument("""--model_path""", default=None, type=str, required=True, help="""Path to the model to convert.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, required=True, help="""Path to the output model.""") parser.add_argument("""--half""", action="""store_true""", help="""Save weights in half precision.""") parser.add_argument( """--use_safetensors""", action="""store_true""", help="""Save weights use safetensors, default is ckpt.""" ) _a = parser.parse_args() assert args.model_path is not None, "Must provide a model path!" assert args.checkpoint_path is not None, "Must provide a checkpoint path!" # Path for safetensors _a = osp.join(args.model_path, """unet""", """diffusion_pytorch_model.safetensors""") _a = osp.join(args.model_path, """vae""", """diffusion_pytorch_model.safetensors""") _a = osp.join(args.model_path, """text_encoder""", """model.safetensors""") # Load models from safetensors if it exists, if it doesn't pytorch if osp.exists(unet_path): _a = load_file(unet_path, device="""cpu""") else: _a = osp.join(args.model_path, """unet""", """diffusion_pytorch_model.bin""") _a = torch.load(unet_path, map_location="""cpu""") if osp.exists(vae_path): _a = load_file(vae_path, device="""cpu""") else: _a = osp.join(args.model_path, """vae""", """diffusion_pytorch_model.bin""") _a = torch.load(vae_path, map_location="""cpu""") if osp.exists(text_enc_path): _a = load_file(text_enc_path, device="""cpu""") else: _a = osp.join(args.model_path, """text_encoder""", """pytorch_model.bin""") _a = torch.load(text_enc_path, map_location="""cpu""") # Convert the UNet model _a = convert_unet_state_dict(unet_state_dict) _a = {"""model.diffusion_model.""" + k: v for k, v in unet_state_dict.items()} # Convert the VAE model _a = convert_vae_state_dict(vae_state_dict) _a = {"""first_stage_model.""" + k: v for k, v in vae_state_dict.items()} # Easiest way to identify v2.0 model seems to be that the text encoder (OpenCLIP) is deeper _a = """text_model.encoder.layers.22.layer_norm2.bias""" in text_enc_dict if is_vaa_model: # Need to add the tag 'transformer' in advance so we can knock it out from the final layer-norm _a = {"""transformer.""" + k: v for k, v in text_enc_dict.items()} _a = convert_text_enc_state_dict_vaa(text_enc_dict) _a = {"""cond_stage_model.model.""" + k: v for k, v in text_enc_dict.items()} else: _a = convert_text_enc_state_dict(text_enc_dict) _a = {"""cond_stage_model.transformer.""" + k: v for k, v in text_enc_dict.items()} # Put together new checkpoint _a = {**unet_state_dict, **vae_state_dict, **text_enc_dict} if args.half: _a = {k: v.half() for k, v in state_dict.items()} if args.use_safetensors: save_file(state_dict, args.checkpoint_path) else: _a = {"""state_dict""": state_dict} torch.save(state_dict, args.checkpoint_path)
78
1
"""simple docstring""" import argparse import json import os from collections import OrderedDict import numpy as np import tensorflow as tf import torch def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" _UpperCamelCase = os.path.join(args.tf_model_dir, '''parameters.json''' ) _UpperCamelCase = json.loads(open(__snake_case ).read() ) if not params: raise ValueError( F'''It seems that the json file at {parameter_file} is empty. Make sure you have a correct json file.''' ) if not args.output.endswith('''.pt''' ): _UpperCamelCase = args.output + '''.pt''' _UpperCamelCase = OrderedDict() with tf.device('''/CPU:0''' ): _UpperCamelCase = tf.train.load_checkpoint(args.tf_model_dir ) _UpperCamelCase = reader.get_variable_to_shape_map() for key_name in shapes.keys(): _UpperCamelCase = reader.get_tensor(__snake_case ).astype(np.floataa ) if key_name.endswith('''/adam_m''' ) or key_name.endswith('''/adam_v''' ): continue if key_name.startswith('''pasts/''' ): if key_name.startswith('''pasts/mlp''' ): _UpperCamelCase = int(key_name[9] ) elif key_name.startswith('''pasts/out''' ): _UpperCamelCase = 8 _UpperCamelCase = '''model.sqout.%d.weight''' % (player * 2) # enter to nn.Sequencial with Tanh, so 2 at a time _UpperCamelCase = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix _UpperCamelCase = torch.tensor(__snake_case ) elif key_name.startswith('''model/moe''' ): _UpperCamelCase = int(key_name[9:].split('''/''' )[0] ) if key_name.endswith('''/switch_gating/kernel''' ): _UpperCamelCase = '''model.blocks.%d.feed_forward.mlp.router.classifier.weight''' % player _UpperCamelCase = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix _UpperCamelCase = torch.tensor(__snake_case ) elif key_name.endswith('''/softmlp/kernel''' ): _UpperCamelCase = '''model.blocks.%d.feed_forward.soft_bypass_mlp.weight''' % player _UpperCamelCase = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix _UpperCamelCase = torch.tensor(__snake_case ) elif key_name.endswith('''/wo/kernel''' ) or key_name.endswith('''/wi/kernel''' ): _UpperCamelCase = key_name[-9:-7] for i in range(16 ): _UpperCamelCase = '''model.blocks.%d.feed_forward.mlp.experts.expert_%d.%s.weight''' % (player, i, nlayer) _UpperCamelCase = ( vnp[i].transpose([1, 0] ).copy() ) # In Mesh-Tensorflow, it is one array, so it is divided _UpperCamelCase = torch.tensor(__snake_case ) elif key_name.startswith('''model/mlp''' ): _UpperCamelCase = int(key_name[9:].split('''/''' )[0] ) if key_name.endswith('''/p1/kernel''' ): _UpperCamelCase = '''model.blocks.%d.feed_forward.mlp.wi.weight''' % player _UpperCamelCase = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix _UpperCamelCase = torch.tensor(__snake_case ) elif key_name.endswith('''/p1/bias''' ): _UpperCamelCase = '''model.blocks.%d.feed_forward.mlp.wi.bias''' % player _UpperCamelCase = vnp.copy() # same because it is one dimensional _UpperCamelCase = torch.tensor(__snake_case ) elif key_name.endswith('''/p2/kernel''' ): _UpperCamelCase = '''model.blocks.%d.feed_forward.mlp.wo.weight''' % player _UpperCamelCase = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix _UpperCamelCase = torch.tensor(__snake_case ) elif key_name.endswith('''/p2/bias''' ): _UpperCamelCase = '''model.blocks.%d.feed_forward.mlp.wo.bias''' % player _UpperCamelCase = vnp.copy() # same because it is one dimensional _UpperCamelCase = torch.tensor(__snake_case ) elif key_name.startswith('''model/ln''' ): _UpperCamelCase = int(key_name[8:].split('''/''' )[0] ) if key_name.endswith('''/b''' ): _UpperCamelCase = '''model.blocks.%d.feed_forward.norm.bias''' % player _UpperCamelCase = vnp.copy() # same because it is one dimensional _UpperCamelCase = torch.tensor(__snake_case ) elif key_name.endswith('''/g''' ): _UpperCamelCase = '''model.blocks.%d.feed_forward.norm.weight''' % player _UpperCamelCase = vnp.copy() # same because it is one dimensional _UpperCamelCase = torch.tensor(__snake_case ) elif key_name.startswith('''model/att''' ): _UpperCamelCase = int(key_name[9:].split('''/''' )[0] ) if key_name.endswith('''/qkv/kernel''' ): _UpperCamelCase = vnp.copy() # Compute same dimension as Mesh-tensorflow using einsum _UpperCamelCase = state[:, 0, :, :] _UpperCamelCase = state[:, 1, :, :] _UpperCamelCase = state[:, 2, :, :] _UpperCamelCase = ( state_q.reshape([state_q.shape[0], state_q.shape[1] * state_q.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix _UpperCamelCase = ( state_k.reshape([state_k.shape[0], state_k.shape[1] * state_k.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix _UpperCamelCase = ( state_v.reshape([state_v.shape[0], state_v.shape[1] * state_v.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix _UpperCamelCase = '''model.blocks.%d.self_attn.self_attn.q_proj.weight''' % player _UpperCamelCase = torch.tensor(__snake_case ) _UpperCamelCase = '''model.blocks.%d.self_attn.self_attn.k_proj.weight''' % player _UpperCamelCase = torch.tensor(__snake_case ) _UpperCamelCase = '''model.blocks.%d.self_attn.self_attn.v_proj.weight''' % player _UpperCamelCase = torch.tensor(__snake_case ) elif key_name.endswith('''/o/kernel''' ): _UpperCamelCase = '''model.blocks.%d.self_attn.self_attn.out_proj.weight''' % player _UpperCamelCase = ( vnp.reshape([vnp.shape[0] * vnp.shape[1], vnp.shape[2]] ).transpose([1, 0] ).copy() ) # Mesh-Tensorflow is a diagonal matrix _UpperCamelCase = torch.tensor(__snake_case ) elif key_name.startswith('''model/an''' ): _UpperCamelCase = int(key_name[8:].split('''/''' )[0] ) if key_name.endswith('''/b''' ): _UpperCamelCase = '''model.blocks.%d.self_attn.norm.bias''' % player _UpperCamelCase = vnp.copy() # same because it is one dimensional _UpperCamelCase = torch.tensor(__snake_case ) elif key_name.endswith('''/g''' ): _UpperCamelCase = '''model.blocks.%d.self_attn.norm.weight''' % player _UpperCamelCase = vnp.copy() # same because it is one dimensional _UpperCamelCase = torch.tensor(__snake_case ) elif ( key_name.startswith('''model/wte''' ) or key_name.startswith('''model/wpe''' ) or key_name.startswith('''model/ete''' ) ): _UpperCamelCase = {'''wte''': '''embed_tokens''', '''wpe''': '''position_embeddings''', '''ete''': '''extra_position_embeddings'''}[ key_name[-3:] ] _UpperCamelCase = '''model.%s.weight''' % nlayer _UpperCamelCase = vnp.copy() # same in embedded _UpperCamelCase = torch.tensor(__snake_case ) if key_name.startswith('''model/wte''' ): _UpperCamelCase = '''lm_head.weight''' _UpperCamelCase = vnp.copy() # same in embedded _UpperCamelCase = torch.tensor(__snake_case ) elif key_name.startswith('''model/wob''' ): _UpperCamelCase = '''final_logits_bias''' _UpperCamelCase = vnp.copy() # same in embedded _UpperCamelCase = state.reshape((1, -1) ) _UpperCamelCase = torch.tensor(__snake_case ) elif key_name == "model/dense/kernel": _UpperCamelCase = '''model.last_project.weight''' _UpperCamelCase = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix _UpperCamelCase = torch.tensor(__snake_case ) elif key_name == "model/dense_1/bias": _UpperCamelCase = '''model.last_project.bias''' _UpperCamelCase = vnp.copy() # same because it is one dimensional _UpperCamelCase = torch.tensor(__snake_case ) torch.save(__snake_case, args.output ) if __name__ == "__main__": _a = argparse.ArgumentParser( description="""model converter.""", formatter_class=argparse.ArgumentDefaultsHelpFormatter ) parser.add_argument("""--tf_model_dir""", metavar="""PATH""", type=str, required=True, help="""import model""") parser.add_argument("""--output""", metavar="""PATH""", type=str, required=True, help="""output model""") _a = parser.parse_args() convert_tf_gptsan_to_pt(args)
78
"""simple docstring""" import argparse import torch from transformers import OpenAIGPTConfig, OpenAIGPTModel, load_tf_weights_in_openai_gpt from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> Union[str, Any]: """simple docstring""" if openai_config_file == "": _UpperCamelCase = OpenAIGPTConfig() else: _UpperCamelCase = OpenAIGPTConfig.from_json_file(__snake_case ) _UpperCamelCase = OpenAIGPTModel(__snake_case ) # Load weights from numpy load_tf_weights_in_openai_gpt(__snake_case, __snake_case, __snake_case ) # Save pytorch-model _UpperCamelCase = pytorch_dump_folder_path + '''/''' + WEIGHTS_NAME _UpperCamelCase = pytorch_dump_folder_path + '''/''' + CONFIG_NAME print(F'''Save PyTorch model to {pytorch_weights_dump_path}''' ) torch.save(model.state_dict(), __snake_case ) print(F'''Save configuration file to {pytorch_config_dump_path}''' ) with open(__snake_case, '''w''', encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( """--openai_checkpoint_folder_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--openai_config_file""", default="""""", type=str, help=( """An optional config json file corresponding to the pre-trained OpenAI model. \n""" """This specifies the model architecture.""" ), ) _a = parser.parse_args() convert_openai_checkpoint_to_pytorch( args.openai_checkpoint_folder_path, args.openai_config_file, args.pytorch_dump_folder_path )
78
1
"""simple docstring""" import warnings from typing import Dict, List, Optional, Tuple from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _a = logging.get_logger(__name__) class _UpperCAmelCase( lowerCamelCase ): lowercase__ = ['input_ids', 'attention_mask'] def __init__( self , __a="</s>" , __a="<unk>" , __a="<pad>" , __a=1_25 , __a=None , **__a , ) -> None: '''simple docstring''' # Add extra_ids to the special token list if extra_ids > 0 and additional_special_tokens is None: _UpperCamelCase = [F'''<extra_id_{i}>''' for i in range(__a)] elif extra_ids > 0 and additional_special_tokens is not None: # Check that we have the right number of extra_id special tokens _UpperCamelCase = len(set(filter(lambda __a: bool('''extra_id''' in str(__a)) , __a))) if extra_tokens != extra_ids: raise ValueError( F'''Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are''' ''' provided to ByT5Tokenizer. In this case the additional_special_tokens must include the''' ''' extra_ids tokens''') _UpperCamelCase = AddedToken(__a , lstrip=__a , rstrip=__a) if isinstance(__a , __a) else pad_token _UpperCamelCase = AddedToken(__a , lstrip=__a , rstrip=__a) if isinstance(__a , __a) else eos_token _UpperCamelCase = AddedToken(__a , lstrip=__a , rstrip=__a) if isinstance(__a , __a) else unk_token super().__init__( eos_token=__a , unk_token=__a , pad_token=__a , extra_ids=__a , additional_special_tokens=__a , **__a , ) _UpperCamelCase = extra_ids _UpperCamelCase = 2**8 # utf is 8 bits # define special tokens dict _UpperCamelCase = { self.pad_token: 0, self.eos_token: 1, self.unk_token: 2, } _UpperCamelCase = len(self.special_tokens_encoder) _UpperCamelCase = len(__a) for i, token in enumerate(__a): _UpperCamelCase = self.vocab_size + i - n _UpperCamelCase = {v: k for k, v in self.special_tokens_encoder.items()} @property def UpperCAmelCase ( self) -> Dict: '''simple docstring''' return self._utf_vocab_size + self._num_special_tokens + self._extra_ids def UpperCAmelCase ( self , __a , __a = None , __a = False) -> List[int]: '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__a , token_ids_a=__a , already_has_special_tokens=__a) # normal case: some special tokens if token_ids_a is None: return ([0] * len(__a)) + [1] return ([0] * len(__a)) + [1] + ([0] * len(__a)) + [1] def UpperCAmelCase ( self , __a) -> List[int]: '''simple docstring''' if len(__a) > 0 and token_ids[-1] == self.eos_token_id: warnings.warn( F'''This sequence already has {self.eos_token}. In future versions this behavior may lead to duplicated''' ''' eos tokens being added.''') return token_ids else: return token_ids + [self.eos_token_id] def UpperCAmelCase ( self , __a , __a = None) -> List[int]: '''simple docstring''' _UpperCamelCase = [self.eos_token_id] if token_ids_a is None: return len(token_ids_a + eos) * [0] return len(token_ids_a + eos + token_ids_a + eos) * [0] def UpperCAmelCase ( self , __a , __a = None) -> List[int]: '''simple docstring''' _UpperCamelCase = self._add_eos_if_not_present(__a) if token_ids_a is None: return token_ids_a else: _UpperCamelCase = self._add_eos_if_not_present(__a) return token_ids_a + token_ids_a def UpperCAmelCase ( self , __a) -> List[str]: '''simple docstring''' _UpperCamelCase = [chr(__a) for i in text.encode('''utf-8''')] return tokens def UpperCAmelCase ( self , __a) -> Any: '''simple docstring''' if token in self.special_tokens_encoder: _UpperCamelCase = self.special_tokens_encoder[token] elif token in self.added_tokens_encoder: _UpperCamelCase = self.added_tokens_encoder[token] elif len(__a) != 1: _UpperCamelCase = self.unk_token_id else: _UpperCamelCase = ord(__a) + self._num_special_tokens return token_id def UpperCAmelCase ( self , __a) -> Tuple: '''simple docstring''' if index in self.special_tokens_decoder: _UpperCamelCase = self.special_tokens_decoder[index] else: _UpperCamelCase = chr(index - self._num_special_tokens) return token def UpperCAmelCase ( self , __a) -> Any: '''simple docstring''' _UpperCamelCase = B'''''' for token in tokens: if token in self.special_tokens_decoder: _UpperCamelCase = self.special_tokens_decoder[token].encode('''utf-8''') elif token in self.added_tokens_decoder: _UpperCamelCase = self.special_tokens_decoder[token].encode('''utf-8''') elif token in self.special_tokens_encoder: _UpperCamelCase = token.encode('''utf-8''') elif token in self.added_tokens_encoder: _UpperCamelCase = token.encode('''utf-8''') else: _UpperCamelCase = bytes([ord(__a)]) bstring += tok_string _UpperCamelCase = bstring.decode('''utf-8''' , errors='''ignore''') return string def UpperCAmelCase ( self , __a , __a = None) -> Tuple[str]: '''simple docstring''' return ()
78
"""simple docstring""" from __future__ import annotations import unittest from transformers import AutoTokenizer, MBartConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFMBartForConditionalGeneration, TFMBartModel @require_tf class _UpperCAmelCase: lowercase__ = MBartConfig lowercase__ = {} lowercase__ = 'gelu' def __init__( self , __a , __a=13 , __a=7 , __a=True , __a=False , __a=99 , __a=32 , __a=2 , __a=4 , __a=37 , __a=0.1 , __a=0.1 , __a=20 , __a=2 , __a=1 , __a=0 , ) -> Any: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = seq_length _UpperCamelCase = is_training _UpperCamelCase = use_labels _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = eos_token_id _UpperCamelCase = pad_token_id _UpperCamelCase = bos_token_id def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size) _UpperCamelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size) , 1) _UpperCamelCase = tf.concat([input_ids, eos_tensor] , axis=1) _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) _UpperCamelCase = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) _UpperCamelCase = prepare_mbart_inputs_dict(__a , __a , __a) return config, inputs_dict def UpperCAmelCase ( self , __a , __a) -> Optional[int]: '''simple docstring''' _UpperCamelCase = TFMBartModel(config=__a).get_decoder() _UpperCamelCase = inputs_dict['''input_ids'''] _UpperCamelCase = input_ids[:1, :] _UpperCamelCase = inputs_dict['''attention_mask'''][:1, :] _UpperCamelCase = inputs_dict['''head_mask'''] _UpperCamelCase = 1 # first forward pass _UpperCamelCase = model(__a , attention_mask=__a , head_mask=__a , use_cache=__a) _UpperCamelCase , _UpperCamelCase = outputs.to_tuple() _UpperCamelCase = past_key_values[1] def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case, __snake_case=None, __snake_case=None, __snake_case=None, __snake_case=None, __snake_case=None, ) -> Optional[int]: """simple docstring""" if attention_mask is None: _UpperCamelCase = tf.cast(tf.math.not_equal(__snake_case, config.pad_token_id ), tf.inta ) if decoder_attention_mask is None: _UpperCamelCase = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape, dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:], config.pad_token_id ), tf.inta ), ], axis=-1, ) if head_mask is None: _UpperCamelCase = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _UpperCamelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: _UpperCamelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class _UpperCAmelCase( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase__ = (TFMBartForConditionalGeneration, TFMBartModel) if is_tf_available() else () lowercase__ = (TFMBartForConditionalGeneration,) if is_tf_available() else () lowercase__ = ( { 'conversational': TFMBartForConditionalGeneration, 'feature-extraction': TFMBartModel, 'summarization': TFMBartForConditionalGeneration, 'text2text-generation': TFMBartForConditionalGeneration, 'translation': TFMBartForConditionalGeneration, } if is_tf_available() else {} ) lowercase__ = True lowercase__ = False lowercase__ = False def UpperCAmelCase ( self , __a , __a , __a , __a , __a) -> Dict: '''simple docstring''' if pipeline_test_casse_name != "FeatureExtractionPipelineTests": # Exception encountered when calling layer '...' return True return False def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = TFMBartModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=__a) def UpperCAmelCase ( self) -> str: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__a) @require_sentencepiece @require_tokenizers @require_tf class _UpperCAmelCase( unittest.TestCase ): lowercase__ = [ ' UN Chief Says There Is No Military Solution in Syria', ] lowercase__ = [ 'Şeful ONU declară că nu există o soluţie militară în Siria', ] lowercase__ = 'facebook/mbart-large-en-ro' @cached_property def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' return AutoTokenizer.from_pretrained(self.model_name) @cached_property def UpperCAmelCase ( self) -> str: '''simple docstring''' _UpperCamelCase = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name) return model def UpperCAmelCase ( self , **__a) -> List[str]: '''simple docstring''' _UpperCamelCase = self.translate_src_text(**__a) self.assertListEqual(self.expected_text , __a) def UpperCAmelCase ( self , **__a) -> Dict: '''simple docstring''' _UpperCamelCase = self.tokenizer(self.src_text , **__a , return_tensors='''tf''') _UpperCamelCase = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2) _UpperCamelCase = self.tokenizer.batch_decode(__a , skip_special_tokens=__a) return generated_words @slow def UpperCAmelCase ( self) -> Any: '''simple docstring''' self._assert_generated_batch_equal_expected()
78
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _a = { """configuration_squeezebert""": [ """SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """SqueezeBertConfig""", """SqueezeBertOnnxConfig""", ], """tokenization_squeezebert""": ["""SqueezeBertTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ["""SqueezeBertTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ """SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """SqueezeBertForMaskedLM""", """SqueezeBertForMultipleChoice""", """SqueezeBertForQuestionAnswering""", """SqueezeBertForSequenceClassification""", """SqueezeBertForTokenClassification""", """SqueezeBertModel""", """SqueezeBertModule""", """SqueezeBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_squeezebert import ( SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, SqueezeBertConfig, SqueezeBertOnnxConfig, ) from .tokenization_squeezebert import SqueezeBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_squeezebert_fast import SqueezeBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_squeezebert import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, SqueezeBertModule, SqueezeBertPreTrainedModel, ) else: import sys _a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
78
"""simple docstring""" from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _a = logging.get_logger(__name__) class _UpperCAmelCase( lowerCamelCase ): lowercase__ = ['pixel_values'] def __init__( self , __a = True , __a = 1 / 2_55 , __a = True , __a = 8 , **__a , ) -> None: '''simple docstring''' super().__init__(**__a) _UpperCamelCase = do_rescale _UpperCamelCase = rescale_factor _UpperCamelCase = do_pad _UpperCamelCase = pad_size def UpperCAmelCase ( self , __a , __a , __a = None , **__a) -> np.ndarray: '''simple docstring''' return rescale(__a , scale=__a , data_format=__a , **__a) def UpperCAmelCase ( self , __a , __a , __a = None) -> List[Any]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = get_image_size(__a) _UpperCamelCase = (old_height // size + 1) * size - old_height _UpperCamelCase = (old_width // size + 1) * size - old_width return pad(__a , ((0, pad_height), (0, pad_width)) , mode='''symmetric''' , data_format=__a) def UpperCAmelCase ( self , __a , __a = None , __a = None , __a = None , __a = None , __a = None , __a = ChannelDimension.FIRST , **__a , ) -> Tuple: '''simple docstring''' _UpperCamelCase = do_rescale if do_rescale is not None else self.do_rescale _UpperCamelCase = rescale_factor if rescale_factor is not None else self.rescale_factor _UpperCamelCase = do_pad if do_pad is not None else self.do_pad _UpperCamelCase = pad_size if pad_size is not None else self.pad_size _UpperCamelCase = make_list_of_images(__a) if not valid_images(__a): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''') if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''') # All transformations expect numpy arrays. _UpperCamelCase = [to_numpy_array(__a) for image in images] if do_rescale: _UpperCamelCase = [self.rescale(image=__a , scale=__a) for image in images] if do_pad: _UpperCamelCase = [self.pad(__a , size=__a) for image in images] _UpperCamelCase = [to_channel_dimension_format(__a , __a) for image in images] _UpperCamelCase = {'''pixel_values''': images} return BatchFeature(data=__a , tensor_type=__a)
78
1
"""simple docstring""" import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / """utils""")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class _UpperCAmelCase( unittest.TestCase ): def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = 0 def UpperCAmelCase ( self) -> str: '''simple docstring''' _UpperCamelCase = AutoImageProcessor.from_pretrained('''openai/clip-vit-base-patch32''') self.assertIsInstance(__a , __a) def UpperCAmelCase ( self) -> Dict: '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdirname: _UpperCamelCase = Path(__a) / '''preprocessor_config.json''' _UpperCamelCase = Path(__a) / '''config.json''' json.dump( {'''image_processor_type''': '''CLIPImageProcessor''', '''processor_class''': '''CLIPProcessor'''} , open(__a , '''w''') , ) json.dump({'''model_type''': '''clip'''} , open(__a , '''w''')) _UpperCamelCase = AutoImageProcessor.from_pretrained(__a) self.assertIsInstance(__a , __a) def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: _UpperCamelCase = Path(__a) / '''preprocessor_config.json''' _UpperCamelCase = Path(__a) / '''config.json''' json.dump( {'''feature_extractor_type''': '''CLIPFeatureExtractor''', '''processor_class''': '''CLIPProcessor'''} , open(__a , '''w''') , ) json.dump({'''model_type''': '''clip'''} , open(__a , '''w''')) _UpperCamelCase = AutoImageProcessor.from_pretrained(__a) self.assertIsInstance(__a , __a) def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdirname: _UpperCamelCase = CLIPConfig() # Create a dummy config file with image_proceesor_type _UpperCamelCase = Path(__a) / '''preprocessor_config.json''' _UpperCamelCase = Path(__a) / '''config.json''' json.dump( {'''image_processor_type''': '''CLIPImageProcessor''', '''processor_class''': '''CLIPProcessor'''} , open(__a , '''w''') , ) json.dump({'''model_type''': '''clip'''} , open(__a , '''w''')) # remove image_processor_type to make sure config.json alone is enough to load image processor locally _UpperCamelCase = AutoImageProcessor.from_pretrained(__a).to_dict() config_dict.pop('''image_processor_type''') _UpperCamelCase = CLIPImageProcessor(**__a) # save in new folder model_config.save_pretrained(__a) config.save_pretrained(__a) _UpperCamelCase = AutoImageProcessor.from_pretrained(__a) # make sure private variable is not incorrectly saved _UpperCamelCase = json.loads(config.to_json_string()) self.assertTrue('''_processor_class''' not in dict_as_saved) self.assertIsInstance(__a , __a) def UpperCAmelCase ( self) -> int: '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdirname: _UpperCamelCase = Path(__a) / '''preprocessor_config.json''' json.dump( {'''image_processor_type''': '''CLIPImageProcessor''', '''processor_class''': '''CLIPProcessor'''} , open(__a , '''w''') , ) _UpperCamelCase = AutoImageProcessor.from_pretrained(__a) self.assertIsInstance(__a , __a) def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' with self.assertRaisesRegex( __a , '''clip-base is not a local folder and is not a valid model identifier'''): _UpperCamelCase = AutoImageProcessor.from_pretrained('''clip-base''') def UpperCAmelCase ( self) -> Dict: '''simple docstring''' with self.assertRaisesRegex( __a , R'''aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)'''): _UpperCamelCase = AutoImageProcessor.from_pretrained(__a , revision='''aaaaaa''') def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' with self.assertRaisesRegex( __a , '''hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.''' , ): _UpperCamelCase = AutoImageProcessor.from_pretrained('''hf-internal-testing/config-no-model''') def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(__a): _UpperCamelCase = AutoImageProcessor.from_pretrained('''hf-internal-testing/test_dynamic_image_processor''') # If remote code is disabled, we can't load this config. with self.assertRaises(__a): _UpperCamelCase = AutoImageProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_image_processor''' , trust_remote_code=__a) _UpperCamelCase = AutoImageProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_image_processor''' , trust_remote_code=__a) self.assertEqual(image_processor.__class__.__name__ , '''NewImageProcessor''') # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(__a) _UpperCamelCase = AutoImageProcessor.from_pretrained(__a , trust_remote_code=__a) self.assertEqual(reloaded_image_processor.__class__.__name__ , '''NewImageProcessor''') def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' try: AutoConfig.register('''custom''' , __a) AutoImageProcessor.register(__a , __a) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(__a): AutoImageProcessor.register(__a , __a) with tempfile.TemporaryDirectory() as tmpdirname: _UpperCamelCase = Path(__a) / '''preprocessor_config.json''' _UpperCamelCase = Path(__a) / '''config.json''' json.dump( {'''feature_extractor_type''': '''CLIPFeatureExtractor''', '''processor_class''': '''CLIPProcessor'''} , open(__a , '''w''') , ) json.dump({'''model_type''': '''clip'''} , open(__a , '''w''')) _UpperCamelCase = CustomImageProcessor.from_pretrained(__a) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(__a) _UpperCamelCase = AutoImageProcessor.from_pretrained(__a) self.assertIsInstance(__a , __a) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' class _UpperCAmelCase( lowerCamelCase ): lowercase__ = True try: AutoConfig.register('''custom''' , __a) AutoImageProcessor.register(__a , __a) # If remote code is not set, the default is to use local _UpperCamelCase = AutoImageProcessor.from_pretrained('''hf-internal-testing/test_dynamic_image_processor''') self.assertEqual(image_processor.__class__.__name__ , '''NewImageProcessor''') self.assertTrue(image_processor.is_local) # If remote code is disabled, we load the local one. _UpperCamelCase = AutoImageProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_image_processor''' , trust_remote_code=__a) self.assertEqual(image_processor.__class__.__name__ , '''NewImageProcessor''') self.assertTrue(image_processor.is_local) # If remote is enabled, we load from the Hub _UpperCamelCase = AutoImageProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_image_processor''' , trust_remote_code=__a) self.assertEqual(image_processor.__class__.__name__ , '''NewImageProcessor''') self.assertTrue(not hasattr(__a , '''is_local''')) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
78
"""simple docstring""" from importlib import import_module from .logging import get_logger _a = get_logger(__name__) class _UpperCAmelCase: def __init__( self , __a , __a=None) -> Dict: '''simple docstring''' _UpperCamelCase = attrs or [] if module is not None: for key in module.__dict__: if key in attrs or not key.startswith('''__'''): setattr(self , __a , getattr(__a , __a)) _UpperCamelCase = module._original_module if isinstance(__a , _PatchedModuleObj) else module class _UpperCAmelCase: lowercase__ = [] def __init__( self , __a , __a , __a , __a=None) -> List[str]: '''simple docstring''' _UpperCamelCase = obj _UpperCamelCase = target _UpperCamelCase = new _UpperCamelCase = target.split('''.''')[0] _UpperCamelCase = {} _UpperCamelCase = attrs or [] def __enter__( self) -> int: '''simple docstring''' *_UpperCamelCase , _UpperCamelCase = self.target.split('''.''') # Patch modules: # it's used to patch attributes of submodules like "os.path.join"; # in this case we need to patch "os" and "os.path" for i in range(len(__a)): try: _UpperCamelCase = import_module('''.'''.join(submodules[: i + 1])) except ModuleNotFoundError: continue # We iterate over all the globals in self.obj in case we find "os" or "os.path" for attr in self.obj.__dir__(): _UpperCamelCase = getattr(self.obj , __a) # We don't check for the name of the global, but rather if its value *is* "os" or "os.path". # This allows to patch renamed modules like "from os import path as ospath". if obj_attr is submodule or ( (isinstance(__a , _PatchedModuleObj) and obj_attr._original_module is submodule) ): _UpperCamelCase = obj_attr # patch at top level setattr(self.obj , __a , _PatchedModuleObj(__a , attrs=self.attrs)) _UpperCamelCase = getattr(self.obj , __a) # construct lower levels patches for key in submodules[i + 1 :]: setattr(__a , __a , _PatchedModuleObj(getattr(__a , __a , __a) , attrs=self.attrs)) _UpperCamelCase = getattr(__a , __a) # finally set the target attribute setattr(__a , __a , self.new) # Patch attribute itself: # it's used for builtins like "open", # and also to patch "os.path.join" we may also need to patch "join" # itself if it was imported as "from os.path import join". if submodules: # if it's an attribute of a submodule like "os.path.join" try: _UpperCamelCase = getattr(import_module('''.'''.join(__a)) , __a) except (AttributeError, ModuleNotFoundError): return # We iterate over all the globals in self.obj in case we find "os.path.join" for attr in self.obj.__dir__(): # We don't check for the name of the global, but rather if its value *is* "os.path.join". # This allows to patch renamed attributes like "from os.path import join as pjoin". if getattr(self.obj , __a) is attr_value: _UpperCamelCase = getattr(self.obj , __a) setattr(self.obj , __a , self.new) elif target_attr in globals()["__builtins__"]: # if it'a s builtin like "open" _UpperCamelCase = globals()['''__builtins__'''][target_attr] setattr(self.obj , __a , self.new) else: raise RuntimeError(F'''Tried to patch attribute {target_attr} instead of a submodule.''') def __exit__( self , *__a) -> Tuple: '''simple docstring''' for attr in list(self.original): setattr(self.obj , __a , self.original.pop(__a)) def UpperCAmelCase ( self) -> Dict: '''simple docstring''' self.__enter__() self._active_patches.append(self) def UpperCAmelCase ( self) -> str: '''simple docstring''' try: self._active_patches.remove(self) except ValueError: # If the patch hasn't been started this will fail return None return self.__exit__()
78
1
"""simple docstring""" from pathlib import Path import fire def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> int: """simple docstring""" _UpperCamelCase = Path(__snake_case ) _UpperCamelCase = Path(__snake_case ) dest_dir.mkdir(exist_ok=__snake_case ) for path in src_dir.iterdir(): _UpperCamelCase = [x.rstrip() for x in list(path.open().readlines() )][:n] _UpperCamelCase = dest_dir.joinpath(path.name ) print(__snake_case ) dest_path.open('''w''' ).write('''\n'''.join(__snake_case ) ) if __name__ == "__main__": fire.Fire(minify)
78
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_flax_available, is_torch_available, is_transformers_available, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .multicontrolnet import MultiControlNetModel from .pipeline_controlnet import StableDiffusionControlNetPipeline from .pipeline_controlnet_imgaimg import StableDiffusionControlNetImgaImgPipeline from .pipeline_controlnet_inpaint import StableDiffusionControlNetInpaintPipeline if is_transformers_available() and is_flax_available(): from .pipeline_flax_controlnet import FlaxStableDiffusionControlNetPipeline
78
1
"""simple docstring""" import itertools import os from collections import Counter, defaultdict from concurrent.futures import ThreadPoolExecutor, as_completed import numpy as np import datasets from .execute import check_correctness _a = """\ @misc{chen2021evaluating, title={Evaluating Large Language Models Trained on Code}, author={Mark Chen and Jerry Tworek and Heewoo Jun and Qiming Yuan \ and Henrique Ponde de Oliveira Pinto and Jared Kaplan and Harri Edwards \ and Yuri Burda and Nicholas Joseph and Greg Brockman and Alex Ray \ and Raul Puri and Gretchen Krueger and Michael Petrov and Heidy Khlaaf \ and Girish Sastry and Pamela Mishkin and Brooke Chan and Scott Gray \ and Nick Ryder and Mikhail Pavlov and Alethea Power and Lukasz Kaiser \ and Mohammad Bavarian and Clemens Winter and Philippe Tillet \ and Felipe Petroski Such and Dave Cummings and Matthias Plappert \ and Fotios Chantzis and Elizabeth Barnes and Ariel Herbert-Voss \ and William Hebgen Guss and Alex Nichol and Alex Paino and Nikolas Tezak \ and Jie Tang and Igor Babuschkin and Suchir Balaji and Shantanu Jain \ and William Saunders and Christopher Hesse and Andrew N. Carr \ and Jan Leike and Josh Achiam and Vedant Misra and Evan Morikawa \ and Alec Radford and Matthew Knight and Miles Brundage and Mira Murati \ and Katie Mayer and Peter Welinder and Bob McGrew and Dario Amodei \ and Sam McCandlish and Ilya Sutskever and Wojciech Zaremba}, year={2021}, eprint={2107.03374}, archivePrefix={arXiv}, primaryClass={cs.LG} } """ _a = """\ This metric implements the evaluation harness for the HumanEval problem solving dataset described in the paper \"Evaluating Large Language Models Trained on Code\" (https://arxiv.org/abs/2107.03374). """ _a = """ Calculates how good are predictions given some references, using certain scores Args: predictions: list of candidates to evaluate. Each candidates should be a list of strings with several code candidates to solve the problem. references: a list with a test for each prediction. Each test should evaluate the correctness of a code candidate. k: number of code candidates to consider in the evaluation (Default: [1, 10, 100]) num_workers: number of workers used to evaluate the canidate programs (Default: 4). timeout: Returns: pass_at_k: dict with pass rates for each k results: dict with granular results of each unittest Examples: >>> code_eval = datasets.load_metric(\"code_eval\") >>> test_cases = [\"assert add(2,3)==5\"] >>> candidates = [[\"def add(a,b): return a*b\", \"def add(a, b): return a+b\"]] >>> pass_at_k, results = code_eval.compute(references=test_cases, predictions=candidates, k=[1, 2]) >>> print(pass_at_k) {'pass@1': 0.5, 'pass@2': 1.0} """ _a = """ ################################################################################ !!!WARNING!!! ################################################################################ The \"code_eval\" metric executes untrusted model-generated code in Python. Although it is highly unlikely that model-generated code will do something overtly malicious in response to this test suite, model-generated code may act destructively due to a lack of model capability or alignment. Users are strongly encouraged to sandbox this evaluation suite so that it does not perform destructive actions on their host or network. For more information on how OpenAI sandboxes its code, see the paper \"Evaluating Large Language Models Trained on Code\" (https://arxiv.org/abs/2107.03374). Once you have read this disclaimer and taken appropriate precautions, set the environment variable HF_ALLOW_CODE_EVAL=\"1\". Within Python you can to this with: >>> import os >>> os.environ[\"HF_ALLOW_CODE_EVAL\"] = \"1\" ################################################################################\ """ _a = """The MIT License Copyright (c) OpenAI (https://openai.com) Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the \"Software\"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.""" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _UpperCAmelCase( datasets.Metric ): def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' return datasets.MetricInfo( # This is the description that will appear on the metrics page. description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''')), '''references''': datasets.Value('''string'''), }) , homepage='''https://github.com/openai/human-eval''' , codebase_urls=['''https://github.com/openai/human-eval'''] , reference_urls=['''https://github.com/openai/human-eval'''] , license=_LICENSE , ) def UpperCAmelCase ( self , __a , __a , __a=[1, 10, 1_00] , __a=4 , __a=3.0) -> Optional[int]: '''simple docstring''' if os.getenv('''HF_ALLOW_CODE_EVAL''' , 0) != "1": raise ValueError(_WARNING) if os.name == "nt": raise NotImplementedError('''This metric is currently not supported on Windows.''') with ThreadPoolExecutor(max_workers=__a) as executor: _UpperCamelCase = [] _UpperCamelCase = Counter() _UpperCamelCase = 0 _UpperCamelCase = defaultdict(__a) for task_id, (candidates, test_case) in enumerate(zip(__a , __a)): for candidate in candidates: _UpperCamelCase = candidate + '''\n''' + test_case _UpperCamelCase = (test_program, timeout, task_id, completion_id[task_id]) _UpperCamelCase = executor.submit(__a , *__a) futures.append(__a) completion_id[task_id] += 1 n_samples += 1 for future in as_completed(__a): _UpperCamelCase = future.result() results[result["task_id"]].append((result['''completion_id'''], result)) _UpperCamelCase , _UpperCamelCase = [], [] for result in results.values(): result.sort() _UpperCamelCase = [r[1]['''passed'''] for r in result] total.append(len(__a)) correct.append(sum(__a)) _UpperCamelCase = np.array(__a) _UpperCamelCase = np.array(__a) _UpperCamelCase = k _UpperCamelCase = {F'''pass@{k}''': estimate_pass_at_k(__a , __a , __a).mean() for k in ks if (total >= k).all()} return pass_at_k, results def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> Dict: """simple docstring""" def estimator(__snake_case, __snake_case, __snake_case ) -> float: if n - c < k: return 1.0 return 1.0 - np.prod(1.0 - k / np.arange(n - c + 1, n + 1 ) ) if isinstance(__snake_case, __snake_case ): _UpperCamelCase = itertools.repeat(__snake_case, len(__snake_case ) ) else: assert len(__snake_case ) == len(__snake_case ) _UpperCamelCase = iter(__snake_case ) return np.array([estimator(int(__snake_case ), int(__snake_case ), __snake_case ) for n, c in zip(__snake_case, __snake_case )] )
78
"""simple docstring""" from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging _a = logging.get_logger(__name__) _a = { """EleutherAI/gpt-neo-1.3B""": """https://huggingface.co/EleutherAI/gpt-neo-1.3B/resolve/main/config.json""", # See all GPTNeo models at https://huggingface.co/models?filter=gpt_neo } class _UpperCAmelCase( lowerCamelCase ): lowercase__ = 'gpt_neo' lowercase__ = ['past_key_values'] lowercase__ = {'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers'} def __init__( self , __a=5_02_57 , __a=20_48 , __a=20_48 , __a=24 , __a=[[["global", "local"], 12]] , __a=16 , __a=None , __a=2_56 , __a="gelu_new" , __a=0.0 , __a=0.0 , __a=0.0 , __a=0.1 , __a=1e-5 , __a=0.02 , __a=True , __a=5_02_56 , __a=5_02_56 , **__a , ) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = vocab_size _UpperCamelCase = max_position_embeddings _UpperCamelCase = hidden_size _UpperCamelCase = num_layers _UpperCamelCase = num_heads _UpperCamelCase = intermediate_size _UpperCamelCase = window_size _UpperCamelCase = activation_function _UpperCamelCase = resid_dropout _UpperCamelCase = embed_dropout _UpperCamelCase = attention_dropout _UpperCamelCase = classifier_dropout _UpperCamelCase = layer_norm_epsilon _UpperCamelCase = initializer_range _UpperCamelCase = use_cache _UpperCamelCase = bos_token_id _UpperCamelCase = eos_token_id _UpperCamelCase = attention_types _UpperCamelCase = self.expand_attention_types_params(__a) if len(self.attention_layers) != self.num_layers: raise ValueError( '''Configuration for convolutional module is incorrect. ''' '''It is required that `len(config.attention_layers)` == `config.num_layers` ''' F'''but is `len(config.attention_layers) = {len(self.attention_layers)}`, ''' F'''`config.num_layers = {self.num_layers}`. ''' '''`config.attention_layers` is prepared using `config.attention_types`. ''' '''Please verify the value of `config.attention_types` argument.''') super().__init__(bos_token_id=__a , eos_token_id=__a , **__a) @staticmethod def UpperCAmelCase ( __a) -> int: '''simple docstring''' _UpperCamelCase = [] for item in attention_types: for _ in range(item[1]): attentions.extend(item[0]) return attentions def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case, __snake_case ) -> str: """simple docstring""" import torch _UpperCamelCase = input.size() _UpperCamelCase = len(__snake_case ) _UpperCamelCase = shape[dimension] _UpperCamelCase = torch.arange(0, __snake_case, __snake_case ) _UpperCamelCase = torch.div(sizedim - size, __snake_case, rounding_mode='''floor''' ) + 1 _UpperCamelCase = torch.arange(__snake_case ) + low_indices[:min_length][:, None] _UpperCamelCase = [slice(__snake_case )] * rank _UpperCamelCase = indices _UpperCamelCase = input[s] _UpperCamelCase = list(range(0, rank + 1 ) ) perm.append(perm.pop(dimension + 1 ) ) return sliced.permute(__snake_case ) def lowerCamelCase__ ( __snake_case, __snake_case ) -> str: """simple docstring""" import torch _UpperCamelCase = torch.arange(1, __snake_case ) _UpperCamelCase = torch.remainder(__snake_case, __snake_case ) _UpperCamelCase = remainders == 0 _UpperCamelCase = candidates[divisor_indices] _UpperCamelCase = torch.max(__snake_case ) return largest_divisor, torch.div(__snake_case, __snake_case, rounding_mode='''floor''' ) class _UpperCAmelCase( lowerCamelCase ): @property def UpperCAmelCase ( self) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' _UpperCamelCase = OrderedDict({'''input_ids''': {0: '''batch''', 1: '''sequence'''}}) if self.use_past: self.fill_with_past_key_values_(__a , direction='''inputs''') _UpperCamelCase = {0: '''batch''', 1: '''past_sequence + sequence'''} else: _UpperCamelCase = {0: '''batch''', 1: '''sequence'''} return common_inputs @property def UpperCAmelCase ( self) -> int: '''simple docstring''' return self._config.num_heads def UpperCAmelCase ( self , __a , __a = -1 , __a = -1 , __a = False , __a = None , ) -> Mapping[str, Any]: '''simple docstring''' _UpperCamelCase = super(__a , self).generate_dummy_inputs( __a , batch_size=__a , seq_length=__a , is_pair=__a , framework=__a) # We need to order the input in the way they appears in the forward() _UpperCamelCase = OrderedDict({'''input_ids''': common_inputs['''input_ids''']}) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''') else: import torch _UpperCamelCase , _UpperCamelCase = common_inputs['''input_ids'''].shape # Not using the same length for past_key_values _UpperCamelCase = seqlen + 2 _UpperCamelCase = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) _UpperCamelCase = [ (torch.zeros(__a), torch.zeros(__a)) for _ in range(self.num_layers) ] _UpperCamelCase = common_inputs['''attention_mask'''] if self.use_past: _UpperCamelCase = ordered_inputs['''attention_mask'''].dtype _UpperCamelCase = torch.cat( [ordered_inputs['''attention_mask'''], torch.ones(__a , __a , dtype=__a)] , dim=1) return ordered_inputs @property def UpperCAmelCase ( self) -> int: '''simple docstring''' return 13
78
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) _a = {"""configuration_fnet""": ["""FNET_PRETRAINED_CONFIG_ARCHIVE_MAP""", """FNetConfig"""]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ["""FNetTokenizer"""] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ["""FNetTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ """FNET_PRETRAINED_MODEL_ARCHIVE_LIST""", """FNetForMaskedLM""", """FNetForMultipleChoice""", """FNetForNextSentencePrediction""", """FNetForPreTraining""", """FNetForQuestionAnswering""", """FNetForSequenceClassification""", """FNetForTokenClassification""", """FNetLayer""", """FNetModel""", """FNetPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_fnet import FNET_PRETRAINED_CONFIG_ARCHIVE_MAP, FNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_fnet import FNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_fnet_fast import FNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_fnet import ( FNET_PRETRAINED_MODEL_ARCHIVE_LIST, FNetForMaskedLM, FNetForMultipleChoice, FNetForNextSentencePrediction, FNetForPreTraining, FNetForQuestionAnswering, FNetForSequenceClassification, FNetForTokenClassification, FNetLayer, FNetModel, FNetPreTrainedModel, ) else: import sys _a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
78
"""simple docstring""" import sys from collections import defaultdict class _UpperCAmelCase: def __init__( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = [] def UpperCAmelCase ( self , __a) -> Optional[Any]: '''simple docstring''' return self.node_position[vertex] def UpperCAmelCase ( self , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = pos def UpperCAmelCase ( self , __a , __a , __a , __a) -> Tuple: '''simple docstring''' if start > size // 2 - 1: return else: if 2 * start + 2 >= size: _UpperCamelCase = 2 * start + 1 else: if heap[2 * start + 1] < heap[2 * start + 2]: _UpperCamelCase = 2 * start + 1 else: _UpperCamelCase = 2 * start + 2 if heap[smallest_child] < heap[start]: _UpperCamelCase , _UpperCamelCase = heap[smallest_child], positions[smallest_child] _UpperCamelCase , _UpperCamelCase = ( heap[start], positions[start], ) _UpperCamelCase , _UpperCamelCase = temp, tempa _UpperCamelCase = self.get_position(positions[smallest_child]) self.set_position( positions[smallest_child] , self.get_position(positions[start])) self.set_position(positions[start] , __a) self.top_to_bottom(__a , __a , __a , __a) def UpperCAmelCase ( self , __a , __a , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = position[index] while index != 0: _UpperCamelCase = int((index - 2) / 2) if index % 2 == 0 else int((index - 1) / 2) if val < heap[parent]: _UpperCamelCase = heap[parent] _UpperCamelCase = position[parent] self.set_position(position[parent] , __a) else: _UpperCamelCase = val _UpperCamelCase = temp self.set_position(__a , __a) break _UpperCamelCase = parent else: _UpperCamelCase = val _UpperCamelCase = temp self.set_position(__a , 0) def UpperCAmelCase ( self , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = len(__a) // 2 - 1 for i in range(__a , -1 , -1): self.top_to_bottom(__a , __a , len(__a) , __a) def UpperCAmelCase ( self , __a , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = positions[0] _UpperCamelCase = sys.maxsize self.top_to_bottom(__a , 0 , len(__a) , __a) return temp def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" _UpperCamelCase = Heap() _UpperCamelCase = [0] * len(__snake_case ) _UpperCamelCase = [-1] * len(__snake_case ) # Neighboring Tree Vertex of selected vertex # Minimum Distance of explored vertex with neighboring vertex of partial tree # formed in graph _UpperCamelCase = [] # Heap of Distance of vertices from their neighboring vertex _UpperCamelCase = [] for vertex in range(len(__snake_case ) ): distance_tv.append(sys.maxsize ) positions.append(__snake_case ) heap.node_position.append(__snake_case ) _UpperCamelCase = [] _UpperCamelCase = 1 _UpperCamelCase = sys.maxsize for neighbor, distance in adjacency_list[0]: _UpperCamelCase = 0 _UpperCamelCase = distance heap.heapify(__snake_case, __snake_case ) for _ in range(1, len(__snake_case ) ): _UpperCamelCase = heap.delete_minimum(__snake_case, __snake_case ) if visited[vertex] == 0: tree_edges.append((nbr_tv[vertex], vertex) ) _UpperCamelCase = 1 for neighbor, distance in adjacency_list[vertex]: if ( visited[neighbor] == 0 and distance < distance_tv[heap.get_position(__snake_case )] ): _UpperCamelCase = distance heap.bottom_to_top( __snake_case, heap.get_position(__snake_case ), __snake_case, __snake_case ) _UpperCamelCase = vertex return tree_edges if __name__ == "__main__": # pragma: no cover # < --------- Prims Algorithm --------- > _a = int(input("""Enter number of edges: """).strip()) _a = defaultdict(list) for _ in range(edges_number): _a = [int(x) for x in input().strip().split()] adjacency_list[edge[0]].append([edge[1], edge[2]]) adjacency_list[edge[1]].append([edge[0], edge[2]]) print(prisms_algorithm(adjacency_list))
78
1
"""simple docstring""" import os def lowerCamelCase__ ( ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = os.path.dirname(os.path.realpath(__snake_case ) ) _UpperCamelCase = os.path.join(__snake_case, '''triangle.txt''' ) with open(__snake_case ) as f: _UpperCamelCase = f.readlines() _UpperCamelCase = [] for line in triangle: _UpperCamelCase = [] for number in line.strip().split(''' ''' ): numbers_from_line.append(int(__snake_case ) ) a.append(__snake_case ) for i in range(1, len(__snake_case ) ): for j in range(len(a[i] ) ): _UpperCamelCase = a[i - 1][j] if j != len(a[i - 1] ) else 0 _UpperCamelCase = a[i - 1][j - 1] if j > 0 else 0 a[i][j] += max(__snake_case, __snake_case ) return max(a[-1] ) if __name__ == "__main__": print(solution())
78
"""simple docstring""" import json import sys def lowerCamelCase__ ( __snake_case, __snake_case ) -> Union[str, Any]: """simple docstring""" with open(__snake_case, encoding='''utf-8''' ) as f: _UpperCamelCase = json.load(__snake_case ) _UpperCamelCase = ['''<details>''', '''<summary>Show updated benchmarks!</summary>''', ''' '''] for benchmark_name in sorted(__snake_case ): _UpperCamelCase = results[benchmark_name] _UpperCamelCase = benchmark_name.split('''/''' )[-1] output_md.append(F'''### Benchmark: {benchmark_file_name}''' ) _UpperCamelCase = '''| metric |''' _UpperCamelCase = '''|--------|''' _UpperCamelCase = '''| new / old (diff) |''' for metric_name in sorted(__snake_case ): _UpperCamelCase = benchmark_res[metric_name] _UpperCamelCase = metric_vals['''new'''] _UpperCamelCase = metric_vals.get('''old''', __snake_case ) _UpperCamelCase = metric_vals.get('''diff''', __snake_case ) _UpperCamelCase = F''' {new_val:f}''' if isinstance(__snake_case, (int, float) ) else '''None''' if old_val is not None: val_str += F''' / {old_val:f}''' if isinstance(__snake_case, (int, float) ) else "None" if dif_val is not None: val_str += F''' ({dif_val:f})''' if isinstance(__snake_case, (int, float) ) else "None" title += " " + metric_name + " |" lines += "---|" value += val_str + " |" output_md += [title, lines, value, " "] output_md.append('''</details>''' ) with open(__snake_case, '''w''', encoding='''utf-8''' ) as f: f.writelines('''\n'''.join(__snake_case ) ) if __name__ == "__main__": _a = sys.argv[1] _a = sys.argv[2] format_json_to_md(input_json_file, output_md_file)
78
1
"""simple docstring""" from collections import namedtuple _a = namedtuple("""from_to""", """from_ to""") _a = { """cubicmeter""": from_to(1, 1), """litre""": from_to(0.001, 1000), """kilolitre""": from_to(1, 1), """gallon""": from_to(0.0_0454, 264.172), """cubicyard""": from_to(0.7_6455, 1.3_0795), """cubicfoot""": from_to(0.028, 35.3147), """cup""": from_to(0.0_0023_6588, 4226.75), } def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> float: """simple docstring""" if from_type not in METRIC_CONVERSION: raise ValueError( F'''Invalid \'from_type\' value: {from_type!r} Supported values are:\n''' + ''', '''.join(__snake_case ) ) if to_type not in METRIC_CONVERSION: raise ValueError( F'''Invalid \'to_type\' value: {to_type!r}. Supported values are:\n''' + ''', '''.join(__snake_case ) ) return value * METRIC_CONVERSION[from_type].from_ * METRIC_CONVERSION[to_type].to if __name__ == "__main__": import doctest doctest.testmod()
78
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTImageProcessor, ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) def lowerCamelCase__ ( __snake_case, __snake_case=False ) -> Tuple: """simple docstring""" _UpperCamelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'''blocks.{i}.norm1.weight''', F'''vit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((F'''blocks.{i}.norm1.bias''', F'''vit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((F'''blocks.{i}.attn.proj.weight''', F'''vit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.attn.proj.bias''', F'''vit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((F'''blocks.{i}.norm2.weight''', F'''vit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((F'''blocks.{i}.norm2.bias''', F'''vit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.weight''', F'''vit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.bias''', F'''vit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.weight''', F'''vit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.bias''', F'''vit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ('''cls_token''', '''vit.embeddings.cls_token'''), ('''patch_embed.proj.weight''', '''vit.embeddings.patch_embeddings.projection.weight'''), ('''patch_embed.proj.bias''', '''vit.embeddings.patch_embeddings.projection.bias'''), ('''pos_embed''', '''vit.embeddings.position_embeddings'''), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ('''pre_logits.fc.weight''', '''pooler.dense.weight'''), ('''pre_logits.fc.bias''', '''pooler.dense.bias'''), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _UpperCamelCase = [(pair[0], pair[1][4:]) if pair[1].startswith('''vit''' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('''norm.weight''', '''vit.layernorm.weight'''), ('''norm.bias''', '''vit.layernorm.bias'''), ('''head.weight''', '''classifier.weight'''), ('''head.bias''', '''classifier.bias'''), ] ) return rename_keys def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case=False ) -> str: """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: _UpperCamelCase = '''''' else: _UpperCamelCase = '''vit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _UpperCamelCase = state_dict.pop(F'''blocks.{i}.attn.qkv.weight''' ) _UpperCamelCase = state_dict.pop(F'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict _UpperCamelCase = in_proj_weight[ : config.hidden_size, : ] _UpperCamelCase = in_proj_bias[: config.hidden_size] _UpperCamelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _UpperCamelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _UpperCamelCase = in_proj_weight[ -config.hidden_size :, : ] _UpperCamelCase = in_proj_bias[-config.hidden_size :] def lowerCamelCase__ ( __snake_case ) -> Dict: """simple docstring""" _UpperCamelCase = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(__snake_case, __snake_case ) def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> Any: """simple docstring""" _UpperCamelCase = dct.pop(__snake_case ) _UpperCamelCase = val def lowerCamelCase__ ( ) -> Dict: """simple docstring""" _UpperCamelCase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _UpperCamelCase = Image.open(requests.get(__snake_case, stream=__snake_case ).raw ) return im @torch.no_grad() def lowerCamelCase__ ( __snake_case, __snake_case ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = ViTConfig() _UpperCamelCase = False # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size if vit_name[-5:] == "in21k": _UpperCamelCase = True _UpperCamelCase = int(vit_name[-12:-10] ) _UpperCamelCase = int(vit_name[-9:-6] ) else: _UpperCamelCase = 10_00 _UpperCamelCase = '''huggingface/label-files''' _UpperCamelCase = '''imagenet-1k-id2label.json''' _UpperCamelCase = json.load(open(hf_hub_download(__snake_case, __snake_case, repo_type='''dataset''' ), '''r''' ) ) _UpperCamelCase = {int(__snake_case ): v for k, v in idalabel.items()} _UpperCamelCase = idalabel _UpperCamelCase = {v: k for k, v in idalabel.items()} _UpperCamelCase = int(vit_name[-6:-4] ) _UpperCamelCase = int(vit_name[-3:] ) # size of the architecture if "deit" in vit_name: if vit_name[9:].startswith('''tiny''' ): _UpperCamelCase = 1_92 _UpperCamelCase = 7_68 _UpperCamelCase = 12 _UpperCamelCase = 3 elif vit_name[9:].startswith('''small''' ): _UpperCamelCase = 3_84 _UpperCamelCase = 15_36 _UpperCamelCase = 12 _UpperCamelCase = 6 else: pass else: if vit_name[4:].startswith('''small''' ): _UpperCamelCase = 7_68 _UpperCamelCase = 23_04 _UpperCamelCase = 8 _UpperCamelCase = 8 elif vit_name[4:].startswith('''base''' ): pass elif vit_name[4:].startswith('''large''' ): _UpperCamelCase = 10_24 _UpperCamelCase = 40_96 _UpperCamelCase = 24 _UpperCamelCase = 16 elif vit_name[4:].startswith('''huge''' ): _UpperCamelCase = 12_80 _UpperCamelCase = 51_20 _UpperCamelCase = 32 _UpperCamelCase = 16 # load original model from timm _UpperCamelCase = timm.create_model(__snake_case, pretrained=__snake_case ) timm_model.eval() # load state_dict of original model, remove and rename some keys _UpperCamelCase = timm_model.state_dict() if base_model: remove_classification_head_(__snake_case ) _UpperCamelCase = create_rename_keys(__snake_case, __snake_case ) for src, dest in rename_keys: rename_key(__snake_case, __snake_case, __snake_case ) read_in_q_k_v(__snake_case, __snake_case, __snake_case ) # load HuggingFace model if vit_name[-5:] == "in21k": _UpperCamelCase = ViTModel(__snake_case ).eval() else: _UpperCamelCase = ViTForImageClassification(__snake_case ).eval() model.load_state_dict(__snake_case ) # Check outputs on an image, prepared by ViTImageProcessor/DeiTImageProcessor if "deit" in vit_name: _UpperCamelCase = DeiTImageProcessor(size=config.image_size ) else: _UpperCamelCase = ViTImageProcessor(size=config.image_size ) _UpperCamelCase = image_processor(images=prepare_img(), return_tensors='''pt''' ) _UpperCamelCase = encoding['''pixel_values'''] _UpperCamelCase = model(__snake_case ) if base_model: _UpperCamelCase = timm_model.forward_features(__snake_case ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(__snake_case, outputs.pooler_output, atol=1e-3 ) else: _UpperCamelCase = timm_model(__snake_case ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__snake_case, outputs.logits, atol=1e-3 ) Path(__snake_case ).mkdir(exist_ok=__snake_case ) print(F'''Saving model {vit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(__snake_case ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(__snake_case ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( """--vit_name""", default="""vit_base_patch16_224""", type=str, help="""Name of the ViT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) _a = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path)
78
1
"""simple docstring""" from abc import ABC, abstractmethod from typing import Optional, Union from .. import Dataset, DatasetDict, Features, IterableDataset, IterableDatasetDict, NamedSplit from ..utils.typing import NestedDataStructureLike, PathLike class _UpperCAmelCase( lowerCamelCase ): def __init__( self , __a = None , __a = None , __a = None , __a = None , __a = False , __a = False , __a = None , **__a , ) -> List[str]: '''simple docstring''' _UpperCamelCase = path_or_paths _UpperCamelCase = split if split or isinstance(__a , __a) else '''train''' _UpperCamelCase = features _UpperCamelCase = cache_dir _UpperCamelCase = keep_in_memory _UpperCamelCase = streaming _UpperCamelCase = num_proc _UpperCamelCase = kwargs @abstractmethod def UpperCAmelCase ( self) -> Union[Dataset, DatasetDict, IterableDataset, IterableDatasetDict]: '''simple docstring''' pass class _UpperCAmelCase( lowerCamelCase ): def __init__( self , __a = None , __a = None , __a = False , __a = False , __a = None , **__a , ) -> Optional[int]: '''simple docstring''' _UpperCamelCase = features _UpperCamelCase = cache_dir _UpperCamelCase = keep_in_memory _UpperCamelCase = streaming _UpperCamelCase = num_proc _UpperCamelCase = kwargs @abstractmethod def UpperCAmelCase ( self) -> Union[Dataset, IterableDataset]: '''simple docstring''' pass
78
"""simple docstring""" import unittest from transformers import AlbertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForPreTraining, AlbertForQuestionAnswering, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertModel, ) from transformers.models.albert.modeling_albert import ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST class _UpperCAmelCase: def __init__( self , __a , __a=13 , __a=7 , __a=True , __a=True , __a=True , __a=True , __a=99 , __a=16 , __a=36 , __a=6 , __a=6 , __a=6 , __a=37 , __a="gelu" , __a=0.1 , __a=0.1 , __a=5_12 , __a=16 , __a=2 , __a=0.02 , __a=3 , __a=4 , __a=None , ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = seq_length _UpperCamelCase = is_training _UpperCamelCase = use_input_mask _UpperCamelCase = use_token_type_ids _UpperCamelCase = use_labels _UpperCamelCase = vocab_size _UpperCamelCase = embedding_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_hidden_groups _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = type_vocab_size _UpperCamelCase = type_sequence_label_size _UpperCamelCase = initializer_range _UpperCamelCase = num_labels _UpperCamelCase = num_choices _UpperCamelCase = scope def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) _UpperCamelCase = None if self.use_input_mask: _UpperCamelCase = random_attention_mask([self.batch_size, self.seq_length]) _UpperCamelCase = None if self.use_token_type_ids: _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size) _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) _UpperCamelCase = ids_tensor([self.batch_size] , self.num_choices) _UpperCamelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase ( self) -> Dict: '''simple docstring''' return AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , num_hidden_groups=self.num_hidden_groups , ) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Optional[int]: '''simple docstring''' _UpperCamelCase = AlbertModel(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , token_type_ids=__a) _UpperCamelCase = model(__a , token_type_ids=__a) _UpperCamelCase = model(__a) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = AlbertForPreTraining(config=__a) model.to(__a) model.eval() _UpperCamelCase = model( __a , attention_mask=__a , token_type_ids=__a , labels=__a , sentence_order_label=__a , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) self.parent.assertEqual(result.sop_logits.shape , (self.batch_size, config.num_labels)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Optional[int]: '''simple docstring''' _UpperCamelCase = AlbertForMaskedLM(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Any: '''simple docstring''' _UpperCamelCase = AlbertForQuestionAnswering(config=__a) model.to(__a) model.eval() _UpperCamelCase = model( __a , attention_mask=__a , token_type_ids=__a , start_positions=__a , end_positions=__a , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> List[Any]: '''simple docstring''' _UpperCamelCase = self.num_labels _UpperCamelCase = AlbertForSequenceClassification(__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> List[str]: '''simple docstring''' _UpperCamelCase = self.num_labels _UpperCamelCase = AlbertForTokenClassification(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.num_choices _UpperCamelCase = AlbertForMultipleChoice(config=__a) model.to(__a) model.eval() _UpperCamelCase = input_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() _UpperCamelCase = token_type_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() _UpperCamelCase = input_mask.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() _UpperCamelCase = model( __a , attention_mask=__a , token_type_ids=__a , labels=__a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices)) def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.prepare_config_and_inputs() ( ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ) = config_and_inputs _UpperCamelCase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class _UpperCAmelCase( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase__ = ( ( AlbertModel, AlbertForPreTraining, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertForQuestionAnswering, ) if is_torch_available() else () ) lowercase__ = ( { 'feature-extraction': AlbertModel, 'fill-mask': AlbertForMaskedLM, 'question-answering': AlbertForQuestionAnswering, 'text-classification': AlbertForSequenceClassification, 'token-classification': AlbertForTokenClassification, 'zero-shot': AlbertForSequenceClassification, } if is_torch_available() else {} ) lowercase__ = True def UpperCAmelCase ( self , __a , __a , __a=False) -> List[str]: '''simple docstring''' _UpperCamelCase = super()._prepare_for_class(__a , __a , return_labels=__a) if return_labels: if model_class in get_values(__a): _UpperCamelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=__a) _UpperCamelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__a) return inputs_dict def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = AlbertModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=__a , hidden_size=37) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*__a) def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__a) def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__a) def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__a) def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__a) def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _UpperCamelCase = type self.model_tester.create_and_check_model(*__a) @slow def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' for model_name in ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase = AlbertModel.from_pretrained(__a) self.assertIsNotNone(__a) @require_torch class _UpperCAmelCase( unittest.TestCase ): @slow def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = AlbertModel.from_pretrained('''albert-base-v2''') _UpperCamelCase = torch.tensor([[0, 3_45, 2_32, 3_28, 7_40, 1_40, 16_95, 69, 60_78, 15_88, 2]]) _UpperCamelCase = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]) with torch.no_grad(): _UpperCamelCase = model(__a , attention_mask=__a)[0] _UpperCamelCase = torch.Size((1, 11, 7_68)) self.assertEqual(output.shape , __a) _UpperCamelCase = torch.tensor( [[[-0.6513, 1.5035, -0.2766], [-0.6515, 1.5046, -0.2780], [-0.6512, 1.5049, -0.2784]]]) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , __a , atol=1e-4))
78
1
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import ( DiffusionPipeline, UnCLIPImageVariationPipeline, UnCLIPScheduler, UNetaDConditionModel, UNetaDModel, ) from diffusers.pipelines.unclip.text_proj import UnCLIPTextProjModel from diffusers.utils import floats_tensor, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, load_image, require_torch_gpu, skip_mps from ..pipeline_params import IMAGE_VARIATION_BATCH_PARAMS, IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class _UpperCAmelCase( lowerCamelCase , unittest.TestCase ): lowercase__ = UnCLIPImageVariationPipeline lowercase__ = IMAGE_VARIATION_PARAMS - {'height', 'width', 'guidance_scale'} lowercase__ = IMAGE_VARIATION_BATCH_PARAMS lowercase__ = [ 'generator', 'return_dict', 'decoder_num_inference_steps', 'super_res_num_inference_steps', ] lowercase__ = False @property def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' return 32 @property def UpperCAmelCase ( self) -> Dict: '''simple docstring''' return 32 @property def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' return self.time_input_dim @property def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' return self.time_input_dim * 4 @property def UpperCAmelCase ( self) -> Dict: '''simple docstring''' return 1_00 @property def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') return tokenizer @property def UpperCAmelCase ( self) -> str: '''simple docstring''' torch.manual_seed(0) _UpperCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) return CLIPTextModelWithProjection(__a) @property def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' torch.manual_seed(0) _UpperCamelCase = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , num_hidden_layers=5 , num_attention_heads=4 , image_size=32 , intermediate_size=37 , patch_size=1 , ) return CLIPVisionModelWithProjection(__a) @property def UpperCAmelCase ( self) -> Any: '''simple docstring''' torch.manual_seed(0) _UpperCamelCase = { '''clip_embeddings_dim''': self.text_embedder_hidden_size, '''time_embed_dim''': self.time_embed_dim, '''cross_attention_dim''': self.cross_attention_dim, } _UpperCamelCase = UnCLIPTextProjModel(**__a) return model @property def UpperCAmelCase ( self) -> Any: '''simple docstring''' torch.manual_seed(0) _UpperCamelCase = { '''sample_size''': 32, # RGB in channels '''in_channels''': 3, # Out channels is double in channels because predicts mean and variance '''out_channels''': 6, '''down_block_types''': ('''ResnetDownsampleBlock2D''', '''SimpleCrossAttnDownBlock2D'''), '''up_block_types''': ('''SimpleCrossAttnUpBlock2D''', '''ResnetUpsampleBlock2D'''), '''mid_block_type''': '''UNetMidBlock2DSimpleCrossAttn''', '''block_out_channels''': (self.block_out_channels_a, self.block_out_channels_a * 2), '''layers_per_block''': 1, '''cross_attention_dim''': self.cross_attention_dim, '''attention_head_dim''': 4, '''resnet_time_scale_shift''': '''scale_shift''', '''class_embed_type''': '''identity''', } _UpperCamelCase = UNetaDConditionModel(**__a) return model @property def UpperCAmelCase ( self) -> int: '''simple docstring''' return { "sample_size": 64, "layers_per_block": 1, "down_block_types": ("ResnetDownsampleBlock2D", "ResnetDownsampleBlock2D"), "up_block_types": ("ResnetUpsampleBlock2D", "ResnetUpsampleBlock2D"), "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "in_channels": 6, "out_channels": 3, } @property def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' torch.manual_seed(0) _UpperCamelCase = UNetaDModel(**self.dummy_super_res_kwargs) return model @property def UpperCAmelCase ( self) -> int: '''simple docstring''' # seeded differently to get different unet than `self.dummy_super_res_first` torch.manual_seed(1) _UpperCamelCase = UNetaDModel(**self.dummy_super_res_kwargs) return model def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = self.dummy_decoder _UpperCamelCase = self.dummy_text_proj _UpperCamelCase = self.dummy_text_encoder _UpperCamelCase = self.dummy_tokenizer _UpperCamelCase = self.dummy_super_res_first _UpperCamelCase = self.dummy_super_res_last _UpperCamelCase = UnCLIPScheduler( variance_type='''learned_range''' , prediction_type='''epsilon''' , num_train_timesteps=10_00 , ) _UpperCamelCase = UnCLIPScheduler( variance_type='''fixed_small_log''' , prediction_type='''epsilon''' , num_train_timesteps=10_00 , ) _UpperCamelCase = CLIPImageProcessor(crop_size=32 , size=32) _UpperCamelCase = self.dummy_image_encoder return { "decoder": decoder, "text_encoder": text_encoder, "tokenizer": tokenizer, "text_proj": text_proj, "feature_extractor": feature_extractor, "image_encoder": image_encoder, "super_res_first": super_res_first, "super_res_last": super_res_last, "decoder_scheduler": decoder_scheduler, "super_res_scheduler": super_res_scheduler, } def UpperCAmelCase ( self , __a , __a=0 , __a=True) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(__a)).to(__a) if str(__a).startswith('''mps'''): _UpperCamelCase = torch.manual_seed(__a) else: _UpperCamelCase = torch.Generator(device=__a).manual_seed(__a) if pil_image: _UpperCamelCase = input_image * 0.5 + 0.5 _UpperCamelCase = input_image.clamp(0 , 1) _UpperCamelCase = input_image.cpu().permute(0 , 2 , 3 , 1).float().numpy() _UpperCamelCase = DiffusionPipeline.numpy_to_pil(__a)[0] return { "image": input_image, "generator": generator, "decoder_num_inference_steps": 2, "super_res_num_inference_steps": 2, "output_type": "np", } def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = '''cpu''' _UpperCamelCase = self.get_dummy_components() _UpperCamelCase = self.pipeline_class(**__a) _UpperCamelCase = pipe.to(__a) pipe.set_progress_bar_config(disable=__a) _UpperCamelCase = self.get_dummy_inputs(__a , pil_image=__a) _UpperCamelCase = pipe(**__a) _UpperCamelCase = output.images _UpperCamelCase = self.get_dummy_inputs(__a , pil_image=__a) _UpperCamelCase = pipe( **__a , return_dict=__a , )[0] _UpperCamelCase = image[0, -3:, -3:, -1] _UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _UpperCamelCase = np.array( [ 0.9997, 0.0002, 0.9997, 0.9997, 0.9969, 0.0023, 0.9997, 0.9969, 0.9970, ]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = '''cpu''' _UpperCamelCase = self.get_dummy_components() _UpperCamelCase = self.pipeline_class(**__a) _UpperCamelCase = pipe.to(__a) pipe.set_progress_bar_config(disable=__a) _UpperCamelCase = self.get_dummy_inputs(__a , pil_image=__a) _UpperCamelCase = pipe(**__a) _UpperCamelCase = output.images _UpperCamelCase = self.get_dummy_inputs(__a , pil_image=__a) _UpperCamelCase = pipe( **__a , return_dict=__a , )[0] _UpperCamelCase = image[0, -3:, -3:, -1] _UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _UpperCamelCase = np.array([0.9997, 0.0003, 0.9997, 0.9997, 0.9970, 0.0024, 0.9997, 0.9971, 0.9971]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = '''cpu''' _UpperCamelCase = self.get_dummy_components() _UpperCamelCase = self.pipeline_class(**__a) _UpperCamelCase = pipe.to(__a) pipe.set_progress_bar_config(disable=__a) _UpperCamelCase = self.get_dummy_inputs(__a , pil_image=__a) _UpperCamelCase = [ pipeline_inputs['''image'''], pipeline_inputs['''image'''], ] _UpperCamelCase = pipe(**__a) _UpperCamelCase = output.images _UpperCamelCase = self.get_dummy_inputs(__a , pil_image=__a) _UpperCamelCase = [ tuple_pipeline_inputs['''image'''], tuple_pipeline_inputs['''image'''], ] _UpperCamelCase = pipe( **__a , return_dict=__a , )[0] _UpperCamelCase = image[0, -3:, -3:, -1] _UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (2, 64, 64, 3) _UpperCamelCase = np.array( [ 0.9997, 0.9989, 0.0008, 0.0021, 0.9960, 0.0018, 0.0014, 0.0002, 0.9933, ]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = torch.device('''cpu''') class _UpperCAmelCase: lowercase__ = 1 _UpperCamelCase = self.get_dummy_components() _UpperCamelCase = self.pipeline_class(**__a) _UpperCamelCase = pipe.to(__a) pipe.set_progress_bar_config(disable=__a) _UpperCamelCase = torch.Generator(device=__a).manual_seed(0) _UpperCamelCase = pipe.decoder.dtype _UpperCamelCase = 1 _UpperCamelCase = ( batch_size, pipe.decoder.config.in_channels, pipe.decoder.config.sample_size, pipe.decoder.config.sample_size, ) _UpperCamelCase = pipe.prepare_latents( __a , dtype=__a , device=__a , generator=__a , latents=__a , scheduler=DummyScheduler()) _UpperCamelCase = ( batch_size, pipe.super_res_first.config.in_channels // 2, pipe.super_res_first.config.sample_size, pipe.super_res_first.config.sample_size, ) _UpperCamelCase = pipe.prepare_latents( __a , dtype=__a , device=__a , generator=__a , latents=__a , scheduler=DummyScheduler()) _UpperCamelCase = self.get_dummy_inputs(__a , pil_image=__a) _UpperCamelCase = pipe( **__a , decoder_latents=__a , super_res_latents=__a).images _UpperCamelCase = self.get_dummy_inputs(__a , pil_image=__a) # Don't pass image, instead pass embedding _UpperCamelCase = pipeline_inputs.pop('''image''') _UpperCamelCase = pipe.image_encoder(__a).image_embeds _UpperCamelCase = pipe( **__a , decoder_latents=__a , super_res_latents=__a , image_embeddings=__a , ).images # make sure passing text embeddings manually is identical assert np.abs(img_out_a - img_out_a).max() < 1e-4 @skip_mps def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = torch_device == '''cpu''' # Check is relaxed because there is not a torch 2.0 sliced attention added kv processor _UpperCamelCase = 1e-2 self._test_attention_slicing_forward_pass( test_max_difference=__a , expected_max_diff=__a) @skip_mps def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = torch_device == '''cpu''' _UpperCamelCase = True _UpperCamelCase = [ '''decoder_num_inference_steps''', '''super_res_num_inference_steps''', ] self._test_inference_batch_single_identical( test_max_difference=__a , relax_max_difference=__a , additional_params_copy_to_batched_inputs=__a , ) def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = [ '''decoder_num_inference_steps''', '''super_res_num_inference_steps''', ] if torch_device == "mps": # TODO: MPS errors with larger batch sizes _UpperCamelCase = [2, 3] self._test_inference_batch_consistent( batch_sizes=__a , additional_params_copy_to_batched_inputs=__a , ) else: self._test_inference_batch_consistent( additional_params_copy_to_batched_inputs=__a) @skip_mps def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' return super().test_dict_tuple_outputs_equivalent() @skip_mps def UpperCAmelCase ( self) -> Dict: '''simple docstring''' return super().test_save_load_local() @skip_mps def UpperCAmelCase ( self) -> str: '''simple docstring''' return super().test_save_load_optional_components() @slow @require_torch_gpu class _UpperCAmelCase( unittest.TestCase ): def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/unclip/cat.png''') _UpperCamelCase = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/unclip/karlo_v1_alpha_cat_variation_fp16.npy''') _UpperCamelCase = UnCLIPImageVariationPipeline.from_pretrained( '''kakaobrain/karlo-v1-alpha-image-variations''' , torch_dtype=torch.floataa) _UpperCamelCase = pipeline.to(__a) pipeline.set_progress_bar_config(disable=__a) _UpperCamelCase = torch.Generator(device='''cpu''').manual_seed(0) _UpperCamelCase = pipeline( __a , generator=__a , output_type='''np''' , ) _UpperCamelCase = output.images[0] assert image.shape == (2_56, 2_56, 3) assert_mean_pixel_difference(__a , __a , 15)
78
"""simple docstring""" import os from typing import BinaryIO, Optional, Union import numpy as np import pyarrow.parquet as pq from .. import Audio, Dataset, Features, Image, NamedSplit, Value, config from ..features.features import FeatureType, _visit from ..formatting import query_table from ..packaged_modules import _PACKAGED_DATASETS_MODULES from ..packaged_modules.parquet.parquet import Parquet from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" _UpperCamelCase = np.inf def set_batch_size(__snake_case ) -> None: nonlocal batch_size if isinstance(__snake_case, __snake_case ): _UpperCamelCase = min(__snake_case, config.PARQUET_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS ) elif isinstance(__snake_case, __snake_case ): _UpperCamelCase = min(__snake_case, config.PARQUET_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS ) elif isinstance(__snake_case, __snake_case ) and feature.dtype == "binary": _UpperCamelCase = min(__snake_case, config.PARQUET_ROW_GROUP_SIZE_FOR_BINARY_DATASETS ) _visit(__snake_case, __snake_case ) return None if batch_size is np.inf else batch_size class _UpperCAmelCase( lowerCamelCase ): def __init__( self , __a , __a = None , __a = None , __a = None , __a = False , __a = False , __a = None , **__a , ) -> Dict: '''simple docstring''' super().__init__( __a , split=__a , features=__a , cache_dir=__a , keep_in_memory=__a , streaming=__a , num_proc=__a , **__a , ) _UpperCamelCase = path_or_paths if isinstance(__a , __a) else {self.split: path_or_paths} _UpperCamelCase = _PACKAGED_DATASETS_MODULES['''parquet'''][1] _UpperCamelCase = Parquet( cache_dir=__a , data_files=__a , features=__a , hash=__a , **__a , ) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' # Build iterable dataset if self.streaming: _UpperCamelCase = self.builder.as_streaming_dataset(split=self.split) # Build regular (map-style) dataset else: _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None self.builder.download_and_prepare( download_config=__a , download_mode=__a , verification_mode=__a , base_path=__a , num_proc=self.num_proc , ) _UpperCamelCase = self.builder.as_dataset( split=self.split , verification_mode=__a , in_memory=self.keep_in_memory) return dataset class _UpperCAmelCase: def __init__( self , __a , __a , __a = None , **__a , ) -> Dict: '''simple docstring''' _UpperCamelCase = dataset _UpperCamelCase = path_or_buf _UpperCamelCase = batch_size or get_writer_batch_size(dataset.features) _UpperCamelCase = parquet_writer_kwargs def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = self.batch_size if self.batch_size else config.DEFAULT_MAX_BATCH_SIZE if isinstance(self.path_or_buf , (str, bytes, os.PathLike)): with open(self.path_or_buf , '''wb+''') as buffer: _UpperCamelCase = self._write(file_obj=__a , batch_size=__a , **self.parquet_writer_kwargs) else: _UpperCamelCase = self._write(file_obj=self.path_or_buf , batch_size=__a , **self.parquet_writer_kwargs) return written def UpperCAmelCase ( self , __a , __a , **__a) -> int: '''simple docstring''' _UpperCamelCase = 0 _UpperCamelCase = parquet_writer_kwargs.pop('''path_or_buf''' , __a) _UpperCamelCase = self.dataset.features.arrow_schema _UpperCamelCase = pq.ParquetWriter(__a , schema=__a , **__a) for offset in logging.tqdm( range(0 , len(self.dataset) , __a) , unit='''ba''' , disable=not logging.is_progress_bar_enabled() , desc='''Creating parquet from Arrow format''' , ): _UpperCamelCase = query_table( table=self.dataset._data , key=slice(__a , offset + batch_size) , indices=self.dataset._indices if self.dataset._indices is not None else None , ) writer.write_table(__a) written += batch.nbytes writer.close() return written
78
1
"""simple docstring""" import functools def lowerCamelCase__ ( __snake_case, __snake_case ) -> int: """simple docstring""" _UpperCamelCase = len(__snake_case ) _UpperCamelCase = len(__snake_case ) @functools.cache def min_distance(__snake_case, __snake_case ) -> int: # if first word index is overflow - delete all from the second word if indexa >= len_worda: return len_worda - indexa # if second word index is overflow - delete all from the first word if indexa >= len_worda: return len_worda - indexa _UpperCamelCase = int(worda[indexa] != worda[indexa] ) # current letters not identical return min( 1 + min_distance(indexa + 1, __snake_case ), 1 + min_distance(__snake_case, indexa + 1 ), diff + min_distance(indexa + 1, indexa + 1 ), ) return min_distance(0, 0 ) if __name__ == "__main__": import doctest doctest.testmod()
78
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class _UpperCAmelCase( unittest.TestCase ): def __init__( self , __a , __a=7 , __a=3 , __a=18 , __a=30 , __a=4_00 , __a=True , __a=None , __a=True , __a=None , __a=True , ) -> int: '''simple docstring''' _UpperCamelCase = size if size is not None else {'''shortest_edge''': 20} _UpperCamelCase = crop_size if crop_size is not None else {'''height''': 18, '''width''': 18} _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = num_channels _UpperCamelCase = image_size _UpperCamelCase = min_resolution _UpperCamelCase = max_resolution _UpperCamelCase = do_resize _UpperCamelCase = size _UpperCamelCase = do_center_crop _UpperCamelCase = crop_size _UpperCamelCase = do_flip_channel_order def UpperCAmelCase ( self) -> str: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_flip_channel_order": self.do_flip_channel_order, } @require_torch @require_vision class _UpperCAmelCase( lowerCamelCase , unittest.TestCase ): lowercase__ = MobileViTImageProcessor if is_vision_available() else None def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = MobileViTImageProcessingTester(self) @property def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(__a , '''do_resize''')) self.assertTrue(hasattr(__a , '''size''')) self.assertTrue(hasattr(__a , '''do_center_crop''')) self.assertTrue(hasattr(__a , '''center_crop''')) self.assertTrue(hasattr(__a , '''do_flip_channel_order''')) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size , {'''shortest_edge''': 20}) self.assertEqual(image_processor.crop_size , {'''height''': 18, '''width''': 18}) _UpperCamelCase = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84) self.assertEqual(image_processor.size , {'''shortest_edge''': 42}) self.assertEqual(image_processor.crop_size , {'''height''': 84, '''width''': 84}) def UpperCAmelCase ( self) -> Dict: '''simple docstring''' pass def UpperCAmelCase ( self) -> str: '''simple docstring''' # Initialize image_processing _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random PIL images _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__a) for image in image_inputs: self.assertIsInstance(__a , Image.Image) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(__a , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' # Initialize image_processing _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__a , numpify=__a) for image in image_inputs: self.assertIsInstance(__a , np.ndarray) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(__a , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def UpperCAmelCase ( self) -> int: '''simple docstring''' # Initialize image_processing _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__a , torchify=__a) for image in image_inputs: self.assertIsInstance(__a , torch.Tensor) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(__a , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
78
1
"""simple docstring""" import unittest from huggingface_hub import hf_hub_download from transformers import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEFeatureExtractor from transformers.pipelines import VideoClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_decord, require_tf, require_torch, require_torch_or_tf, require_vision, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf @require_vision @require_decord class _UpperCAmelCase( unittest.TestCase ): lowercase__ = MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING def UpperCAmelCase ( self , __a , __a , __a) -> List[str]: '''simple docstring''' _UpperCamelCase = hf_hub_download( repo_id='''nateraw/video-demo''' , filename='''archery.mp4''' , repo_type='''dataset''') _UpperCamelCase = VideoClassificationPipeline(model=__a , image_processor=__a , top_k=2) _UpperCamelCase = [ example_video_filepath, '''https://huggingface.co/datasets/nateraw/video-demo/resolve/main/archery.mp4''', ] return video_classifier, examples def UpperCAmelCase ( self , __a , __a) -> int: '''simple docstring''' for example in examples: _UpperCamelCase = video_classifier(__a) self.assertEqual( __a , [ {'''score''': ANY(__a), '''label''': ANY(__a)}, {'''score''': ANY(__a), '''label''': ANY(__a)}, ] , ) @require_torch def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = '''hf-internal-testing/tiny-random-VideoMAEForVideoClassification''' _UpperCamelCase = VideoMAEFeatureExtractor( size={'''shortest_edge''': 10} , crop_size={'''height''': 10, '''width''': 10}) _UpperCamelCase = pipeline( '''video-classification''' , model=__a , feature_extractor=__a , frame_sampling_rate=4) _UpperCamelCase = hf_hub_download(repo_id='''nateraw/video-demo''' , filename='''archery.mp4''' , repo_type='''dataset''') _UpperCamelCase = video_classifier(__a , top_k=2) self.assertEqual( nested_simplify(__a , decimals=4) , [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}] , ) _UpperCamelCase = video_classifier( [ video_file_path, video_file_path, ] , top_k=2 , ) self.assertEqual( nested_simplify(__a , decimals=4) , [ [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}], [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}], ] , ) @require_tf def UpperCAmelCase ( self) -> int: '''simple docstring''' pass
78
"""simple docstring""" import warnings from typing import List import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import is_flax_available, is_tf_available, is_torch_available class _UpperCAmelCase( lowerCamelCase ): lowercase__ = ['image_processor', 'tokenizer'] lowercase__ = 'OwlViTImageProcessor' lowercase__ = ('CLIPTokenizer', 'CLIPTokenizerFast') def __init__( self , __a=None , __a=None , **__a) -> List[Any]: '''simple docstring''' _UpperCamelCase = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , __a , ) _UpperCamelCase = kwargs.pop('''feature_extractor''') _UpperCamelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''') if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''') super().__init__(__a , __a) def __call__( self , __a=None , __a=None , __a=None , __a="max_length" , __a="np" , **__a) -> List[str]: '''simple docstring''' if text is None and query_images is None and images is None: raise ValueError( '''You have to specify at least one text or query image or image. All three cannot be none.''') if text is not None: if isinstance(__a , __a) or (isinstance(__a , __a) and not isinstance(text[0] , __a)): _UpperCamelCase = [self.tokenizer(__a , padding=__a , return_tensors=__a , **__a)] elif isinstance(__a , __a) and isinstance(text[0] , __a): _UpperCamelCase = [] # Maximum number of queries across batch _UpperCamelCase = max([len(__a) for t in text]) # Pad all batch samples to max number of text queries for t in text: if len(__a) != max_num_queries: _UpperCamelCase = t + [''' '''] * (max_num_queries - len(__a)) _UpperCamelCase = self.tokenizer(__a , padding=__a , return_tensors=__a , **__a) encodings.append(__a) else: raise TypeError('''Input text should be a string, a list of strings or a nested list of strings''') if return_tensors == "np": _UpperCamelCase = np.concatenate([encoding['''input_ids'''] for encoding in encodings] , axis=0) _UpperCamelCase = np.concatenate([encoding['''attention_mask'''] for encoding in encodings] , axis=0) elif return_tensors == "jax" and is_flax_available(): import jax.numpy as jnp _UpperCamelCase = jnp.concatenate([encoding['''input_ids'''] for encoding in encodings] , axis=0) _UpperCamelCase = jnp.concatenate([encoding['''attention_mask'''] for encoding in encodings] , axis=0) elif return_tensors == "pt" and is_torch_available(): import torch _UpperCamelCase = torch.cat([encoding['''input_ids'''] for encoding in encodings] , dim=0) _UpperCamelCase = torch.cat([encoding['''attention_mask'''] for encoding in encodings] , dim=0) elif return_tensors == "tf" and is_tf_available(): import tensorflow as tf _UpperCamelCase = tf.stack([encoding['''input_ids'''] for encoding in encodings] , axis=0) _UpperCamelCase = tf.stack([encoding['''attention_mask'''] for encoding in encodings] , axis=0) else: raise ValueError('''Target return tensor type could not be returned''') _UpperCamelCase = BatchEncoding() _UpperCamelCase = input_ids _UpperCamelCase = attention_mask if query_images is not None: _UpperCamelCase = BatchEncoding() _UpperCamelCase = self.image_processor( __a , return_tensors=__a , **__a).pixel_values _UpperCamelCase = query_pixel_values if images is not None: _UpperCamelCase = self.image_processor(__a , return_tensors=__a , **__a) if text is not None and images is not None: _UpperCamelCase = image_features.pixel_values return encoding elif query_images is not None and images is not None: _UpperCamelCase = image_features.pixel_values return encoding elif text is not None or query_images is not None: return encoding else: return BatchEncoding(data=dict(**__a) , tensor_type=__a) def UpperCAmelCase ( self , *__a , **__a) -> str: '''simple docstring''' return self.image_processor.post_process(*__a , **__a) def UpperCAmelCase ( self , *__a , **__a) -> Dict: '''simple docstring''' return self.image_processor.post_process_object_detection(*__a , **__a) def UpperCAmelCase ( self , *__a , **__a) -> Optional[Any]: '''simple docstring''' return self.image_processor.post_process_image_guided_detection(*__a , **__a) def UpperCAmelCase ( self , *__a , **__a) -> Optional[int]: '''simple docstring''' return self.tokenizer.batch_decode(*__a , **__a) def UpperCAmelCase ( self , *__a , **__a) -> Optional[Any]: '''simple docstring''' return self.tokenizer.decode(*__a , **__a) @property def UpperCAmelCase ( self) -> str: '''simple docstring''' warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , __a , ) return self.image_processor_class @property def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , __a , ) return self.image_processor
78
1
"""simple docstring""" from importlib import import_module from .logging import get_logger _a = get_logger(__name__) class _UpperCAmelCase: def __init__( self , __a , __a=None) -> Dict: '''simple docstring''' _UpperCamelCase = attrs or [] if module is not None: for key in module.__dict__: if key in attrs or not key.startswith('''__'''): setattr(self , __a , getattr(__a , __a)) _UpperCamelCase = module._original_module if isinstance(__a , _PatchedModuleObj) else module class _UpperCAmelCase: lowercase__ = [] def __init__( self , __a , __a , __a , __a=None) -> List[str]: '''simple docstring''' _UpperCamelCase = obj _UpperCamelCase = target _UpperCamelCase = new _UpperCamelCase = target.split('''.''')[0] _UpperCamelCase = {} _UpperCamelCase = attrs or [] def __enter__( self) -> int: '''simple docstring''' *_UpperCamelCase , _UpperCamelCase = self.target.split('''.''') # Patch modules: # it's used to patch attributes of submodules like "os.path.join"; # in this case we need to patch "os" and "os.path" for i in range(len(__a)): try: _UpperCamelCase = import_module('''.'''.join(submodules[: i + 1])) except ModuleNotFoundError: continue # We iterate over all the globals in self.obj in case we find "os" or "os.path" for attr in self.obj.__dir__(): _UpperCamelCase = getattr(self.obj , __a) # We don't check for the name of the global, but rather if its value *is* "os" or "os.path". # This allows to patch renamed modules like "from os import path as ospath". if obj_attr is submodule or ( (isinstance(__a , _PatchedModuleObj) and obj_attr._original_module is submodule) ): _UpperCamelCase = obj_attr # patch at top level setattr(self.obj , __a , _PatchedModuleObj(__a , attrs=self.attrs)) _UpperCamelCase = getattr(self.obj , __a) # construct lower levels patches for key in submodules[i + 1 :]: setattr(__a , __a , _PatchedModuleObj(getattr(__a , __a , __a) , attrs=self.attrs)) _UpperCamelCase = getattr(__a , __a) # finally set the target attribute setattr(__a , __a , self.new) # Patch attribute itself: # it's used for builtins like "open", # and also to patch "os.path.join" we may also need to patch "join" # itself if it was imported as "from os.path import join". if submodules: # if it's an attribute of a submodule like "os.path.join" try: _UpperCamelCase = getattr(import_module('''.'''.join(__a)) , __a) except (AttributeError, ModuleNotFoundError): return # We iterate over all the globals in self.obj in case we find "os.path.join" for attr in self.obj.__dir__(): # We don't check for the name of the global, but rather if its value *is* "os.path.join". # This allows to patch renamed attributes like "from os.path import join as pjoin". if getattr(self.obj , __a) is attr_value: _UpperCamelCase = getattr(self.obj , __a) setattr(self.obj , __a , self.new) elif target_attr in globals()["__builtins__"]: # if it'a s builtin like "open" _UpperCamelCase = globals()['''__builtins__'''][target_attr] setattr(self.obj , __a , self.new) else: raise RuntimeError(F'''Tried to patch attribute {target_attr} instead of a submodule.''') def __exit__( self , *__a) -> Tuple: '''simple docstring''' for attr in list(self.original): setattr(self.obj , __a , self.original.pop(__a)) def UpperCAmelCase ( self) -> Dict: '''simple docstring''' self.__enter__() self._active_patches.append(self) def UpperCAmelCase ( self) -> str: '''simple docstring''' try: self._active_patches.remove(self) except ValueError: # If the patch hasn't been started this will fail return None return self.__exit__()
78
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) _a = { """configuration_perceiver""": ["""PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """PerceiverConfig""", """PerceiverOnnxConfig"""], """tokenization_perceiver""": ["""PerceiverTokenizer"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ["""PerceiverFeatureExtractor"""] _a = ["""PerceiverImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ """PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST""", """PerceiverForImageClassificationConvProcessing""", """PerceiverForImageClassificationFourier""", """PerceiverForImageClassificationLearned""", """PerceiverForMaskedLM""", """PerceiverForMultimodalAutoencoding""", """PerceiverForOpticalFlow""", """PerceiverForSequenceClassification""", """PerceiverLayer""", """PerceiverModel""", """PerceiverPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys _a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
78
1
"""simple docstring""" from typing import List, Optional, Union import torch from transformers import ( XLMRobertaTokenizer, ) from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) from .text_encoder import MultilingualCLIP _a = logging.get_logger(__name__) # pylint: disable=invalid-name _a = """ Examples: ```py >>> from diffusers import KandinskyPipeline, KandinskyPriorPipeline >>> import torch >>> pipe_prior = KandinskyPriorPipeline.from_pretrained(\"kandinsky-community/Kandinsky-2-1-prior\") >>> pipe_prior.to(\"cuda\") >>> prompt = \"red cat, 4k photo\" >>> out = pipe_prior(prompt) >>> image_emb = out.image_embeds >>> negative_image_emb = out.negative_image_embeds >>> pipe = KandinskyPipeline.from_pretrained(\"kandinsky-community/kandinsky-2-1\") >>> pipe.to(\"cuda\") >>> image = pipe( ... prompt, ... image_embeds=image_emb, ... negative_image_embeds=negative_image_emb, ... height=768, ... width=768, ... num_inference_steps=100, ... ).images >>> image[0].save(\"cat.png\") ``` """ def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case=8 ) -> List[str]: """simple docstring""" _UpperCamelCase = h // scale_factor**2 if h % scale_factor**2 != 0: new_h += 1 _UpperCamelCase = w // scale_factor**2 if w % scale_factor**2 != 0: new_w += 1 return new_h * scale_factor, new_w * scale_factor class _UpperCAmelCase( lowerCamelCase ): def __init__( self , __a , __a , __a , __a , __a , ) -> str: '''simple docstring''' super().__init__() self.register_modules( text_encoder=__a , tokenizer=__a , unet=__a , scheduler=__a , movq=__a , ) _UpperCamelCase = 2 ** (len(self.movq.config.block_out_channels) - 1) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a) -> Optional[Any]: '''simple docstring''' if latents is None: _UpperCamelCase = randn_tensor(__a , generator=__a , device=__a , dtype=__a) else: if latents.shape != shape: raise ValueError(F'''Unexpected latents shape, got {latents.shape}, expected {shape}''') _UpperCamelCase = latents.to(__a) _UpperCamelCase = latents * scheduler.init_noise_sigma return latents def UpperCAmelCase ( self , __a , __a , __a , __a , __a=None , ) -> List[Any]: '''simple docstring''' _UpperCamelCase = len(__a) if isinstance(__a , __a) else 1 # get prompt text embeddings _UpperCamelCase = self.tokenizer( __a , padding='''max_length''' , truncation=__a , max_length=77 , return_attention_mask=__a , add_special_tokens=__a , return_tensors='''pt''' , ) _UpperCamelCase = text_inputs.input_ids _UpperCamelCase = self.tokenizer(__a , padding='''longest''' , return_tensors='''pt''').input_ids if untruncated_ids.shape[-1] >= text_input_ids.shape[-1] and not torch.equal(__a , __a): _UpperCamelCase = self.tokenizer.batch_decode(untruncated_ids[:, self.tokenizer.model_max_length - 1 : -1]) logger.warning( '''The following part of your input was truncated because CLIP can only handle sequences up to''' F''' {self.tokenizer.model_max_length} tokens: {removed_text}''') _UpperCamelCase = text_input_ids.to(__a) _UpperCamelCase = text_inputs.attention_mask.to(__a) _UpperCamelCase , _UpperCamelCase = self.text_encoder( input_ids=__a , attention_mask=__a) _UpperCamelCase = prompt_embeds.repeat_interleave(__a , dim=0) _UpperCamelCase = text_encoder_hidden_states.repeat_interleave(__a , dim=0) _UpperCamelCase = text_mask.repeat_interleave(__a , dim=0) if do_classifier_free_guidance: _UpperCamelCase = 42 if negative_prompt is None: _UpperCamelCase = [''''''] * batch_size elif type(__a) is not type(__a): raise TypeError( F'''`negative_prompt` should be the same type to `prompt`, but got {type(__a)} !=''' F''' {type(__a)}.''') elif isinstance(__a , __a): _UpperCamelCase = [negative_prompt] elif batch_size != len(__a): raise ValueError( F'''`negative_prompt`: {negative_prompt} has batch size {len(__a)}, but `prompt`:''' F''' {prompt} has batch size {batch_size}. Please make sure that passed `negative_prompt` matches''' ''' the batch size of `prompt`.''') else: _UpperCamelCase = negative_prompt _UpperCamelCase = self.tokenizer( __a , padding='''max_length''' , max_length=77 , truncation=__a , return_attention_mask=__a , add_special_tokens=__a , return_tensors='''pt''' , ) _UpperCamelCase = uncond_input.input_ids.to(__a) _UpperCamelCase = uncond_input.attention_mask.to(__a) _UpperCamelCase , _UpperCamelCase = self.text_encoder( input_ids=__a , attention_mask=__a) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method _UpperCamelCase = negative_prompt_embeds.shape[1] _UpperCamelCase = negative_prompt_embeds.repeat(1 , __a) _UpperCamelCase = negative_prompt_embeds.view(batch_size * num_images_per_prompt , __a) _UpperCamelCase = uncond_text_encoder_hidden_states.shape[1] _UpperCamelCase = uncond_text_encoder_hidden_states.repeat(1 , __a , 1) _UpperCamelCase = uncond_text_encoder_hidden_states.view( batch_size * num_images_per_prompt , __a , -1) _UpperCamelCase = uncond_text_mask.repeat_interleave(__a , dim=0) # done duplicates # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes _UpperCamelCase = torch.cat([negative_prompt_embeds, prompt_embeds]) _UpperCamelCase = torch.cat([uncond_text_encoder_hidden_states, text_encoder_hidden_states]) _UpperCamelCase = torch.cat([uncond_text_mask, text_mask]) return prompt_embeds, text_encoder_hidden_states, text_mask def UpperCAmelCase ( self , __a=0) -> Optional[int]: '''simple docstring''' if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('''Please install accelerate via `pip install accelerate`''') _UpperCamelCase = torch.device(F'''cuda:{gpu_id}''') _UpperCamelCase = [ self.unet, self.text_encoder, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(__a , __a) def UpperCAmelCase ( self , __a=0) -> Optional[int]: '''simple docstring''' if is_accelerate_available() and is_accelerate_version('''>=''' , '''0.17.0.dev0'''): from accelerate import cpu_offload_with_hook else: raise ImportError('''`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.''') _UpperCamelCase = torch.device(F'''cuda:{gpu_id}''') if self.device.type != "cpu": self.to('''cpu''' , silence_dtype_warnings=__a) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) _UpperCamelCase = None for cpu_offloaded_model in [self.text_encoder, self.unet, self.movq]: _UpperCamelCase , _UpperCamelCase = cpu_offload_with_hook(__a , __a , prev_module_hook=__a) if self.safety_checker is not None: _UpperCamelCase , _UpperCamelCase = cpu_offload_with_hook(self.safety_checker , __a , prev_module_hook=__a) # We'll offload the last model manually. _UpperCamelCase = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' if not hasattr(self.unet , '''_hf_hook'''): return self.device for module in self.unet.modules(): if ( hasattr(__a , '''_hf_hook''') and hasattr(module._hf_hook , '''execution_device''') and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device) return self.device @torch.no_grad() @replace_example_docstring(__a) def __call__( self , __a , __a , __a , __a = None , __a = 5_12 , __a = 5_12 , __a = 1_00 , __a = 4.0 , __a = 1 , __a = None , __a = None , __a = "pil" , __a = True , ) -> Optional[int]: '''simple docstring''' if isinstance(__a , __a): _UpperCamelCase = 1 elif isinstance(__a , __a): _UpperCamelCase = len(__a) else: raise ValueError(F'''`prompt` has to be of type `str` or `list` but is {type(__a)}''') _UpperCamelCase = self._execution_device _UpperCamelCase = batch_size * num_images_per_prompt _UpperCamelCase = guidance_scale > 1.0 _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = self._encode_prompt( __a , __a , __a , __a , __a) if isinstance(__a , __a): _UpperCamelCase = torch.cat(__a , dim=0) if isinstance(__a , __a): _UpperCamelCase = torch.cat(__a , dim=0) if do_classifier_free_guidance: _UpperCamelCase = image_embeds.repeat_interleave(__a , dim=0) _UpperCamelCase = negative_image_embeds.repeat_interleave(__a , dim=0) _UpperCamelCase = torch.cat([negative_image_embeds, image_embeds] , dim=0).to( dtype=prompt_embeds.dtype , device=__a) self.scheduler.set_timesteps(__a , device=__a) _UpperCamelCase = self.scheduler.timesteps _UpperCamelCase = self.unet.config.in_channels _UpperCamelCase , _UpperCamelCase = get_new_h_w(__a , __a , self.movq_scale_factor) # create initial latent _UpperCamelCase = self.prepare_latents( (batch_size, num_channels_latents, height, width) , text_encoder_hidden_states.dtype , __a , __a , __a , self.scheduler , ) for i, t in enumerate(self.progress_bar(__a)): # expand the latents if we are doing classifier free guidance _UpperCamelCase = torch.cat([latents] * 2) if do_classifier_free_guidance else latents _UpperCamelCase = {'''text_embeds''': prompt_embeds, '''image_embeds''': image_embeds} _UpperCamelCase = self.unet( sample=__a , timestep=__a , encoder_hidden_states=__a , added_cond_kwargs=__a , return_dict=__a , )[0] if do_classifier_free_guidance: _UpperCamelCase , _UpperCamelCase = noise_pred.split(latents.shape[1] , dim=1) _UpperCamelCase , _UpperCamelCase = noise_pred.chunk(2) _UpperCamelCase , _UpperCamelCase = variance_pred.chunk(2) _UpperCamelCase = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) _UpperCamelCase = torch.cat([noise_pred, variance_pred_text] , dim=1) if not ( hasattr(self.scheduler.config , '''variance_type''') and self.scheduler.config.variance_type in ["learned", "learned_range"] ): _UpperCamelCase , _UpperCamelCase = noise_pred.split(latents.shape[1] , dim=1) # compute the previous noisy sample x_t -> x_t-1 _UpperCamelCase = self.scheduler.step( __a , __a , __a , generator=__a , ).prev_sample # post-processing _UpperCamelCase = self.movq.decode(__a , force_not_quantize=__a)['''sample'''] if output_type not in ["pt", "np", "pil"]: raise ValueError(F'''Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}''') if output_type in ["np", "pil"]: _UpperCamelCase = image * 0.5 + 0.5 _UpperCamelCase = image.clamp(0 , 1) _UpperCamelCase = image.cpu().permute(0 , 2 , 3 , 1).float().numpy() if output_type == "pil": _UpperCamelCase = self.numpy_to_pil(__a) if not return_dict: return (image,) return ImagePipelineOutput(images=__a)
78
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class _UpperCAmelCase: def __init__( self , __a , __a=13 , __a=2 , __a=24 , __a=16 , __a=True , __a=True , __a=32 , __a=5 , __a=4 , __a=37 , __a="gelu" , __a=0.1 , __a=0.1 , __a=10 , __a=0.02 , __a=None , __a=2 , __a=2 , ) -> List[str]: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = patch_size _UpperCamelCase = max_length _UpperCamelCase = num_mel_bins _UpperCamelCase = is_training _UpperCamelCase = use_labels _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = type_sequence_label_size _UpperCamelCase = initializer_range _UpperCamelCase = scope _UpperCamelCase = frequency_stride _UpperCamelCase = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) _UpperCamelCase = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 _UpperCamelCase = (self.max_length - self.patch_size) // self.time_stride + 1 _UpperCamelCase = frequency_out_dimension * time_out_dimension _UpperCamelCase = num_patches + 2 def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins]) _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size) _UpperCamelCase = self.get_config() return config, input_values, labels def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__a , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def UpperCAmelCase ( self , __a , __a , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = ASTModel(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.prepare_config_and_inputs() ( ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ) = config_and_inputs _UpperCamelCase = {'''input_values''': input_values} return config, inputs_dict @require_torch class _UpperCAmelCase( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase__ = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) lowercase__ = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) lowercase__ = False lowercase__ = False lowercase__ = False lowercase__ = False def UpperCAmelCase ( self , __a , __a , __a , __a , __a) -> Optional[Any]: '''simple docstring''' if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = ASTModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=__a , has_text_modality=__a , hidden_size=37) def UpperCAmelCase ( self) -> int: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='''AST does not use inputs_embeds''') def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' pass def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = model_class(__a) self.assertIsInstance(model.get_input_embeddings() , (nn.Module)) _UpperCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__a , nn.Linear)) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = model_class(__a) _UpperCamelCase = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCamelCase = [*signature.parameters.keys()] _UpperCamelCase = ['''input_values'''] self.assertListEqual(arg_names[:1] , __a) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a) @slow def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase = ASTModel.from_pretrained(__a) self.assertIsNotNone(__a) def lowerCamelCase__ ( ) -> List[str]: """simple docstring""" _UpperCamelCase = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''', filename='''sample_audio.flac''', repo_type='''dataset''' ) _UpperCamelCase , _UpperCamelCase = torchaudio.load(__snake_case ) return audio, sampling_rate @require_torch @require_torchaudio class _UpperCAmelCase( unittest.TestCase ): @cached_property def UpperCAmelCase ( self) -> Dict: '''simple docstring''' return ( ASTFeatureExtractor.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''') if is_torchaudio_available() else None ) @slow def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.default_feature_extractor _UpperCamelCase = ASTForAudioClassification.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''').to(__a) _UpperCamelCase = self.default_feature_extractor _UpperCamelCase , _UpperCamelCase = prepare_audio() _UpperCamelCase = audio.squeeze().numpy() _UpperCamelCase = feature_extractor(__a , sampling_rate=__a , return_tensors='''pt''').to(__a) # forward pass with torch.no_grad(): _UpperCamelCase = model(**__a) # verify the logits _UpperCamelCase = torch.Size((1, 5_27)) self.assertEqual(outputs.logits.shape , __a) _UpperCamelCase = torch.tensor([-0.8760, -7.0042, -8.6602]).to(__a) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __a , atol=1e-4))
78
1
"""simple docstring""" import os import random import sys from . import cryptomath_module as cryptoMath # noqa: N812 from . import rabin_miller as rabinMiller # noqa: N812 def lowerCamelCase__ ( ) -> None: """simple docstring""" print('''Making key files...''' ) make_key_files('''rsa''', 10_24 ) print('''Key files generation successful.''' ) def lowerCamelCase__ ( __snake_case ) -> tuple[tuple[int, int], tuple[int, int]]: """simple docstring""" print('''Generating prime p...''' ) _UpperCamelCase = rabinMiller.generate_large_prime(__snake_case ) print('''Generating prime q...''' ) _UpperCamelCase = rabinMiller.generate_large_prime(__snake_case ) _UpperCamelCase = p * q print('''Generating e that is relatively prime to (p - 1) * (q - 1)...''' ) while True: _UpperCamelCase = random.randrange(2 ** (key_size - 1), 2 ** (key_size) ) if cryptoMath.gcd(__snake_case, (p - 1) * (q - 1) ) == 1: break print('''Calculating d that is mod inverse of e...''' ) _UpperCamelCase = cryptoMath.find_mod_inverse(__snake_case, (p - 1) * (q - 1) ) _UpperCamelCase = (n, e) _UpperCamelCase = (n, d) return (public_key, private_key) def lowerCamelCase__ ( __snake_case, __snake_case ) -> None: """simple docstring""" if os.path.exists(F'''{name}_pubkey.txt''' ) or os.path.exists(F'''{name}_privkey.txt''' ): print('''\nWARNING:''' ) print( F'''"{name}_pubkey.txt" or "{name}_privkey.txt" already exists. \n''' '''Use a different name or delete these files and re-run this program.''' ) sys.exit() _UpperCamelCase , _UpperCamelCase = generate_key(__snake_case ) print(F'''\nWriting public key to file {name}_pubkey.txt...''' ) with open(F'''{name}_pubkey.txt''', '''w''' ) as out_file: out_file.write(F'''{key_size},{public_key[0]},{public_key[1]}''' ) print(F'''Writing private key to file {name}_privkey.txt...''' ) with open(F'''{name}_privkey.txt''', '''w''' ) as out_file: out_file.write(F'''{key_size},{private_key[0]},{private_key[1]}''' ) if __name__ == "__main__": main()
78
"""simple docstring""" def lowerCamelCase__ ( ) -> list[list[int]]: """simple docstring""" return [list(range(10_00 - i, -10_00 - i, -1 ) ) for i in range(10_00 )] _a = generate_large_matrix() _a = ( [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]], [[3, 2], [1, 0]], [[7, 7, 6]], [[7, 7, 6], [-1, -2, -3]], grid, ) def lowerCamelCase__ ( __snake_case ) -> None: """simple docstring""" assert all(row == sorted(__snake_case, reverse=__snake_case ) for row in grid ) assert all(list(__snake_case ) == sorted(__snake_case, reverse=__snake_case ) for col in zip(*__snake_case ) ) def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" _UpperCamelCase = 0 _UpperCamelCase = len(__snake_case ) - 1 # Edge cases such as no values or all numbers are negative. if not array or array[0] < 0: return 0 while right + 1 > left: _UpperCamelCase = (left + right) // 2 _UpperCamelCase = array[mid] # Num must be negative and the index must be greater than or equal to 0. if num < 0 and array[mid - 1] >= 0: return mid if num >= 0: _UpperCamelCase = mid + 1 else: _UpperCamelCase = mid - 1 # No negative numbers so return the last index of the array + 1 which is the length. return len(__snake_case ) def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" _UpperCamelCase = 0 _UpperCamelCase = len(grid[0] ) for i in range(len(__snake_case ) ): _UpperCamelCase = find_negative_index(grid[i][:bound] ) total += bound return (len(__snake_case ) * len(grid[0] )) - total def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" return len([number for row in grid for number in row if number < 0] ) def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" _UpperCamelCase = 0 for row in grid: for i, number in enumerate(__snake_case ): if number < 0: total += len(__snake_case ) - i break return total def lowerCamelCase__ ( ) -> None: """simple docstring""" from timeit import timeit print('''Running benchmarks''' ) _UpperCamelCase = ( '''from __main__ import count_negatives_binary_search, ''' '''count_negatives_brute_force, count_negatives_brute_force_with_break, grid''' ) for func in ( "count_negatives_binary_search", # took 0.7727 seconds "count_negatives_brute_force_with_break", # took 4.6505 seconds "count_negatives_brute_force", # took 12.8160 seconds ): _UpperCamelCase = timeit(F'''{func}(grid=grid)''', setup=__snake_case, number=5_00 ) print(F'''{func}() took {time:0.4f} seconds''' ) if __name__ == "__main__": import doctest doctest.testmod() benchmark()
78
1
"""simple docstring""" def lowerCamelCase__ ( __snake_case ) -> float: """simple docstring""" _UpperCamelCase = 0 while len(__snake_case ) > 1: _UpperCamelCase = 0 # Consider two files with minimum cost to be merged for _ in range(2 ): _UpperCamelCase = files.index(min(__snake_case ) ) temp += files[min_index] files.pop(__snake_case ) files.append(__snake_case ) optimal_merge_cost += temp return optimal_merge_cost if __name__ == "__main__": import doctest doctest.testmod()
78
"""simple docstring""" import copy import re class _UpperCAmelCase: lowercase__ = 'hp' lowercase__ = {} lowercase__ = None @classmethod def UpperCAmelCase ( cls , __a , __a) -> Dict: '''simple docstring''' _UpperCamelCase = prefix _UpperCamelCase = defaults cls.build_naming_info() @staticmethod def UpperCAmelCase ( __a , __a) -> Union[str, Any]: '''simple docstring''' if len(__a) == 0: return "" _UpperCamelCase = None if any(char.isdigit() for char in word): raise Exception(F'''Parameters should not contain numbers: \'{word}\' contains a number''') if word in info["short_word"]: return info["short_word"][word] for prefix_len in range(1 , len(__a) + 1): _UpperCamelCase = word[:prefix_len] if prefix in info["reverse_short_word"]: continue else: _UpperCamelCase = prefix break if short_word is None: # Paranoid fallback def int_to_alphabetic(__a): _UpperCamelCase = '''''' while integer != 0: _UpperCamelCase = chr(ord('''A''') + integer % 10) + s integer //= 10 return s _UpperCamelCase = 0 while True: _UpperCamelCase = word + '''#''' + int_to_alphabetic(__a) if sword in info["reverse_short_word"]: continue else: _UpperCamelCase = sword break _UpperCamelCase = short_word _UpperCamelCase = word return short_word @staticmethod def UpperCAmelCase ( __a , __a) -> Any: '''simple docstring''' _UpperCamelCase = param_name.split('''_''') _UpperCamelCase = [TrialShortNamer.shortname_for_word(__a , __a) for word in words] # We try to create a separatorless short name, but if there is a collision we have to fallback # to a separated short name _UpperCamelCase = ['''''', '''_'''] for separator in separators: _UpperCamelCase = separator.join(__a) if shortname not in info["reverse_short_param"]: _UpperCamelCase = shortname _UpperCamelCase = param_name return shortname return param_name @staticmethod def UpperCAmelCase ( __a , __a) -> List[str]: '''simple docstring''' _UpperCamelCase = TrialShortNamer.shortname_for_key(__a , __a) _UpperCamelCase = short_name _UpperCamelCase = param_name @classmethod def UpperCAmelCase ( cls) -> Any: '''simple docstring''' if cls.NAMING_INFO is not None: return _UpperCamelCase = { '''short_word''': {}, '''reverse_short_word''': {}, '''short_param''': {}, '''reverse_short_param''': {}, } _UpperCamelCase = list(cls.DEFAULTS.keys()) for k in field_keys: cls.add_new_param_name(__a , __a) _UpperCamelCase = info @classmethod def UpperCAmelCase ( cls , __a) -> Optional[Any]: '''simple docstring''' cls.build_naming_info() assert cls.PREFIX is not None _UpperCamelCase = [copy.copy(cls.PREFIX)] for k, v in params.items(): if k not in cls.DEFAULTS: raise Exception(F'''You should provide a default value for the param name {k} with value {v}''') if v == cls.DEFAULTS[k]: # The default value is not added to the name continue _UpperCamelCase = cls.NAMING_INFO['''short_param'''][k] if isinstance(__a , __a): _UpperCamelCase = 1 if v else 0 _UpperCamelCase = '''''' if isinstance(__a , (int, float)) else '''-''' _UpperCamelCase = F'''{key}{sep}{v}''' name.append(__a) return "_".join(__a) @classmethod def UpperCAmelCase ( cls , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = repr[len(cls.PREFIX) + 1 :] if repr == "": _UpperCamelCase = [] else: _UpperCamelCase = repr.split('''_''') _UpperCamelCase = {} for value in values: if "-" in value: _UpperCamelCase , _UpperCamelCase = value.split('''-''') else: _UpperCamelCase = re.sub('''[0-9.]''' , '''''' , __a) _UpperCamelCase = float(re.sub('''[^0-9.]''' , '''''' , __a)) _UpperCamelCase = cls.NAMING_INFO['''reverse_short_param'''][p_k] _UpperCamelCase = p_v for k in cls.DEFAULTS: if k not in parameters: _UpperCamelCase = cls.DEFAULTS[k] return parameters
78
1
"""simple docstring""" def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> Tuple: """simple docstring""" if n == 0: return 1 elif n % 2 == 1: return (binary_exponentiation(__snake_case, n - 1, __snake_case ) * a) % mod else: _UpperCamelCase = binary_exponentiation(__snake_case, n / 2, __snake_case ) return (b * b) % mod # a prime number _a = 701 _a = 10_0000_0000 _a = 10 # using binary exponentiation function, O(log(p)): print((a / b) % p == (a * binary_exponentiation(b, p - 2, p)) % p) print((a / b) % p == (a * b ** (p - 2)) % p)
78
"""simple docstring""" import os import time import pytest from datasets.utils.filelock import FileLock, Timeout def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = FileLock(str(tmpdir / '''foo.lock''' ) ) _UpperCamelCase = FileLock(str(tmpdir / '''foo.lock''' ) ) _UpperCamelCase = 0.01 with locka.acquire(): with pytest.raises(__snake_case ): _UpperCamelCase = time.time() locka.acquire(__snake_case ) assert time.time() - _start > timeout def lowerCamelCase__ ( __snake_case ) -> List[Any]: """simple docstring""" _UpperCamelCase = '''a''' * 10_00 + '''.lock''' _UpperCamelCase = FileLock(str(tmpdir / filename ) ) assert locka._lock_file.endswith('''.lock''' ) assert not locka._lock_file.endswith(__snake_case ) assert len(os.path.basename(locka._lock_file ) ) <= 2_55 _UpperCamelCase = FileLock(tmpdir / filename ) with locka.acquire(): with pytest.raises(__snake_case ): locka.acquire(0 )
78
1
"""simple docstring""" _a = [ """Audio""", """Array2D""", """Array3D""", """Array4D""", """Array5D""", """ClassLabel""", """Features""", """Sequence""", """Value""", """Image""", """Translation""", """TranslationVariableLanguages""", ] from .audio import Audio from .features import ArrayaD, ArrayaD, ArrayaD, ArrayaD, ClassLabel, Features, Sequence, Value from .image import Image from .translation import Translation, TranslationVariableLanguages
78
"""simple docstring""" from math import sqrt def lowerCamelCase__ ( __snake_case ) -> bool: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and ( number >= 0 ), "'number' must been an int and positive" _UpperCamelCase = True # 0 and 1 are none primes. if number <= 1: _UpperCamelCase = False for divisor in range(2, int(round(sqrt(__snake_case ) ) ) + 1 ): # if 'number' divisible by 'divisor' then sets 'status' # of false and break up the loop. if number % divisor == 0: _UpperCamelCase = False break # precondition assert isinstance(__snake_case, __snake_case ), "'status' must been from type bool" return status def lowerCamelCase__ ( __snake_case ) -> Dict: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n > 2), "'N' must been an int and > 2" # beginList: contains all natural numbers from 2 up to N _UpperCamelCase = list(range(2, n + 1 ) ) _UpperCamelCase = [] # this list will be returns. # actual sieve of erathostenes for i in range(len(__snake_case ) ): for j in range(i + 1, len(__snake_case ) ): if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0): _UpperCamelCase = 0 # filters actual prime numbers. _UpperCamelCase = [x for x in begin_list if x != 0] # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type list" return ans def lowerCamelCase__ ( __snake_case ) -> List[str]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n > 2), "'N' must been an int and > 2" _UpperCamelCase = [] # iterates over all numbers between 2 up to N+1 # if a number is prime then appends to list 'ans' for number in range(2, n + 1 ): if is_prime(__snake_case ): ans.append(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type list" return ans def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and number >= 0, "'number' must been an int and >= 0" _UpperCamelCase = [] # this list will be returns of the function. # potential prime number factors. _UpperCamelCase = 2 _UpperCamelCase = number if number == 0 or number == 1: ans.append(__snake_case ) # if 'number' not prime then builds the prime factorization of 'number' elif not is_prime(__snake_case ): while quotient != 1: if is_prime(__snake_case ) and (quotient % factor == 0): ans.append(__snake_case ) quotient /= factor else: factor += 1 else: ans.append(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type list" return ans def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and ( number >= 0 ), "'number' bust been an int and >= 0" _UpperCamelCase = 0 # prime factorization of 'number' _UpperCamelCase = prime_factorization(__snake_case ) _UpperCamelCase = max(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type int" return ans def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and ( number >= 0 ), "'number' bust been an int and >= 0" _UpperCamelCase = 0 # prime factorization of 'number' _UpperCamelCase = prime_factorization(__snake_case ) _UpperCamelCase = min(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type int" return ans def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ), "'number' must been an int" assert isinstance(number % 2 == 0, __snake_case ), "compare bust been from type bool" return number % 2 == 0 def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ), "'number' must been an int" assert isinstance(number % 2 != 0, __snake_case ), "compare bust been from type bool" return number % 2 != 0 def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" assert ( isinstance(__snake_case, __snake_case ) and (number > 2) and is_even(__snake_case ) ), "'number' must been an int, even and > 2" _UpperCamelCase = [] # this list will returned # creates a list of prime numbers between 2 up to 'number' _UpperCamelCase = get_prime_numbers(__snake_case ) _UpperCamelCase = len(__snake_case ) # run variable for while-loops. _UpperCamelCase = 0 _UpperCamelCase = None # exit variable. for break up the loops _UpperCamelCase = True while i < len_pn and loop: _UpperCamelCase = i + 1 while j < len_pn and loop: if prime_numbers[i] + prime_numbers[j] == number: _UpperCamelCase = False ans.append(prime_numbers[i] ) ans.append(prime_numbers[j] ) j += 1 i += 1 # precondition assert ( isinstance(__snake_case, __snake_case ) and (len(__snake_case ) == 2) and (ans[0] + ans[1] == number) and is_prime(ans[0] ) and is_prime(ans[1] ) ), "'ans' must contains two primes. And sum of elements must been eq 'number'" return ans def lowerCamelCase__ ( __snake_case, __snake_case ) -> str: """simple docstring""" assert ( isinstance(__snake_case, __snake_case ) and isinstance(__snake_case, __snake_case ) and (numbera >= 0) and (numbera >= 0) ), "'number1' and 'number2' must been positive integer." _UpperCamelCase = 0 while numbera != 0: _UpperCamelCase = numbera % numbera _UpperCamelCase = numbera _UpperCamelCase = rest # precondition assert isinstance(__snake_case, __snake_case ) and ( numbera >= 0 ), "'number' must been from type int and positive" return numbera def lowerCamelCase__ ( __snake_case, __snake_case ) -> List[str]: """simple docstring""" assert ( isinstance(__snake_case, __snake_case ) and isinstance(__snake_case, __snake_case ) and (numbera >= 1) and (numbera >= 1) ), "'number1' and 'number2' must been positive integer." _UpperCamelCase = 1 # actual answer that will be return. # for kgV (x,1) if numbera > 1 and numbera > 1: # builds the prime factorization of 'number1' and 'number2' _UpperCamelCase = prime_factorization(__snake_case ) _UpperCamelCase = prime_factorization(__snake_case ) elif numbera == 1 or numbera == 1: _UpperCamelCase = [] _UpperCamelCase = [] _UpperCamelCase = max(__snake_case, __snake_case ) _UpperCamelCase = 0 _UpperCamelCase = 0 _UpperCamelCase = [] # captured numbers int both 'primeFac1' and 'primeFac2' # iterates through primeFac1 for n in prime_fac_a: if n not in done: if n in prime_fac_a: _UpperCamelCase = prime_fac_a.count(__snake_case ) _UpperCamelCase = prime_fac_a.count(__snake_case ) for _ in range(max(__snake_case, __snake_case ) ): ans *= n else: _UpperCamelCase = prime_fac_a.count(__snake_case ) for _ in range(__snake_case ): ans *= n done.append(__snake_case ) # iterates through primeFac2 for n in prime_fac_a: if n not in done: _UpperCamelCase = prime_fac_a.count(__snake_case ) for _ in range(__snake_case ): ans *= n done.append(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ) and ( ans >= 0 ), "'ans' must been from type int and positive" return ans def lowerCamelCase__ ( __snake_case ) -> Tuple: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n >= 0), "'number' must been a positive int" _UpperCamelCase = 0 _UpperCamelCase = 2 # this variable holds the answer while index < n: index += 1 ans += 1 # counts to the next number # if ans not prime then # runs to the next prime number. while not is_prime(__snake_case ): ans += 1 # precondition assert isinstance(__snake_case, __snake_case ) and is_prime( __snake_case ), "'ans' must been a prime number and from type int" return ans def lowerCamelCase__ ( __snake_case, __snake_case ) -> Tuple: """simple docstring""" assert ( is_prime(__snake_case ) and is_prime(__snake_case ) and (p_number_a < p_number_a) ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'" _UpperCamelCase = p_number_a + 1 # jump to the next number _UpperCamelCase = [] # this list will be returns. # if number is not prime then # fetch the next prime number. while not is_prime(__snake_case ): number += 1 while number < p_number_a: ans.append(__snake_case ) number += 1 # fetch the next prime number. while not is_prime(__snake_case ): number += 1 # precondition assert ( isinstance(__snake_case, __snake_case ) and ans[0] != p_number_a and ans[len(__snake_case ) - 1] != p_number_a ), "'ans' must been a list without the arguments" # 'ans' contains not 'pNumber1' and 'pNumber2' ! return ans def lowerCamelCase__ ( __snake_case ) -> List[Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n >= 1), "'n' must been int and >= 1" _UpperCamelCase = [] # will be returned. for divisor in range(1, n + 1 ): if n % divisor == 0: ans.append(__snake_case ) # precondition assert ans[0] == 1 and ans[len(__snake_case ) - 1] == n, "Error in function getDivisiors(...)" return ans def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and ( number > 1 ), "'number' must been an int and >= 1" _UpperCamelCase = get_divisors(__snake_case ) # precondition assert ( isinstance(__snake_case, __snake_case ) and (divisors[0] == 1) and (divisors[len(__snake_case ) - 1] == number) ), "Error in help-function getDivisiors(...)" # summed all divisors up to 'number' (exclusive), hence [:-1] return sum(divisors[:-1] ) == number def lowerCamelCase__ ( __snake_case, __snake_case ) -> Optional[Any]: """simple docstring""" assert ( isinstance(__snake_case, __snake_case ) and isinstance(__snake_case, __snake_case ) and (denominator != 0) ), "The arguments must been from type int and 'denominator' != 0" # build the greatest common divisor of numerator and denominator. _UpperCamelCase = gcd(abs(__snake_case ), abs(__snake_case ) ) # precondition assert ( isinstance(__snake_case, __snake_case ) and (numerator % gcd_of_fraction == 0) and (denominator % gcd_of_fraction == 0) ), "Error in function gcd(...,...)" return (numerator // gcd_of_fraction, denominator // gcd_of_fraction) def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n >= 0), "'n' must been a int and >= 0" _UpperCamelCase = 1 # this will be return. for factor in range(1, n + 1 ): ans *= factor return ans def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n >= 0), "'n' must been an int and >= 0" _UpperCamelCase = 0 _UpperCamelCase = 1 _UpperCamelCase = 1 # this will be return for _ in range(n - 1 ): _UpperCamelCase = ans ans += fiba _UpperCamelCase = tmp return ans
78
1
"""simple docstring""" import gc import unittest import torch from parameterized import parameterized from diffusers import AutoencoderKL from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class _UpperCAmelCase( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase__ = AutoencoderKL lowercase__ = 'sample' lowercase__ = 1E-2 @property def UpperCAmelCase ( self) -> Dict: '''simple docstring''' _UpperCamelCase = 4 _UpperCamelCase = 3 _UpperCamelCase = (32, 32) _UpperCamelCase = floats_tensor((batch_size, num_channels) + sizes).to(__a) return {"sample": image} @property def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' return (3, 32, 32) @property def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' return (3, 32, 32) def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = { '''block_out_channels''': [32, 64], '''in_channels''': 3, '''out_channels''': 3, '''down_block_types''': ['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''], '''up_block_types''': ['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''], '''latent_channels''': 4, } _UpperCamelCase = self.dummy_input return init_dict, inputs_dict def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' pass def UpperCAmelCase ( self) -> str: '''simple docstring''' pass @unittest.skipIf(torch_device == '''mps''' , '''Gradient checkpointing skipped on MPS''') def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' # enable deterministic behavior for gradient checkpointing _UpperCamelCase , _UpperCamelCase = self.prepare_init_args_and_inputs_for_common() _UpperCamelCase = self.model_class(**__a) model.to(__a) assert not model.is_gradient_checkpointing and model.training _UpperCamelCase = model(**__a).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model.zero_grad() _UpperCamelCase = torch.randn_like(__a) _UpperCamelCase = (out - labels).mean() loss.backward() # re-instantiate the model now enabling gradient checkpointing _UpperCamelCase = self.model_class(**__a) # clone model model_a.load_state_dict(model.state_dict()) model_a.to(__a) model_a.enable_gradient_checkpointing() assert model_a.is_gradient_checkpointing and model_a.training _UpperCamelCase = model_a(**__a).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model_a.zero_grad() _UpperCamelCase = (out_a - labels).mean() loss_a.backward() # compare the output and parameters gradients self.assertTrue((loss - loss_a).abs() < 1e-5) _UpperCamelCase = dict(model.named_parameters()) _UpperCamelCase = dict(model_a.named_parameters()) for name, param in named_params.items(): self.assertTrue(torch_all_close(param.grad.data , named_params_a[name].grad.data , atol=5e-5)) def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = AutoencoderKL.from_pretrained('''fusing/autoencoder-kl-dummy''' , output_loading_info=__a) self.assertIsNotNone(__a) self.assertEqual(len(loading_info['''missing_keys''']) , 0) model.to(__a) _UpperCamelCase = model(**self.dummy_input) assert image is not None, "Make sure output is not None" def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = AutoencoderKL.from_pretrained('''fusing/autoencoder-kl-dummy''') _UpperCamelCase = model.to(__a) model.eval() if torch_device == "mps": _UpperCamelCase = torch.manual_seed(0) else: _UpperCamelCase = torch.Generator(device=__a).manual_seed(0) _UpperCamelCase = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0) , ) _UpperCamelCase = image.to(__a) with torch.no_grad(): _UpperCamelCase = model(__a , sample_posterior=__a , generator=__a).sample _UpperCamelCase = output[0, -1, -3:, -3:].flatten().cpu() # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. if torch_device == "mps": _UpperCamelCase = torch.tensor( [ -4.0_078e-01, -3.8_323e-04, -1.2_681e-01, -1.1_462e-01, 2.0_095e-01, 1.0_893e-01, -8.8_247e-02, -3.0_361e-01, -9.8_644e-03, ]) elif torch_device == "cpu": _UpperCamelCase = torch.tensor( [-0.1352, 0.0878, 0.0419, -0.0818, -0.1069, 0.0688, -0.1458, -0.4446, -0.0026]) else: _UpperCamelCase = torch.tensor( [-0.2421, 0.4642, 0.2507, -0.0438, 0.0682, 0.3160, -0.2018, -0.0727, 0.2485]) self.assertTrue(torch_all_close(__a , __a , rtol=1e-2)) @slow class _UpperCAmelCase( unittest.TestCase ): def UpperCAmelCase ( self , __a , __a) -> Optional[Any]: '''simple docstring''' return F'''gaussian_noise_s={seed}_shape={"_".join([str(__a) for s in shape])}.npy''' def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase ( self , __a=0 , __a=(4, 3, 5_12, 5_12) , __a=False) -> str: '''simple docstring''' _UpperCamelCase = torch.floataa if fpaa else torch.floataa _UpperCamelCase = torch.from_numpy(load_hf_numpy(self.get_file_format(__a , __a))).to(__a).to(__a) return image def UpperCAmelCase ( self , __a="CompVis/stable-diffusion-v1-4" , __a=False) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = '''fp16''' if fpaa else None _UpperCamelCase = torch.floataa if fpaa else torch.floataa _UpperCamelCase = AutoencoderKL.from_pretrained( __a , subfolder='''vae''' , torch_dtype=__a , revision=__a , ) model.to(__a).eval() return model def UpperCAmelCase ( self , __a=0) -> Tuple: '''simple docstring''' if torch_device == "mps": return torch.manual_seed(__a) return torch.Generator(device=__a).manual_seed(__a) @parameterized.expand( [ # fmt: off [33, [-0.1603, 0.9878, -0.0495, -0.0790, -0.2709, 0.8375, -0.2060, -0.0824], [-0.2395, 0.0098, 0.0102, -0.0709, -0.2840, -0.0274, -0.0718, -0.1824]], [47, [-0.2376, 0.1168, 0.1332, -0.4840, -0.2508, -0.0791, -0.0493, -0.4089], [0.0350, 0.0847, 0.0467, 0.0344, -0.0842, -0.0547, -0.0633, -0.1131]], # fmt: on ]) def UpperCAmelCase ( self , __a , __a , __a) -> Any: '''simple docstring''' _UpperCamelCase = self.get_sd_vae_model() _UpperCamelCase = self.get_sd_image(__a) _UpperCamelCase = self.get_generator(__a) with torch.no_grad(): _UpperCamelCase = model(__a , generator=__a , sample_posterior=__a).sample assert sample.shape == image.shape _UpperCamelCase = sample[-1, -2:, -2:, :2].flatten().float().cpu() _UpperCamelCase = torch.tensor(expected_slice_mps if torch_device == '''mps''' else expected_slice) assert torch_all_close(__a , __a , atol=3e-3) @parameterized.expand( [ # fmt: off [33, [-0.0513, 0.0289, 1.3799, 0.2166, -0.2573, -0.0871, 0.5103, -0.0999]], [47, [-0.4128, -0.1320, -0.3704, 0.1965, -0.4116, -0.2332, -0.3340, 0.2247]], # fmt: on ]) @require_torch_gpu def UpperCAmelCase ( self , __a , __a) -> List[Any]: '''simple docstring''' _UpperCamelCase = self.get_sd_vae_model(fpaa=__a) _UpperCamelCase = self.get_sd_image(__a , fpaa=__a) _UpperCamelCase = self.get_generator(__a) with torch.no_grad(): _UpperCamelCase = model(__a , generator=__a , sample_posterior=__a).sample assert sample.shape == image.shape _UpperCamelCase = sample[-1, -2:, :2, -2:].flatten().float().cpu() _UpperCamelCase = torch.tensor(__a) assert torch_all_close(__a , __a , atol=1e-2) @parameterized.expand( [ # fmt: off [33, [-0.1609, 0.9866, -0.0487, -0.0777, -0.2716, 0.8368, -0.2055, -0.0814], [-0.2395, 0.0098, 0.0102, -0.0709, -0.2840, -0.0274, -0.0718, -0.1824]], [47, [-0.2377, 0.1147, 0.1333, -0.4841, -0.2506, -0.0805, -0.0491, -0.4085], [0.0350, 0.0847, 0.0467, 0.0344, -0.0842, -0.0547, -0.0633, -0.1131]], # fmt: on ]) def UpperCAmelCase ( self , __a , __a , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = self.get_sd_vae_model() _UpperCamelCase = self.get_sd_image(__a) with torch.no_grad(): _UpperCamelCase = model(__a).sample assert sample.shape == image.shape _UpperCamelCase = sample[-1, -2:, -2:, :2].flatten().float().cpu() _UpperCamelCase = torch.tensor(expected_slice_mps if torch_device == '''mps''' else expected_slice) assert torch_all_close(__a , __a , atol=3e-3) @parameterized.expand( [ # fmt: off [13, [-0.2051, -0.1803, -0.2311, -0.2114, -0.3292, -0.3574, -0.2953, -0.3323]], [37, [-0.2632, -0.2625, -0.2199, -0.2741, -0.4539, -0.4990, -0.3720, -0.4925]], # fmt: on ]) @require_torch_gpu def UpperCAmelCase ( self , __a , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = self.get_sd_vae_model() _UpperCamelCase = self.get_sd_image(__a , shape=(3, 4, 64, 64)) with torch.no_grad(): _UpperCamelCase = model.decode(__a).sample assert list(sample.shape) == [3, 3, 5_12, 5_12] _UpperCamelCase = sample[-1, -2:, :2, -2:].flatten().cpu() _UpperCamelCase = torch.tensor(__a) assert torch_all_close(__a , __a , atol=1e-3) @parameterized.expand( [ # fmt: off [27, [-0.0369, 0.0207, -0.0776, -0.0682, -0.1747, -0.1930, -0.1465, -0.2039]], [16, [-0.1628, -0.2134, -0.2747, -0.2642, -0.3774, -0.4404, -0.3687, -0.4277]], # fmt: on ]) @require_torch_gpu def UpperCAmelCase ( self , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.get_sd_vae_model(fpaa=__a) _UpperCamelCase = self.get_sd_image(__a , shape=(3, 4, 64, 64) , fpaa=__a) with torch.no_grad(): _UpperCamelCase = model.decode(__a).sample assert list(sample.shape) == [3, 3, 5_12, 5_12] _UpperCamelCase = sample[-1, -2:, :2, -2:].flatten().float().cpu() _UpperCamelCase = torch.tensor(__a) assert torch_all_close(__a , __a , atol=5e-3) @parameterized.expand([(13,), (16,), (27,)]) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason='''xformers is not required when using PyTorch 2.0.''') def UpperCAmelCase ( self , __a) -> str: '''simple docstring''' _UpperCamelCase = self.get_sd_vae_model(fpaa=__a) _UpperCamelCase = self.get_sd_image(__a , shape=(3, 4, 64, 64) , fpaa=__a) with torch.no_grad(): _UpperCamelCase = model.decode(__a).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): _UpperCamelCase = model.decode(__a).sample assert list(sample.shape) == [3, 3, 5_12, 5_12] assert torch_all_close(__a , __a , atol=1e-1) @parameterized.expand([(13,), (16,), (37,)]) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason='''xformers is not required when using PyTorch 2.0.''') def UpperCAmelCase ( self , __a) -> int: '''simple docstring''' _UpperCamelCase = self.get_sd_vae_model() _UpperCamelCase = self.get_sd_image(__a , shape=(3, 4, 64, 64)) with torch.no_grad(): _UpperCamelCase = model.decode(__a).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): _UpperCamelCase = model.decode(__a).sample assert list(sample.shape) == [3, 3, 5_12, 5_12] assert torch_all_close(__a , __a , atol=1e-2) @parameterized.expand( [ # fmt: off [33, [-0.3001, 0.0918, -2.6984, -3.9720, -3.2099, -5.0353, 1.7338, -0.2065, 3.4267]], [47, [-1.5030, -4.3871, -6.0355, -9.1157, -1.6661, -2.7853, 2.1607, -5.0823, 2.5633]], # fmt: on ]) def UpperCAmelCase ( self , __a , __a) -> List[Any]: '''simple docstring''' _UpperCamelCase = self.get_sd_vae_model() _UpperCamelCase = self.get_sd_image(__a) _UpperCamelCase = self.get_generator(__a) with torch.no_grad(): _UpperCamelCase = model.encode(__a).latent_dist _UpperCamelCase = dist.sample(generator=__a) assert list(sample.shape) == [image.shape[0], 4] + [i // 8 for i in image.shape[2:]] _UpperCamelCase = sample[0, -1, -3:, -3:].flatten().cpu() _UpperCamelCase = torch.tensor(__a) _UpperCamelCase = 3e-3 if torch_device != '''mps''' else 1e-2 assert torch_all_close(__a , __a , atol=__a)
78
"""simple docstring""" import logging from dataclasses import dataclass, field from pathlib import Path from typing import Optional, Union from .generation.configuration_utils import GenerationConfig from .training_args import TrainingArguments from .utils import add_start_docstrings _a = logging.getLogger(__name__) @dataclass @add_start_docstrings(TrainingArguments.__doc__ ) class _UpperCAmelCase( lowerCamelCase ): lowercase__ = field(default=lowerCamelCase , metadata={'help': 'Whether to use SortishSampler or not.'} ) lowercase__ = field( default=lowerCamelCase , metadata={'help': 'Whether to use generate to calculate generative metrics (ROUGE, BLEU).'} ) lowercase__ = field( default=lowerCamelCase , metadata={ 'help': ( 'The `max_length` to use on each evaluation loop when `predict_with_generate=True`. Will default ' 'to the `max_length` value of the model configuration.' ) } , ) lowercase__ = field( default=lowerCamelCase , metadata={ 'help': ( 'The `num_beams` to use on each evaluation loop when `predict_with_generate=True`. Will default ' 'to the `num_beams` value of the model configuration.' ) } , ) lowercase__ = field( default=lowerCamelCase , metadata={ 'help': 'Model id, file path or url pointing to a GenerationConfig json file, to use during prediction.' } , ) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = super().to_dict() for k, v in d.items(): if isinstance(__a , __a): _UpperCamelCase = v.to_dict() return d
78
1
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(""">=""", """4.25.0""")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
78
"""simple docstring""" import argparse import torch from transformers import BlenderbotConfig, BlenderbotForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) _a = [ ["""attention""", """attn"""], ["""encoder_attention""", """encoder_attn"""], ["""q_lin""", """q_proj"""], ["""k_lin""", """k_proj"""], ["""v_lin""", """v_proj"""], ["""out_lin""", """out_proj"""], ["""norm_embeddings""", """layernorm_embedding"""], ["""position_embeddings""", """embed_positions"""], ["""embeddings""", """embed_tokens"""], ["""ffn.lin""", """fc"""], ] def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" if k == "embeddings.weight": return "shared.weight" for parlai_name, hf_name in PATTERNS: _UpperCamelCase = k.replace(__snake_case, __snake_case ) if k.startswith('''encoder''' ): _UpperCamelCase = k.replace('''.attn''', '''.self_attn''' ) _UpperCamelCase = k.replace('''norm1''', '''self_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm2''', '''final_layer_norm''' ) elif k.startswith('''decoder''' ): _UpperCamelCase = k.replace('''norm1''', '''self_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm2''', '''encoder_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm3''', '''final_layer_norm''' ) return k def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" _UpperCamelCase = [ '''model.encoder.layernorm_embedding.weight''', '''model.encoder.layernorm_embedding.bias''', '''model.decoder.layernorm_embedding.weight''', '''model.decoder.layernorm_embedding.bias''', ] for k in keys: _UpperCamelCase = sd.pop(__snake_case ) _UpperCamelCase = k.replace('''layernorm_embedding''', '''layer_norm''' ) assert new_k not in sd _UpperCamelCase = v _a = ["""START"""] @torch.no_grad() def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> int: """simple docstring""" _UpperCamelCase = torch.load(__snake_case, map_location='''cpu''' ) _UpperCamelCase = model['''model'''] _UpperCamelCase = BlenderbotConfig.from_json_file(__snake_case ) _UpperCamelCase = BlenderbotForConditionalGeneration(__snake_case ) _UpperCamelCase = m.model.state_dict().keys() _UpperCamelCase = [] _UpperCamelCase = {} for k, v in sd.items(): if k in IGNORE_KEYS: continue _UpperCamelCase = rename_state_dict_key(__snake_case ) if new_k not in valid_keys: failures.append([k, new_k] ) else: _UpperCamelCase = v if cfg.normalize_before: # Blenderbot-3B checkpoints. Rename layernorm_embedding -> layer_norm rename_layernorm_keys(__snake_case ) m.model.load_state_dict(__snake_case, strict=__snake_case ) m.half() m.save_pretrained(__snake_case ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument("""--src_path""", type=str, help="""like blenderbot-model.bin""") parser.add_argument("""--save_dir""", default="""hf_blenderbot""", type=str, help="""Where to save converted model.""") parser.add_argument( """--hf_config_json""", default="""blenderbot-3b-config.json""", type=str, help="""Path to config to use""" ) _a = parser.parse_args() convert_parlai_checkpoint(args.src_path, args.save_dir, args.hf_config_json)
78
1
"""simple docstring""" from __future__ import annotations def lowerCamelCase__ ( __snake_case ) -> bool: """simple docstring""" _UpperCamelCase = str(__snake_case ) return n == n[::-1] def lowerCamelCase__ ( __snake_case = 1_00_00_00 ) -> Optional[Any]: """simple docstring""" _UpperCamelCase = 0 for i in range(1, __snake_case ): if is_palindrome(__snake_case ) and is_palindrome(bin(__snake_case ).split('''b''' )[1] ): total += i return total if __name__ == "__main__": print(solution(int(str(input().strip()))))
78
"""simple docstring""" import argparse import os.path as osp import re import torch from safetensors.torch import load_file, save_file # =================# # UNet Conversion # # =================# _a = [ # (stable-diffusion, HF Diffusers) ("""time_embed.0.weight""", """time_embedding.linear_1.weight"""), ("""time_embed.0.bias""", """time_embedding.linear_1.bias"""), ("""time_embed.2.weight""", """time_embedding.linear_2.weight"""), ("""time_embed.2.bias""", """time_embedding.linear_2.bias"""), ("""input_blocks.0.0.weight""", """conv_in.weight"""), ("""input_blocks.0.0.bias""", """conv_in.bias"""), ("""out.0.weight""", """conv_norm_out.weight"""), ("""out.0.bias""", """conv_norm_out.bias"""), ("""out.2.weight""", """conv_out.weight"""), ("""out.2.bias""", """conv_out.bias"""), ] _a = [ # (stable-diffusion, HF Diffusers) ("""in_layers.0""", """norm1"""), ("""in_layers.2""", """conv1"""), ("""out_layers.0""", """norm2"""), ("""out_layers.3""", """conv2"""), ("""emb_layers.1""", """time_emb_proj"""), ("""skip_connection""", """conv_shortcut"""), ] _a = [] # hardcoded number of downblocks and resnets/attentions... # would need smarter logic for other networks. for i in range(4): # loop over downblocks/upblocks for j in range(2): # loop over resnets/attentions for downblocks _a = F"""down_blocks.{i}.resnets.{j}.""" _a = F"""input_blocks.{3*i + j + 1}.0.""" unet_conversion_map_layer.append((sd_down_res_prefix, hf_down_res_prefix)) if i < 3: # no attention layers in down_blocks.3 _a = F"""down_blocks.{i}.attentions.{j}.""" _a = F"""input_blocks.{3*i + j + 1}.1.""" unet_conversion_map_layer.append((sd_down_atn_prefix, hf_down_atn_prefix)) for j in range(3): # loop over resnets/attentions for upblocks _a = F"""up_blocks.{i}.resnets.{j}.""" _a = F"""output_blocks.{3*i + j}.0.""" unet_conversion_map_layer.append((sd_up_res_prefix, hf_up_res_prefix)) if i > 0: # no attention layers in up_blocks.0 _a = F"""up_blocks.{i}.attentions.{j}.""" _a = F"""output_blocks.{3*i + j}.1.""" unet_conversion_map_layer.append((sd_up_atn_prefix, hf_up_atn_prefix)) if i < 3: # no downsample in down_blocks.3 _a = F"""down_blocks.{i}.downsamplers.0.conv.""" _a = F"""input_blocks.{3*(i+1)}.0.op.""" unet_conversion_map_layer.append((sd_downsample_prefix, hf_downsample_prefix)) # no upsample in up_blocks.3 _a = F"""up_blocks.{i}.upsamplers.0.""" _a = F"""output_blocks.{3*i + 2}.{1 if i == 0 else 2}.""" unet_conversion_map_layer.append((sd_upsample_prefix, hf_upsample_prefix)) _a = """mid_block.attentions.0.""" _a = """middle_block.1.""" unet_conversion_map_layer.append((sd_mid_atn_prefix, hf_mid_atn_prefix)) for j in range(2): _a = F"""mid_block.resnets.{j}.""" _a = F"""middle_block.{2*j}.""" unet_conversion_map_layer.append((sd_mid_res_prefix, hf_mid_res_prefix)) def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" _UpperCamelCase = {k: k for k in unet_state_dict.keys()} for sd_name, hf_name in unet_conversion_map: _UpperCamelCase = sd_name for k, v in mapping.items(): if "resnets" in k: for sd_part, hf_part in unet_conversion_map_resnet: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v for k, v in mapping.items(): for sd_part, hf_part in unet_conversion_map_layer: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v _UpperCamelCase = {v: unet_state_dict[k] for k, v in mapping.items()} return new_state_dict # ================# # VAE Conversion # # ================# _a = [ # (stable-diffusion, HF Diffusers) ("""nin_shortcut""", """conv_shortcut"""), ("""norm_out""", """conv_norm_out"""), ("""mid.attn_1.""", """mid_block.attentions.0."""), ] for i in range(4): # down_blocks have two resnets for j in range(2): _a = F"""encoder.down_blocks.{i}.resnets.{j}.""" _a = F"""encoder.down.{i}.block.{j}.""" vae_conversion_map.append((sd_down_prefix, hf_down_prefix)) if i < 3: _a = F"""down_blocks.{i}.downsamplers.0.""" _a = F"""down.{i}.downsample.""" vae_conversion_map.append((sd_downsample_prefix, hf_downsample_prefix)) _a = F"""up_blocks.{i}.upsamplers.0.""" _a = F"""up.{3-i}.upsample.""" vae_conversion_map.append((sd_upsample_prefix, hf_upsample_prefix)) # up_blocks have three resnets # also, up blocks in hf are numbered in reverse from sd for j in range(3): _a = F"""decoder.up_blocks.{i}.resnets.{j}.""" _a = F"""decoder.up.{3-i}.block.{j}.""" vae_conversion_map.append((sd_up_prefix, hf_up_prefix)) # this part accounts for mid blocks in both the encoder and the decoder for i in range(2): _a = F"""mid_block.resnets.{i}.""" _a = F"""mid.block_{i+1}.""" vae_conversion_map.append((sd_mid_res_prefix, hf_mid_res_prefix)) _a = [ # (stable-diffusion, HF Diffusers) ("""norm.""", """group_norm."""), ("""q.""", """query."""), ("""k.""", """key."""), ("""v.""", """value."""), ("""proj_out.""", """proj_attn."""), ] def lowerCamelCase__ ( __snake_case ) -> List[str]: """simple docstring""" return w.reshape(*w.shape, 1, 1 ) def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" _UpperCamelCase = {k: k for k in vae_state_dict.keys()} for k, v in mapping.items(): for sd_part, hf_part in vae_conversion_map: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v for k, v in mapping.items(): if "attentions" in k: for sd_part, hf_part in vae_conversion_map_attn: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v _UpperCamelCase = {v: vae_state_dict[k] for k, v in mapping.items()} _UpperCamelCase = ['''q''', '''k''', '''v''', '''proj_out'''] for k, v in new_state_dict.items(): for weight_name in weights_to_convert: if F'''mid.attn_1.{weight_name}.weight''' in k: print(F'''Reshaping {k} for SD format''' ) _UpperCamelCase = reshape_weight_for_sd(__snake_case ) return new_state_dict # =========================# # Text Encoder Conversion # # =========================# _a = [ # (stable-diffusion, HF Diffusers) ("""resblocks.""", """text_model.encoder.layers."""), ("""ln_1""", """layer_norm1"""), ("""ln_2""", """layer_norm2"""), (""".c_fc.""", """.fc1."""), (""".c_proj.""", """.fc2."""), (""".attn""", """.self_attn"""), ("""ln_final.""", """transformer.text_model.final_layer_norm."""), ("""token_embedding.weight""", """transformer.text_model.embeddings.token_embedding.weight"""), ("""positional_embedding""", """transformer.text_model.embeddings.position_embedding.weight"""), ] _a = {re.escape(x[1]): x[0] for x in textenc_conversion_lst} _a = re.compile("""|""".join(protected.keys())) # Ordering is from https://github.com/pytorch/pytorch/blob/master/test/cpp/api/modules.cpp _a = {"""q""": 0, """k""": 1, """v""": 2} def lowerCamelCase__ ( __snake_case ) -> Any: """simple docstring""" _UpperCamelCase = {} _UpperCamelCase = {} _UpperCamelCase = {} for k, v in text_enc_dict.items(): if ( k.endswith('''.self_attn.q_proj.weight''' ) or k.endswith('''.self_attn.k_proj.weight''' ) or k.endswith('''.self_attn.v_proj.weight''' ) ): _UpperCamelCase = k[: -len('''.q_proj.weight''' )] _UpperCamelCase = k[-len('''q_proj.weight''' )] if k_pre not in capture_qkv_weight: _UpperCamelCase = [None, None, None] _UpperCamelCase = v continue if ( k.endswith('''.self_attn.q_proj.bias''' ) or k.endswith('''.self_attn.k_proj.bias''' ) or k.endswith('''.self_attn.v_proj.bias''' ) ): _UpperCamelCase = k[: -len('''.q_proj.bias''' )] _UpperCamelCase = k[-len('''q_proj.bias''' )] if k_pre not in capture_qkv_bias: _UpperCamelCase = [None, None, None] _UpperCamelCase = v continue _UpperCamelCase = textenc_pattern.sub(lambda __snake_case : protected[re.escape(m.group(0 ) )], __snake_case ) _UpperCamelCase = v for k_pre, tensors in capture_qkv_weight.items(): if None in tensors: raise Exception('''CORRUPTED MODEL: one of the q-k-v values for the text encoder was missing''' ) _UpperCamelCase = textenc_pattern.sub(lambda __snake_case : protected[re.escape(m.group(0 ) )], __snake_case ) _UpperCamelCase = torch.cat(__snake_case ) for k_pre, tensors in capture_qkv_bias.items(): if None in tensors: raise Exception('''CORRUPTED MODEL: one of the q-k-v values for the text encoder was missing''' ) _UpperCamelCase = textenc_pattern.sub(lambda __snake_case : protected[re.escape(m.group(0 ) )], __snake_case ) _UpperCamelCase = torch.cat(__snake_case ) return new_state_dict def lowerCamelCase__ ( __snake_case ) -> Tuple: """simple docstring""" return text_enc_dict if __name__ == "__main__": _a = argparse.ArgumentParser() parser.add_argument("""--model_path""", default=None, type=str, required=True, help="""Path to the model to convert.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, required=True, help="""Path to the output model.""") parser.add_argument("""--half""", action="""store_true""", help="""Save weights in half precision.""") parser.add_argument( """--use_safetensors""", action="""store_true""", help="""Save weights use safetensors, default is ckpt.""" ) _a = parser.parse_args() assert args.model_path is not None, "Must provide a model path!" assert args.checkpoint_path is not None, "Must provide a checkpoint path!" # Path for safetensors _a = osp.join(args.model_path, """unet""", """diffusion_pytorch_model.safetensors""") _a = osp.join(args.model_path, """vae""", """diffusion_pytorch_model.safetensors""") _a = osp.join(args.model_path, """text_encoder""", """model.safetensors""") # Load models from safetensors if it exists, if it doesn't pytorch if osp.exists(unet_path): _a = load_file(unet_path, device="""cpu""") else: _a = osp.join(args.model_path, """unet""", """diffusion_pytorch_model.bin""") _a = torch.load(unet_path, map_location="""cpu""") if osp.exists(vae_path): _a = load_file(vae_path, device="""cpu""") else: _a = osp.join(args.model_path, """vae""", """diffusion_pytorch_model.bin""") _a = torch.load(vae_path, map_location="""cpu""") if osp.exists(text_enc_path): _a = load_file(text_enc_path, device="""cpu""") else: _a = osp.join(args.model_path, """text_encoder""", """pytorch_model.bin""") _a = torch.load(text_enc_path, map_location="""cpu""") # Convert the UNet model _a = convert_unet_state_dict(unet_state_dict) _a = {"""model.diffusion_model.""" + k: v for k, v in unet_state_dict.items()} # Convert the VAE model _a = convert_vae_state_dict(vae_state_dict) _a = {"""first_stage_model.""" + k: v for k, v in vae_state_dict.items()} # Easiest way to identify v2.0 model seems to be that the text encoder (OpenCLIP) is deeper _a = """text_model.encoder.layers.22.layer_norm2.bias""" in text_enc_dict if is_vaa_model: # Need to add the tag 'transformer' in advance so we can knock it out from the final layer-norm _a = {"""transformer.""" + k: v for k, v in text_enc_dict.items()} _a = convert_text_enc_state_dict_vaa(text_enc_dict) _a = {"""cond_stage_model.model.""" + k: v for k, v in text_enc_dict.items()} else: _a = convert_text_enc_state_dict(text_enc_dict) _a = {"""cond_stage_model.transformer.""" + k: v for k, v in text_enc_dict.items()} # Put together new checkpoint _a = {**unet_state_dict, **vae_state_dict, **text_enc_dict} if args.half: _a = {k: v.half() for k, v in state_dict.items()} if args.use_safetensors: save_file(state_dict, args.checkpoint_path) else: _a = {"""state_dict""": state_dict} torch.save(state_dict, args.checkpoint_path)
78
1
"""simple docstring""" import unittest from transformers.models.xlm_prophetnet.tokenization_xlm_prophetnet import SPIECE_UNDERLINE, XLMProphetNetTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _a = get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece class _UpperCAmelCase( lowerCamelCase , unittest.TestCase ): lowercase__ = XLMProphetNetTokenizer lowercase__ = False lowercase__ = True def UpperCAmelCase ( self) -> int: '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing _UpperCamelCase = XLMProphetNetTokenizer(__a , keep_accents=__a) tokenizer.save_pretrained(self.tmpdirname) def UpperCAmelCase ( self) -> str: '''simple docstring''' _UpperCamelCase = '''[PAD]''' _UpperCamelCase = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__a) , __a) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__a) , __a) def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = list(self.get_tokenizer().get_vocab().keys()) self.assertEqual(vocab_keys[0] , '''[PAD]''') self.assertEqual(vocab_keys[1] , '''[CLS]''') self.assertEqual(vocab_keys[-1] , '''j''') self.assertEqual(len(__a) , 10_12) def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 10_12) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = XLMProphetNetTokenizer(__a , keep_accents=__a) _UpperCamelCase = tokenizer.tokenize('''This is a test''') self.assertListEqual(__a , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est''']) self.assertListEqual( tokenizer.convert_tokens_to_ids(__a) , [value + tokenizer.fairseq_offset for value in [2_85, 46, 10, 1_70, 3_82]] , ) _UpperCamelCase = tokenizer.tokenize('''I was born in 92000, and this is falsé.''') self.assertListEqual( __a , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) _UpperCamelCase = tokenizer.convert_tokens_to_ids(__a) self.assertListEqual( __a , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, -9, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, -9, 4] ] , ) _UpperCamelCase = tokenizer.convert_ids_to_tokens(__a) self.assertListEqual( __a , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''[UNK]''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''[UNK]''', '''.''', ] , ) @cached_property def UpperCAmelCase ( self) -> Any: '''simple docstring''' return XLMProphetNetTokenizer.from_pretrained('''microsoft/xprophetnet-large-wiki100-cased''') @slow def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = '''Hello World!''' _UpperCamelCase = [3_53_89, 66_72, 49, 2] self.assertListEqual(__a , self.big_tokenizer.encode(__a)) @slow def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' # fmt: off _UpperCamelCase = {'''input_ids''': [[1_10_73, 8_27_83, 18, 26, 8_27_83, 5_49, 5_15_40, 2_48, 1_72_09, 13_01, 2_17, 20, 21_51_86, 13_25, 1_47, 1_72_09, 13_01, 2_17, 20, 5_63_70, 53, 12_20_20, 20, 1_64_77, 27, 8_73_55, 45_48, 20, 47_28, 7_83_92, 17, 15_99_69, 18, 26, 2_44_91, 6_29, 15, 5_38, 2_27_04, 54_39, 15, 27_88, 2_44_91, 98_85, 15, 4_35_34, 6_05, 15, 8_14, 1_84_03, 3_32_00, 29, 15, 4_35_34, 2_44_58, 1_24_10, 1_11, 2_49_66, 8_36_69, 96_37, 14_40_68, 26, 8_50, 2_23_46, 27, 1_47, 2_49_66, 8_36_69, 8_34_90, 26, 3_91_13, 7_35, 27, 6_89, 6_56, 28_00, 13_39, 46_00, 53, 12_20_20, 11_57_85, 34, 8_16, 13_39, 4_68_87, 18, 1_47, 5_39_05, 19_51, 4_22_38, 4_11_70, 1_77_32, 8_34, 4_36, 15, 2_75_23, 9_87_33, 2_17, 1_47, 55_42, 49_81, 9_30, 1_73_47, 16, 2], [2_00_91, 6_29, 94, 8_27_86, 58, 4_90, 20, 15_28, 84, 5_39_05, 3_44, 8_05_92, 11_01_28, 1_88_22, 52_67, 13_06, 62, 15_25_37, 3_08, 79_97, 4_01, 12_44_27, 5_49, 3_54_42, 2_25, 1_09, 1_50_55, 2_57_48, 1_47, 71_19, 4_37_12, 34, 7_67, 13_53_66, 18, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [5_92, 6_37_84, 11_94_66, 17, 14_78_08, 8_82_14, 18, 6_56, 81, 32, 32_96, 1_02_80, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__a , model_name='''microsoft/xprophetnet-large-wiki100-cased''' , revision='''1acad1643ddd54a44df6a1b797ada8373685d90e''' , )
78
"""simple docstring""" import argparse import torch from transformers import OpenAIGPTConfig, OpenAIGPTModel, load_tf_weights_in_openai_gpt from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> Union[str, Any]: """simple docstring""" if openai_config_file == "": _UpperCamelCase = OpenAIGPTConfig() else: _UpperCamelCase = OpenAIGPTConfig.from_json_file(__snake_case ) _UpperCamelCase = OpenAIGPTModel(__snake_case ) # Load weights from numpy load_tf_weights_in_openai_gpt(__snake_case, __snake_case, __snake_case ) # Save pytorch-model _UpperCamelCase = pytorch_dump_folder_path + '''/''' + WEIGHTS_NAME _UpperCamelCase = pytorch_dump_folder_path + '''/''' + CONFIG_NAME print(F'''Save PyTorch model to {pytorch_weights_dump_path}''' ) torch.save(model.state_dict(), __snake_case ) print(F'''Save configuration file to {pytorch_config_dump_path}''' ) with open(__snake_case, '''w''', encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( """--openai_checkpoint_folder_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--openai_config_file""", default="""""", type=str, help=( """An optional config json file corresponding to the pre-trained OpenAI model. \n""" """This specifies the model architecture.""" ), ) _a = parser.parse_args() convert_openai_checkpoint_to_pytorch( args.openai_checkpoint_folder_path, args.openai_config_file, args.pytorch_dump_folder_path )
78
1
"""simple docstring""" import flax.linen as nn import jax.numpy as jnp from .attention_flax import FlaxTransformeraDModel from .resnet_flax import FlaxDownsampleaD, FlaxResnetBlockaD, FlaxUpsampleaD class _UpperCAmelCase( nn.Module ): lowercase__ = 42 lowercase__ = 42 lowercase__ = 0.0 lowercase__ = 1 lowercase__ = 1 lowercase__ = True lowercase__ = False lowercase__ = False lowercase__ = False lowercase__ = jnp.floataa def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = [] _UpperCamelCase = [] for i in range(self.num_layers): _UpperCamelCase = self.in_channels if i == 0 else self.out_channels _UpperCamelCase = FlaxResnetBlockaD( in_channels=__a , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(__a) _UpperCamelCase = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(__a) _UpperCamelCase = resnets _UpperCamelCase = attentions if self.add_downsample: _UpperCamelCase = FlaxDownsampleaD(self.out_channels , dtype=self.dtype) def __call__( self , __a , __a , __a , __a=True) -> Optional[int]: '''simple docstring''' _UpperCamelCase = () for resnet, attn in zip(self.resnets , self.attentions): _UpperCamelCase = resnet(__a , __a , deterministic=__a) _UpperCamelCase = attn(__a , __a , deterministic=__a) output_states += (hidden_states,) if self.add_downsample: _UpperCamelCase = self.downsamplers_a(__a) output_states += (hidden_states,) return hidden_states, output_states class _UpperCAmelCase( nn.Module ): lowercase__ = 42 lowercase__ = 42 lowercase__ = 0.0 lowercase__ = 1 lowercase__ = True lowercase__ = jnp.floataa def UpperCAmelCase ( self) -> Dict: '''simple docstring''' _UpperCamelCase = [] for i in range(self.num_layers): _UpperCamelCase = self.in_channels if i == 0 else self.out_channels _UpperCamelCase = FlaxResnetBlockaD( in_channels=__a , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(__a) _UpperCamelCase = resnets if self.add_downsample: _UpperCamelCase = FlaxDownsampleaD(self.out_channels , dtype=self.dtype) def __call__( self , __a , __a , __a=True) -> str: '''simple docstring''' _UpperCamelCase = () for resnet in self.resnets: _UpperCamelCase = resnet(__a , __a , deterministic=__a) output_states += (hidden_states,) if self.add_downsample: _UpperCamelCase = self.downsamplers_a(__a) output_states += (hidden_states,) return hidden_states, output_states class _UpperCAmelCase( nn.Module ): lowercase__ = 42 lowercase__ = 42 lowercase__ = 42 lowercase__ = 0.0 lowercase__ = 1 lowercase__ = 1 lowercase__ = True lowercase__ = False lowercase__ = False lowercase__ = False lowercase__ = jnp.floataa def UpperCAmelCase ( self) -> str: '''simple docstring''' _UpperCamelCase = [] _UpperCamelCase = [] for i in range(self.num_layers): _UpperCamelCase = self.in_channels if (i == self.num_layers - 1) else self.out_channels _UpperCamelCase = self.prev_output_channel if i == 0 else self.out_channels _UpperCamelCase = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(__a) _UpperCamelCase = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(__a) _UpperCamelCase = resnets _UpperCamelCase = attentions if self.add_upsample: _UpperCamelCase = FlaxUpsampleaD(self.out_channels , dtype=self.dtype) def __call__( self , __a , __a , __a , __a , __a=True) -> Dict: '''simple docstring''' for resnet, attn in zip(self.resnets , self.attentions): # pop res hidden states _UpperCamelCase = res_hidden_states_tuple[-1] _UpperCamelCase = res_hidden_states_tuple[:-1] _UpperCamelCase = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1) _UpperCamelCase = resnet(__a , __a , deterministic=__a) _UpperCamelCase = attn(__a , __a , deterministic=__a) if self.add_upsample: _UpperCamelCase = self.upsamplers_a(__a) return hidden_states class _UpperCAmelCase( nn.Module ): lowercase__ = 42 lowercase__ = 42 lowercase__ = 42 lowercase__ = 0.0 lowercase__ = 1 lowercase__ = True lowercase__ = jnp.floataa def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = [] for i in range(self.num_layers): _UpperCamelCase = self.in_channels if (i == self.num_layers - 1) else self.out_channels _UpperCamelCase = self.prev_output_channel if i == 0 else self.out_channels _UpperCamelCase = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(__a) _UpperCamelCase = resnets if self.add_upsample: _UpperCamelCase = FlaxUpsampleaD(self.out_channels , dtype=self.dtype) def __call__( self , __a , __a , __a , __a=True) -> Any: '''simple docstring''' for resnet in self.resnets: # pop res hidden states _UpperCamelCase = res_hidden_states_tuple[-1] _UpperCamelCase = res_hidden_states_tuple[:-1] _UpperCamelCase = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1) _UpperCamelCase = resnet(__a , __a , deterministic=__a) if self.add_upsample: _UpperCamelCase = self.upsamplers_a(__a) return hidden_states class _UpperCAmelCase( nn.Module ): lowercase__ = 42 lowercase__ = 0.0 lowercase__ = 1 lowercase__ = 1 lowercase__ = False lowercase__ = False lowercase__ = jnp.floataa def UpperCAmelCase ( self) -> Dict: '''simple docstring''' # there is always at least one resnet _UpperCamelCase = [ FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) ] _UpperCamelCase = [] for _ in range(self.num_layers): _UpperCamelCase = FlaxTransformeraDModel( in_channels=self.in_channels , n_heads=self.num_attention_heads , d_head=self.in_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(__a) _UpperCamelCase = FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(__a) _UpperCamelCase = resnets _UpperCamelCase = attentions def __call__( self , __a , __a , __a , __a=True) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.resnets[0](__a , __a) for attn, resnet in zip(self.attentions , self.resnets[1:]): _UpperCamelCase = attn(__a , __a , deterministic=__a) _UpperCamelCase = resnet(__a , __a , deterministic=__a) return hidden_states
78
"""simple docstring""" from __future__ import annotations import unittest from transformers import AutoTokenizer, MBartConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFMBartForConditionalGeneration, TFMBartModel @require_tf class _UpperCAmelCase: lowercase__ = MBartConfig lowercase__ = {} lowercase__ = 'gelu' def __init__( self , __a , __a=13 , __a=7 , __a=True , __a=False , __a=99 , __a=32 , __a=2 , __a=4 , __a=37 , __a=0.1 , __a=0.1 , __a=20 , __a=2 , __a=1 , __a=0 , ) -> Any: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = seq_length _UpperCamelCase = is_training _UpperCamelCase = use_labels _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = eos_token_id _UpperCamelCase = pad_token_id _UpperCamelCase = bos_token_id def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size) _UpperCamelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size) , 1) _UpperCamelCase = tf.concat([input_ids, eos_tensor] , axis=1) _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) _UpperCamelCase = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) _UpperCamelCase = prepare_mbart_inputs_dict(__a , __a , __a) return config, inputs_dict def UpperCAmelCase ( self , __a , __a) -> Optional[int]: '''simple docstring''' _UpperCamelCase = TFMBartModel(config=__a).get_decoder() _UpperCamelCase = inputs_dict['''input_ids'''] _UpperCamelCase = input_ids[:1, :] _UpperCamelCase = inputs_dict['''attention_mask'''][:1, :] _UpperCamelCase = inputs_dict['''head_mask'''] _UpperCamelCase = 1 # first forward pass _UpperCamelCase = model(__a , attention_mask=__a , head_mask=__a , use_cache=__a) _UpperCamelCase , _UpperCamelCase = outputs.to_tuple() _UpperCamelCase = past_key_values[1] def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case, __snake_case=None, __snake_case=None, __snake_case=None, __snake_case=None, __snake_case=None, ) -> Optional[int]: """simple docstring""" if attention_mask is None: _UpperCamelCase = tf.cast(tf.math.not_equal(__snake_case, config.pad_token_id ), tf.inta ) if decoder_attention_mask is None: _UpperCamelCase = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape, dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:], config.pad_token_id ), tf.inta ), ], axis=-1, ) if head_mask is None: _UpperCamelCase = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _UpperCamelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: _UpperCamelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class _UpperCAmelCase( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase__ = (TFMBartForConditionalGeneration, TFMBartModel) if is_tf_available() else () lowercase__ = (TFMBartForConditionalGeneration,) if is_tf_available() else () lowercase__ = ( { 'conversational': TFMBartForConditionalGeneration, 'feature-extraction': TFMBartModel, 'summarization': TFMBartForConditionalGeneration, 'text2text-generation': TFMBartForConditionalGeneration, 'translation': TFMBartForConditionalGeneration, } if is_tf_available() else {} ) lowercase__ = True lowercase__ = False lowercase__ = False def UpperCAmelCase ( self , __a , __a , __a , __a , __a) -> Dict: '''simple docstring''' if pipeline_test_casse_name != "FeatureExtractionPipelineTests": # Exception encountered when calling layer '...' return True return False def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = TFMBartModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=__a) def UpperCAmelCase ( self) -> str: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__a) @require_sentencepiece @require_tokenizers @require_tf class _UpperCAmelCase( unittest.TestCase ): lowercase__ = [ ' UN Chief Says There Is No Military Solution in Syria', ] lowercase__ = [ 'Şeful ONU declară că nu există o soluţie militară în Siria', ] lowercase__ = 'facebook/mbart-large-en-ro' @cached_property def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' return AutoTokenizer.from_pretrained(self.model_name) @cached_property def UpperCAmelCase ( self) -> str: '''simple docstring''' _UpperCamelCase = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name) return model def UpperCAmelCase ( self , **__a) -> List[str]: '''simple docstring''' _UpperCamelCase = self.translate_src_text(**__a) self.assertListEqual(self.expected_text , __a) def UpperCAmelCase ( self , **__a) -> Dict: '''simple docstring''' _UpperCamelCase = self.tokenizer(self.src_text , **__a , return_tensors='''tf''') _UpperCamelCase = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2) _UpperCamelCase = self.tokenizer.batch_decode(__a , skip_special_tokens=__a) return generated_words @slow def UpperCAmelCase ( self) -> Any: '''simple docstring''' self._assert_generated_batch_equal_expected()
78
1
"""simple docstring""" import unittest from transformers import ( MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TextGenerationPipeline, logging, pipeline, ) from transformers.testing_utils import ( CaptureLogger, is_pipeline_test, require_accelerate, require_tf, require_torch, require_torch_gpu, require_torch_or_tf, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf class _UpperCAmelCase( unittest.TestCase ): lowercase__ = MODEL_FOR_CAUSAL_LM_MAPPING lowercase__ = TF_MODEL_FOR_CAUSAL_LM_MAPPING @require_torch def UpperCAmelCase ( self) -> Dict: '''simple docstring''' _UpperCamelCase = pipeline(task='''text-generation''' , model='''sshleifer/tiny-ctrl''' , framework='''pt''') # Using `do_sample=False` to force deterministic output _UpperCamelCase = text_generator('''This is a test''' , do_sample=__a) self.assertEqual( __a , [ { '''generated_text''': ( '''This is a test ☃ ☃ segmental segmental segmental 议议eski eski flutter flutter Lacy oscope.''' ''' oscope. FiliFili@@''' ) } ] , ) _UpperCamelCase = text_generator(['''This is a test''', '''This is a second test''']) self.assertEqual( __a , [ [ { '''generated_text''': ( '''This is a test ☃ ☃ segmental segmental segmental 议议eski eski flutter flutter Lacy oscope.''' ''' oscope. FiliFili@@''' ) } ], [ { '''generated_text''': ( '''This is a second test ☃ segmental segmental segmental 议议eski eski flutter flutter Lacy''' ''' oscope. oscope. FiliFili@@''' ) } ], ] , ) _UpperCamelCase = text_generator('''This is a test''' , do_sample=__a , num_return_sequences=2 , return_tensors=__a) self.assertEqual( __a , [ {'''generated_token_ids''': ANY(__a)}, {'''generated_token_ids''': ANY(__a)}, ] , ) _UpperCamelCase = text_generator.model.config.eos_token_id _UpperCamelCase = '''<pad>''' _UpperCamelCase = text_generator( ['''This is a test''', '''This is a second test'''] , do_sample=__a , num_return_sequences=2 , batch_size=2 , return_tensors=__a , ) self.assertEqual( __a , [ [ {'''generated_token_ids''': ANY(__a)}, {'''generated_token_ids''': ANY(__a)}, ], [ {'''generated_token_ids''': ANY(__a)}, {'''generated_token_ids''': ANY(__a)}, ], ] , ) @require_tf def UpperCAmelCase ( self) -> str: '''simple docstring''' _UpperCamelCase = pipeline(task='''text-generation''' , model='''sshleifer/tiny-ctrl''' , framework='''tf''') # Using `do_sample=False` to force deterministic output _UpperCamelCase = text_generator('''This is a test''' , do_sample=__a) self.assertEqual( __a , [ { '''generated_text''': ( '''This is a test FeyFeyFey(Croatis.), s.), Cannes Cannes Cannes 閲閲Cannes Cannes Cannes 攵''' ''' please,''' ) } ] , ) _UpperCamelCase = text_generator(['''This is a test''', '''This is a second test'''] , do_sample=__a) self.assertEqual( __a , [ [ { '''generated_text''': ( '''This is a test FeyFeyFey(Croatis.), s.), Cannes Cannes Cannes 閲閲Cannes Cannes Cannes 攵''' ''' please,''' ) } ], [ { '''generated_text''': ( '''This is a second test Chieftain Chieftain prefecture prefecture prefecture Cannes Cannes''' ''' Cannes 閲閲Cannes Cannes Cannes 攵 please,''' ) } ], ] , ) def UpperCAmelCase ( self , __a , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = TextGenerationPipeline(model=__a , tokenizer=__a) return text_generator, ["This is a test", "Another test"] def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = '''Hello I believe in''' _UpperCamelCase = pipeline('''text-generation''' , model='''hf-internal-testing/tiny-random-gpt2''') _UpperCamelCase = text_generator(__a) self.assertEqual( __a , [{'''generated_text''': '''Hello I believe in fe fe fe fe fe fe fe fe fe fe fe fe'''}] , ) _UpperCamelCase = text_generator(__a , stop_sequence=''' fe''') self.assertEqual(__a , [{'''generated_text''': '''Hello I believe in fe'''}]) def UpperCAmelCase ( self , __a , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = text_generator.model _UpperCamelCase = text_generator.tokenizer _UpperCamelCase = text_generator('''This is a test''') self.assertEqual(__a , [{'''generated_text''': ANY(__a)}]) self.assertTrue(outputs[0]['''generated_text'''].startswith('''This is a test''')) _UpperCamelCase = text_generator('''This is a test''' , return_full_text=__a) self.assertEqual(__a , [{'''generated_text''': ANY(__a)}]) self.assertNotIn('''This is a test''' , outputs[0]['''generated_text''']) _UpperCamelCase = pipeline(task='''text-generation''' , model=__a , tokenizer=__a , return_full_text=__a) _UpperCamelCase = text_generator('''This is a test''') self.assertEqual(__a , [{'''generated_text''': ANY(__a)}]) self.assertNotIn('''This is a test''' , outputs[0]['''generated_text''']) _UpperCamelCase = text_generator('''This is a test''' , return_full_text=__a) self.assertEqual(__a , [{'''generated_text''': ANY(__a)}]) self.assertTrue(outputs[0]['''generated_text'''].startswith('''This is a test''')) _UpperCamelCase = text_generator(['''This is great !''', '''Something else'''] , num_return_sequences=2 , do_sample=__a) self.assertEqual( __a , [ [{'''generated_text''': ANY(__a)}, {'''generated_text''': ANY(__a)}], [{'''generated_text''': ANY(__a)}, {'''generated_text''': ANY(__a)}], ] , ) if text_generator.tokenizer.pad_token is not None: _UpperCamelCase = text_generator( ['''This is great !''', '''Something else'''] , num_return_sequences=2 , batch_size=2 , do_sample=__a) self.assertEqual( __a , [ [{'''generated_text''': ANY(__a)}, {'''generated_text''': ANY(__a)}], [{'''generated_text''': ANY(__a)}, {'''generated_text''': ANY(__a)}], ] , ) with self.assertRaises(__a): _UpperCamelCase = text_generator('''test''' , return_full_text=__a , return_text=__a) with self.assertRaises(__a): _UpperCamelCase = text_generator('''test''' , return_full_text=__a , return_tensors=__a) with self.assertRaises(__a): _UpperCamelCase = text_generator('''test''' , return_text=__a , return_tensors=__a) # Empty prompt is slighly special # it requires BOS token to exist. # Special case for Pegasus which will always append EOS so will # work even without BOS. if ( text_generator.tokenizer.bos_token_id is not None or "Pegasus" in tokenizer.__class__.__name__ or "Git" in model.__class__.__name__ ): _UpperCamelCase = text_generator('''''') self.assertEqual(__a , [{'''generated_text''': ANY(__a)}]) else: with self.assertRaises((ValueError, AssertionError)): _UpperCamelCase = text_generator('''''') if text_generator.framework == "tf": # TF generation does not support max_new_tokens, and it's impossible # to control long generation with only max_length without # fancy calculation, dismissing tests for now. return # We don't care about infinite range models. # They already work. # Skip this test for XGLM, since it uses sinusoidal positional embeddings which are resized on-the-fly. _UpperCamelCase = ['''RwkvForCausalLM''', '''XGLMForCausalLM''', '''GPTNeoXForCausalLM'''] if ( tokenizer.model_max_length < 1_00_00 and text_generator.model.__class__.__name__ not in EXTRA_MODELS_CAN_HANDLE_LONG_INPUTS ): # Handling of large generations with self.assertRaises((RuntimeError, IndexError, ValueError, AssertionError)): text_generator('''This is a test''' * 5_00 , max_new_tokens=20) _UpperCamelCase = text_generator('''This is a test''' * 5_00 , handle_long_generation='''hole''' , max_new_tokens=20) # Hole strategy cannot work with self.assertRaises(__a): text_generator( '''This is a test''' * 5_00 , handle_long_generation='''hole''' , max_new_tokens=tokenizer.model_max_length + 10 , ) @require_torch @require_accelerate @require_torch_gpu def UpperCAmelCase ( self) -> str: '''simple docstring''' import torch # Classic `model_kwargs` _UpperCamelCase = pipeline( model='''hf-internal-testing/tiny-random-bloom''' , model_kwargs={'''device_map''': '''auto''', '''torch_dtype''': torch.bfloataa} , ) self.assertEqual(pipe.model.device , torch.device(0)) self.assertEqual(pipe.model.lm_head.weight.dtype , torch.bfloataa) _UpperCamelCase = pipe('''This is a test''') self.assertEqual( __a , [ { '''generated_text''': ( '''This is a test test test test test test test test test test test test test test test test''' ''' test''' ) } ] , ) # Upgraded those two to real pipeline arguments (they just get sent for the model as they're unlikely to mean anything else.) _UpperCamelCase = pipeline(model='''hf-internal-testing/tiny-random-bloom''' , device_map='''auto''' , torch_dtype=torch.bfloataa) self.assertEqual(pipe.model.device , torch.device(0)) self.assertEqual(pipe.model.lm_head.weight.dtype , torch.bfloataa) _UpperCamelCase = pipe('''This is a test''') self.assertEqual( __a , [ { '''generated_text''': ( '''This is a test test test test test test test test test test test test test test test test''' ''' test''' ) } ] , ) # torch_dtype will be automatically set to float32 if not provided - check: https://github.com/huggingface/transformers/pull/20602 _UpperCamelCase = pipeline(model='''hf-internal-testing/tiny-random-bloom''' , device_map='''auto''') self.assertEqual(pipe.model.device , torch.device(0)) self.assertEqual(pipe.model.lm_head.weight.dtype , torch.floataa) _UpperCamelCase = pipe('''This is a test''') self.assertEqual( __a , [ { '''generated_text''': ( '''This is a test test test test test test test test test test test test test test test test''' ''' test''' ) } ] , ) @require_torch @require_torch_gpu def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' import torch _UpperCamelCase = pipeline(model='''hf-internal-testing/tiny-random-bloom''' , device=0 , torch_dtype=torch.floataa) pipe('''This is a test''') @require_torch @require_accelerate @require_torch_gpu def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' import torch _UpperCamelCase = pipeline(model='''hf-internal-testing/tiny-random-bloom''' , device_map='''auto''' , torch_dtype=torch.floataa) pipe('''This is a test''' , do_sample=__a , top_p=0.5) def UpperCAmelCase ( self) -> Dict: '''simple docstring''' _UpperCamelCase = '''Hello world''' _UpperCamelCase = pipeline('''text-generation''' , model='''hf-internal-testing/tiny-random-gpt2''') if text_generator.model.framework == "tf": _UpperCamelCase = logging.get_logger('''transformers.generation.tf_utils''') else: _UpperCamelCase = logging.get_logger('''transformers.generation.utils''') _UpperCamelCase = '''Both `max_new_tokens`''' # The beggining of the message to be checked in this test # Both are set by the user -> log warning with CaptureLogger(__a) as cl: _UpperCamelCase = text_generator(__a , max_length=10 , max_new_tokens=1) self.assertIn(__a , cl.out) # The user only sets one -> no warning with CaptureLogger(__a) as cl: _UpperCamelCase = text_generator(__a , max_new_tokens=1) self.assertNotIn(__a , cl.out) with CaptureLogger(__a) as cl: _UpperCamelCase = text_generator(__a , max_length=10) self.assertNotIn(__a , cl.out)
78
"""simple docstring""" from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _a = logging.get_logger(__name__) class _UpperCAmelCase( lowerCamelCase ): lowercase__ = ['pixel_values'] def __init__( self , __a = True , __a = 1 / 2_55 , __a = True , __a = 8 , **__a , ) -> None: '''simple docstring''' super().__init__(**__a) _UpperCamelCase = do_rescale _UpperCamelCase = rescale_factor _UpperCamelCase = do_pad _UpperCamelCase = pad_size def UpperCAmelCase ( self , __a , __a , __a = None , **__a) -> np.ndarray: '''simple docstring''' return rescale(__a , scale=__a , data_format=__a , **__a) def UpperCAmelCase ( self , __a , __a , __a = None) -> List[Any]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = get_image_size(__a) _UpperCamelCase = (old_height // size + 1) * size - old_height _UpperCamelCase = (old_width // size + 1) * size - old_width return pad(__a , ((0, pad_height), (0, pad_width)) , mode='''symmetric''' , data_format=__a) def UpperCAmelCase ( self , __a , __a = None , __a = None , __a = None , __a = None , __a = None , __a = ChannelDimension.FIRST , **__a , ) -> Tuple: '''simple docstring''' _UpperCamelCase = do_rescale if do_rescale is not None else self.do_rescale _UpperCamelCase = rescale_factor if rescale_factor is not None else self.rescale_factor _UpperCamelCase = do_pad if do_pad is not None else self.do_pad _UpperCamelCase = pad_size if pad_size is not None else self.pad_size _UpperCamelCase = make_list_of_images(__a) if not valid_images(__a): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''') if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''') # All transformations expect numpy arrays. _UpperCamelCase = [to_numpy_array(__a) for image in images] if do_rescale: _UpperCamelCase = [self.rescale(image=__a , scale=__a) for image in images] if do_pad: _UpperCamelCase = [self.pad(__a , size=__a) for image in images] _UpperCamelCase = [to_channel_dimension_format(__a , __a) for image in images] _UpperCamelCase = {'''pixel_values''': images} return BatchFeature(data=__a , tensor_type=__a)
78
1
"""simple docstring""" import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import DetrConfig, MaskFormerConfig, SwinConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskFormerForInstanceSegmentation, MaskFormerModel if is_vision_available(): from transformers import MaskFormerImageProcessor if is_vision_available(): from PIL import Image class _UpperCAmelCase: def __init__( self , __a , __a=2 , __a=True , __a=False , __a=10 , __a=3 , __a=32 * 4 , __a=32 * 6 , __a=4 , __a=32 , ) -> Dict: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = is_training _UpperCamelCase = use_auxiliary_loss _UpperCamelCase = num_queries _UpperCamelCase = num_channels _UpperCamelCase = min_size _UpperCamelCase = max_size _UpperCamelCase = num_labels _UpperCamelCase = mask_feature_size def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size]).to( __a) _UpperCamelCase = torch.ones([self.batch_size, self.min_size, self.max_size] , device=__a) _UpperCamelCase = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=__a) > 0.5 ).float() _UpperCamelCase = (torch.rand((self.batch_size, self.num_labels) , device=__a) > 0.5).long() _UpperCamelCase = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def UpperCAmelCase ( self) -> int: '''simple docstring''' return MaskFormerConfig.from_backbone_and_decoder_configs( backbone_config=SwinConfig( depths=[1, 1, 1, 1] , ) , decoder_config=DetrConfig( decoder_ffn_dim=1_28 , num_queries=self.num_queries , decoder_attention_heads=2 , d_model=self.mask_feature_size , ) , mask_feature_size=self.mask_feature_size , fpn_feature_size=self.mask_feature_size , num_channels=self.num_channels , num_labels=self.num_labels , ) def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = self.prepare_config_and_inputs() _UpperCamelCase = {'''pixel_values''': pixel_values, '''pixel_mask''': pixel_mask} return config, inputs_dict def UpperCAmelCase ( self , __a , __a) -> Optional[int]: '''simple docstring''' _UpperCamelCase = output.encoder_hidden_states _UpperCamelCase = output.pixel_decoder_hidden_states _UpperCamelCase = output.transformer_decoder_hidden_states self.parent.assertTrue(len(__a) , len(config.backbone_config.depths)) self.parent.assertTrue(len(__a) , len(config.backbone_config.depths)) self.parent.assertTrue(len(__a) , config.decoder_config.decoder_layers) def UpperCAmelCase ( self , __a , __a , __a , __a=False) -> Tuple: '''simple docstring''' with torch.no_grad(): _UpperCamelCase = MaskFormerModel(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(pixel_values=__a , pixel_mask=__a) _UpperCamelCase = model(__a , output_hidden_states=__a) # the correct shape of output.transformer_decoder_hidden_states ensure the correcteness of the # encoder and pixel decoder self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.mask_feature_size) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None) self.parent.assertTrue(output.encoder_last_hidden_state is not None) if output_hidden_states: self.check_output_hidden_state(__a , __a) def UpperCAmelCase ( self , __a , __a , __a , __a , __a) -> Dict: '''simple docstring''' _UpperCamelCase = MaskFormerForInstanceSegmentation(config=__a) model.to(__a) model.eval() def comm_check_on_output(__a): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None) self.parent.assertTrue(result.encoder_last_hidden_state is not None) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1)) with torch.no_grad(): _UpperCamelCase = model(pixel_values=__a , pixel_mask=__a) _UpperCamelCase = model(__a) comm_check_on_output(__a) _UpperCamelCase = model( pixel_values=__a , pixel_mask=__a , mask_labels=__a , class_labels=__a) comm_check_on_output(__a) self.parent.assertTrue(result.loss is not None) self.parent.assertEqual(result.loss.shape , torch.Size([1])) @require_torch class _UpperCAmelCase( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase__ = (MaskFormerModel, MaskFormerForInstanceSegmentation) if is_torch_available() else () lowercase__ = ( {'feature-extraction': MaskFormerModel, 'image-segmentation': MaskFormerForInstanceSegmentation} if is_torch_available() else {} ) lowercase__ = False lowercase__ = False lowercase__ = False lowercase__ = False def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = MaskFormerModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=__a , has_text_modality=__a) def UpperCAmelCase ( self) -> Any: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(__a , **__a , output_hidden_states=__a) def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskformer_instance_segmentation_head_model(*__a) @unittest.skip(reason='''MaskFormer does not use inputs_embeds''') def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' pass @unittest.skip(reason='''MaskFormer does not have a get_input_embeddings method''') def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' pass @unittest.skip(reason='''MaskFormer is not a generative model''') def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' pass @unittest.skip(reason='''MaskFormer does not use token embeddings''') def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip( reason='''MaskFormer has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`''') def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''') def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' pass def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = model_class(__a) _UpperCamelCase = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCamelCase = [*signature.parameters.keys()] _UpperCamelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , __a) @slow def UpperCAmelCase ( self) -> Dict: '''simple docstring''' for model_name in ["facebook/maskformer-swin-small-coco"]: _UpperCamelCase = MaskFormerModel.from_pretrained(__a) self.assertIsNotNone(__a) def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = (self.model_tester.min_size,) * 2 _UpperCamelCase = { '''pixel_values''': torch.randn((2, 3, *size) , device=__a), '''mask_labels''': torch.randn((2, 10, *size) , device=__a), '''class_labels''': torch.zeros(2 , 10 , device=__a).long(), } _UpperCamelCase = MaskFormerForInstanceSegmentation(MaskFormerConfig()).to(__a) _UpperCamelCase = model(**__a) self.assertTrue(outputs.loss is not None) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(__a , **__a , output_hidden_states=__a) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = model_class(__a).to(__a) _UpperCamelCase = model(**__a , output_attentions=__a) self.assertTrue(outputs.attentions is not None) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' if not self.model_tester.is_training: return # only MaskFormerForInstanceSegmentation has the loss _UpperCamelCase = self.all_model_classes[1] _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs() _UpperCamelCase = model_class(__a) model.to(__a) model.train() _UpperCamelCase = model(__a , mask_labels=__a , class_labels=__a).loss loss.backward() def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' # only MaskFormerForInstanceSegmentation has the loss _UpperCamelCase = self.all_model_classes[1] _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs() _UpperCamelCase = True _UpperCamelCase = True _UpperCamelCase = model_class(__a) model.to(__a) model.train() _UpperCamelCase = model(__a , mask_labels=__a , class_labels=__a) _UpperCamelCase = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() _UpperCamelCase = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() # we requires_grad=True in inputs_embeds (line 2152), the original implementation don't _UpperCamelCase = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() _UpperCamelCase = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=__a) self.assertIsNotNone(encoder_hidden_states.grad) self.assertIsNotNone(pixel_decoder_hidden_states.grad) self.assertIsNotNone(transformer_decoder_hidden_states.grad) self.assertIsNotNone(attentions.grad) _a = 1E-4 def lowerCamelCase__ ( ) -> Optional[int]: """simple docstring""" _UpperCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_vision @slow class _UpperCAmelCase( unittest.TestCase ): @cached_property def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' return ( MaskFormerImageProcessor.from_pretrained('''facebook/maskformer-swin-small-coco''') if is_vision_available() else None ) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = MaskFormerModel.from_pretrained('''facebook/maskformer-swin-small-coco''').to(__a) _UpperCamelCase = self.default_image_processor _UpperCamelCase = prepare_img() _UpperCamelCase = image_processor(__a , return_tensors='''pt''').to(__a) _UpperCamelCase = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0) # check size self.assertEqual(__a , (1, 3, 8_00, 10_88)) with torch.no_grad(): _UpperCamelCase = model(**__a) _UpperCamelCase = torch.tensor( [[-0.0482, 0.9228, 0.4951], [-0.2547, 0.8017, 0.8527], [-0.0069, 0.3385, -0.0089]]).to(__a) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , __a , atol=__a)) _UpperCamelCase = torch.tensor( [[-0.8422, -0.8434, -0.9718], [-1.0144, -0.5565, -0.4195], [-1.0038, -0.4484, -0.1961]]).to(__a) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , __a , atol=__a)) _UpperCamelCase = torch.tensor( [[0.2852, -0.0159, 0.9735], [0.6254, 0.1858, 0.8529], [-0.0680, -0.4116, 1.8413]]).to(__a) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , __a , atol=__a)) def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-swin-small-coco''') .to(__a) .eval() ) _UpperCamelCase = self.default_image_processor _UpperCamelCase = prepare_img() _UpperCamelCase = image_processor(__a , return_tensors='''pt''').to(__a) _UpperCamelCase = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0) # check size self.assertEqual(__a , (1, 3, 8_00, 10_88)) with torch.no_grad(): _UpperCamelCase = model(**__a) # masks_queries_logits _UpperCamelCase = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) _UpperCamelCase = [ [-1.373_7124, -1.772_4937, -1.936_4233], [-1.597_7281, -1.986_7939, -2.152_3695], [-1.579_5398, -1.926_9832, -2.09_3942], ] _UpperCamelCase = torch.tensor(__a).to(__a) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , __a , atol=__a)) # class_queries_logits _UpperCamelCase = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1)) _UpperCamelCase = torch.tensor( [ [1.6_512e00, -5.2_572e00, -3.3_519e00], [3.6_169e-02, -5.9_025e00, -2.9_313e00], [1.0_766e-04, -7.7_630e00, -5.1_263e00], ]).to(__a) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , __a , atol=__a)) def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-resnet101-coco-stuff''') .to(__a) .eval() ) _UpperCamelCase = self.default_image_processor _UpperCamelCase = prepare_img() _UpperCamelCase = image_processor(__a , return_tensors='''pt''').to(__a) _UpperCamelCase = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0) # check size self.assertEqual(__a , (1, 3, 8_00, 10_88)) with torch.no_grad(): _UpperCamelCase = model(**__a) # masks_queries_logits _UpperCamelCase = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) _UpperCamelCase = [[-0.9046, -2.6366, -4.6062], [-3.4179, -5.7890, -8.8057], [-4.9179, -7.6560, -10.7711]] _UpperCamelCase = torch.tensor(__a).to(__a) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , __a , atol=__a)) # class_queries_logits _UpperCamelCase = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1)) _UpperCamelCase = torch.tensor( [[4.7188, -3.2585, -2.8857], [6.6871, -2.9181, -1.2487], [7.2449, -2.2764, -2.1874]]).to(__a) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , __a , atol=__a)) def UpperCAmelCase ( self) -> Dict: '''simple docstring''' _UpperCamelCase = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-swin-small-coco''') .to(__a) .eval() ) _UpperCamelCase = self.default_image_processor _UpperCamelCase = image_processor( [np.zeros((3, 8_00, 13_33)), np.zeros((3, 8_00, 13_33))] , segmentation_maps=[np.zeros((3_84, 3_84)).astype(np.floataa), np.zeros((3_84, 3_84)).astype(np.floataa)] , return_tensors='''pt''' , ) _UpperCamelCase = inputs['''pixel_values'''].to(__a) _UpperCamelCase = [el.to(__a) for el in inputs['''mask_labels''']] _UpperCamelCase = [el.to(__a) for el in inputs['''class_labels''']] with torch.no_grad(): _UpperCamelCase = model(**__a) self.assertTrue(outputs.loss is not None)
78
"""simple docstring""" from importlib import import_module from .logging import get_logger _a = get_logger(__name__) class _UpperCAmelCase: def __init__( self , __a , __a=None) -> Dict: '''simple docstring''' _UpperCamelCase = attrs or [] if module is not None: for key in module.__dict__: if key in attrs or not key.startswith('''__'''): setattr(self , __a , getattr(__a , __a)) _UpperCamelCase = module._original_module if isinstance(__a , _PatchedModuleObj) else module class _UpperCAmelCase: lowercase__ = [] def __init__( self , __a , __a , __a , __a=None) -> List[str]: '''simple docstring''' _UpperCamelCase = obj _UpperCamelCase = target _UpperCamelCase = new _UpperCamelCase = target.split('''.''')[0] _UpperCamelCase = {} _UpperCamelCase = attrs or [] def __enter__( self) -> int: '''simple docstring''' *_UpperCamelCase , _UpperCamelCase = self.target.split('''.''') # Patch modules: # it's used to patch attributes of submodules like "os.path.join"; # in this case we need to patch "os" and "os.path" for i in range(len(__a)): try: _UpperCamelCase = import_module('''.'''.join(submodules[: i + 1])) except ModuleNotFoundError: continue # We iterate over all the globals in self.obj in case we find "os" or "os.path" for attr in self.obj.__dir__(): _UpperCamelCase = getattr(self.obj , __a) # We don't check for the name of the global, but rather if its value *is* "os" or "os.path". # This allows to patch renamed modules like "from os import path as ospath". if obj_attr is submodule or ( (isinstance(__a , _PatchedModuleObj) and obj_attr._original_module is submodule) ): _UpperCamelCase = obj_attr # patch at top level setattr(self.obj , __a , _PatchedModuleObj(__a , attrs=self.attrs)) _UpperCamelCase = getattr(self.obj , __a) # construct lower levels patches for key in submodules[i + 1 :]: setattr(__a , __a , _PatchedModuleObj(getattr(__a , __a , __a) , attrs=self.attrs)) _UpperCamelCase = getattr(__a , __a) # finally set the target attribute setattr(__a , __a , self.new) # Patch attribute itself: # it's used for builtins like "open", # and also to patch "os.path.join" we may also need to patch "join" # itself if it was imported as "from os.path import join". if submodules: # if it's an attribute of a submodule like "os.path.join" try: _UpperCamelCase = getattr(import_module('''.'''.join(__a)) , __a) except (AttributeError, ModuleNotFoundError): return # We iterate over all the globals in self.obj in case we find "os.path.join" for attr in self.obj.__dir__(): # We don't check for the name of the global, but rather if its value *is* "os.path.join". # This allows to patch renamed attributes like "from os.path import join as pjoin". if getattr(self.obj , __a) is attr_value: _UpperCamelCase = getattr(self.obj , __a) setattr(self.obj , __a , self.new) elif target_attr in globals()["__builtins__"]: # if it'a s builtin like "open" _UpperCamelCase = globals()['''__builtins__'''][target_attr] setattr(self.obj , __a , self.new) else: raise RuntimeError(F'''Tried to patch attribute {target_attr} instead of a submodule.''') def __exit__( self , *__a) -> Tuple: '''simple docstring''' for attr in list(self.original): setattr(self.obj , __a , self.original.pop(__a)) def UpperCAmelCase ( self) -> Dict: '''simple docstring''' self.__enter__() self._active_patches.append(self) def UpperCAmelCase ( self) -> str: '''simple docstring''' try: self._active_patches.remove(self) except ValueError: # If the patch hasn't been started this will fail return None return self.__exit__()
78
1
"""simple docstring""" import inspect import os import unittest import torch import accelerate from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_multi_gpu from accelerate.utils import patch_environment class _UpperCAmelCase( unittest.TestCase ): def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = inspect.getfile(accelerate.test_utils) _UpperCamelCase = os.path.sep.join(mod_file.split(os.path.sep)[:-1] + ['''scripts''', '''test_script.py''']) _UpperCamelCase = os.path.sep.join( mod_file.split(os.path.sep)[:-1] + ['''scripts''', '''test_distributed_data_loop.py''']) _UpperCamelCase = os.path.sep.join(mod_file.split(os.path.sep)[:-1] + ['''scripts''', '''test_ops.py''']) @require_multi_gpu def UpperCAmelCase ( self) -> int: '''simple docstring''' print(F'''Found {torch.cuda.device_count()} devices.''') _UpperCamelCase = ['''torchrun''', F'''--nproc_per_node={torch.cuda.device_count()}''', self.test_file_path] with patch_environment(omp_num_threads=1): execute_subprocess_async(__a , env=os.environ.copy()) @require_multi_gpu def UpperCAmelCase ( self) -> str: '''simple docstring''' print(F'''Found {torch.cuda.device_count()} devices.''') _UpperCamelCase = ['''torchrun''', F'''--nproc_per_node={torch.cuda.device_count()}''', self.operation_file_path] print(F'''Command: {cmd}''') with patch_environment(omp_num_threads=1): execute_subprocess_async(__a , env=os.environ.copy()) @require_multi_gpu def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = ['''torchrun''', F'''--nproc_per_node={torch.cuda.device_count()}''', inspect.getfile(self.__class__)] with patch_environment(omp_num_threads=1): execute_subprocess_async(__a , env=os.environ.copy()) @require_multi_gpu def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' print(F'''Found {torch.cuda.device_count()} devices, using 2 devices only''') _UpperCamelCase = ['''torchrun''', F'''--nproc_per_node={torch.cuda.device_count()}''', self.data_loop_file_path] with patch_environment(omp_num_threads=1 , cuda_visible_devices='''0,1'''): execute_subprocess_async(__a , env=os.environ.copy()) if __name__ == "__main__": _a = Accelerator() _a = (accelerator.state.process_index + 2, 10) _a = torch.randint(0, 10, shape).to(accelerator.device) _a = """""" _a = accelerator.pad_across_processes(tensor) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." if not torch.equal(tensora[: accelerator.state.process_index + 2], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[accelerator.state.process_index + 2 :] == 0): error_msg += "Padding was not done with the right value (0)." _a = accelerator.pad_across_processes(tensor, pad_first=True) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." _a = accelerator.state.num_processes - accelerator.state.process_index - 1 if not torch.equal(tensora[index:], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[:index] == 0): error_msg += "Padding was not done with the right value (0)." # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
78
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_flax_available, is_torch_available, is_transformers_available, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .multicontrolnet import MultiControlNetModel from .pipeline_controlnet import StableDiffusionControlNetPipeline from .pipeline_controlnet_imgaimg import StableDiffusionControlNetImgaImgPipeline from .pipeline_controlnet_inpaint import StableDiffusionControlNetInpaintPipeline if is_transformers_available() and is_flax_available(): from .pipeline_flax_controlnet import FlaxStableDiffusionControlNetPipeline
78
1
"""simple docstring""" import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" _UpperCamelCase = filter(lambda __snake_case : p.requires_grad, model.parameters() ) _UpperCamelCase = sum([np.prod(p.size() ) for p in model_parameters] ) return params _a = logging.getLogger(__name__) def lowerCamelCase__ ( __snake_case, __snake_case ) -> Union[str, Any]: """simple docstring""" if metric == "rouge2": _UpperCamelCase = '''{val_avg_rouge2:.4f}-{step_count}''' elif metric == "bleu": _UpperCamelCase = '''{val_avg_bleu:.4f}-{step_count}''' elif metric == "em": _UpperCamelCase = '''{val_avg_em:.4f}-{step_count}''' else: raise NotImplementedError( F'''seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this''' ''' function.''' ) _UpperCamelCase = ModelCheckpoint( dirpath=__snake_case, filename=__snake_case, monitor=F'''val_{metric}''', mode='''max''', save_top_k=3, every_n_epochs=1, ) return checkpoint_callback def lowerCamelCase__ ( __snake_case, __snake_case ) -> List[str]: """simple docstring""" return EarlyStopping( monitor=F'''val_{metric}''', mode='''min''' if '''loss''' in metric else '''max''', patience=__snake_case, verbose=__snake_case, ) class _UpperCAmelCase( pl.Callback ): def UpperCAmelCase ( self , __a , __a) -> int: '''simple docstring''' _UpperCamelCase = {F'''lr_group_{i}''': param['''lr'''] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups)} pl_module.logger.log_metrics(__a) @rank_zero_only def UpperCAmelCase ( self , __a , __a , __a , __a=True) -> None: '''simple docstring''' logger.info(F'''***** {type_path} results at step {trainer.global_step:05d} *****''') _UpperCamelCase = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ['''log''', '''progress_bar''', '''preds''']}) # Log results _UpperCamelCase = Path(pl_module.hparams.output_dir) if type_path == "test": _UpperCamelCase = od / '''test_results.txt''' _UpperCamelCase = od / '''test_generations.txt''' else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. _UpperCamelCase = od / F'''{type_path}_results/{trainer.global_step:05d}.txt''' _UpperCamelCase = od / F'''{type_path}_generations/{trainer.global_step:05d}.txt''' results_file.parent.mkdir(exist_ok=__a) generations_file.parent.mkdir(exist_ok=__a) with open(__a , '''a+''') as writer: for key in sorted(__a): if key in ["log", "progress_bar", "preds"]: continue _UpperCamelCase = metrics[key] if isinstance(__a , torch.Tensor): _UpperCamelCase = val.item() _UpperCamelCase = F'''{key}: {val:.6f}\n''' writer.write(__a) if not save_generations: return if "preds" in metrics: _UpperCamelCase = '''\n'''.join(metrics['''preds''']) generations_file.open('''w+''').write(__a) @rank_zero_only def UpperCAmelCase ( self , __a , __a) -> List[str]: '''simple docstring''' try: _UpperCamelCase = pl_module.model.model.num_parameters() except AttributeError: _UpperCamelCase = pl_module.model.num_parameters() _UpperCamelCase = count_trainable_parameters(__a) # mp stands for million parameters trainer.logger.log_metrics({'''n_params''': npars, '''mp''': npars / 1e6, '''grad_mp''': n_trainable_pars / 1e6}) @rank_zero_only def UpperCAmelCase ( self , __a , __a) -> int: '''simple docstring''' save_json(pl_module.metrics , pl_module.metrics_save_path) return self._write_logs(__a , __a , '''test''') @rank_zero_only def UpperCAmelCase ( self , __a , __a) -> Any: '''simple docstring''' save_json(pl_module.metrics , pl_module.metrics_save_path) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
78
"""simple docstring""" from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging _a = logging.get_logger(__name__) _a = { """EleutherAI/gpt-neo-1.3B""": """https://huggingface.co/EleutherAI/gpt-neo-1.3B/resolve/main/config.json""", # See all GPTNeo models at https://huggingface.co/models?filter=gpt_neo } class _UpperCAmelCase( lowerCamelCase ): lowercase__ = 'gpt_neo' lowercase__ = ['past_key_values'] lowercase__ = {'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers'} def __init__( self , __a=5_02_57 , __a=20_48 , __a=20_48 , __a=24 , __a=[[["global", "local"], 12]] , __a=16 , __a=None , __a=2_56 , __a="gelu_new" , __a=0.0 , __a=0.0 , __a=0.0 , __a=0.1 , __a=1e-5 , __a=0.02 , __a=True , __a=5_02_56 , __a=5_02_56 , **__a , ) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = vocab_size _UpperCamelCase = max_position_embeddings _UpperCamelCase = hidden_size _UpperCamelCase = num_layers _UpperCamelCase = num_heads _UpperCamelCase = intermediate_size _UpperCamelCase = window_size _UpperCamelCase = activation_function _UpperCamelCase = resid_dropout _UpperCamelCase = embed_dropout _UpperCamelCase = attention_dropout _UpperCamelCase = classifier_dropout _UpperCamelCase = layer_norm_epsilon _UpperCamelCase = initializer_range _UpperCamelCase = use_cache _UpperCamelCase = bos_token_id _UpperCamelCase = eos_token_id _UpperCamelCase = attention_types _UpperCamelCase = self.expand_attention_types_params(__a) if len(self.attention_layers) != self.num_layers: raise ValueError( '''Configuration for convolutional module is incorrect. ''' '''It is required that `len(config.attention_layers)` == `config.num_layers` ''' F'''but is `len(config.attention_layers) = {len(self.attention_layers)}`, ''' F'''`config.num_layers = {self.num_layers}`. ''' '''`config.attention_layers` is prepared using `config.attention_types`. ''' '''Please verify the value of `config.attention_types` argument.''') super().__init__(bos_token_id=__a , eos_token_id=__a , **__a) @staticmethod def UpperCAmelCase ( __a) -> int: '''simple docstring''' _UpperCamelCase = [] for item in attention_types: for _ in range(item[1]): attentions.extend(item[0]) return attentions def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case, __snake_case ) -> str: """simple docstring""" import torch _UpperCamelCase = input.size() _UpperCamelCase = len(__snake_case ) _UpperCamelCase = shape[dimension] _UpperCamelCase = torch.arange(0, __snake_case, __snake_case ) _UpperCamelCase = torch.div(sizedim - size, __snake_case, rounding_mode='''floor''' ) + 1 _UpperCamelCase = torch.arange(__snake_case ) + low_indices[:min_length][:, None] _UpperCamelCase = [slice(__snake_case )] * rank _UpperCamelCase = indices _UpperCamelCase = input[s] _UpperCamelCase = list(range(0, rank + 1 ) ) perm.append(perm.pop(dimension + 1 ) ) return sliced.permute(__snake_case ) def lowerCamelCase__ ( __snake_case, __snake_case ) -> str: """simple docstring""" import torch _UpperCamelCase = torch.arange(1, __snake_case ) _UpperCamelCase = torch.remainder(__snake_case, __snake_case ) _UpperCamelCase = remainders == 0 _UpperCamelCase = candidates[divisor_indices] _UpperCamelCase = torch.max(__snake_case ) return largest_divisor, torch.div(__snake_case, __snake_case, rounding_mode='''floor''' ) class _UpperCAmelCase( lowerCamelCase ): @property def UpperCAmelCase ( self) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' _UpperCamelCase = OrderedDict({'''input_ids''': {0: '''batch''', 1: '''sequence'''}}) if self.use_past: self.fill_with_past_key_values_(__a , direction='''inputs''') _UpperCamelCase = {0: '''batch''', 1: '''past_sequence + sequence'''} else: _UpperCamelCase = {0: '''batch''', 1: '''sequence'''} return common_inputs @property def UpperCAmelCase ( self) -> int: '''simple docstring''' return self._config.num_heads def UpperCAmelCase ( self , __a , __a = -1 , __a = -1 , __a = False , __a = None , ) -> Mapping[str, Any]: '''simple docstring''' _UpperCamelCase = super(__a , self).generate_dummy_inputs( __a , batch_size=__a , seq_length=__a , is_pair=__a , framework=__a) # We need to order the input in the way they appears in the forward() _UpperCamelCase = OrderedDict({'''input_ids''': common_inputs['''input_ids''']}) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''') else: import torch _UpperCamelCase , _UpperCamelCase = common_inputs['''input_ids'''].shape # Not using the same length for past_key_values _UpperCamelCase = seqlen + 2 _UpperCamelCase = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) _UpperCamelCase = [ (torch.zeros(__a), torch.zeros(__a)) for _ in range(self.num_layers) ] _UpperCamelCase = common_inputs['''attention_mask'''] if self.use_past: _UpperCamelCase = ordered_inputs['''attention_mask'''].dtype _UpperCamelCase = torch.cat( [ordered_inputs['''attention_mask'''], torch.ones(__a , __a , dtype=__a)] , dim=1) return ordered_inputs @property def UpperCAmelCase ( self) -> int: '''simple docstring''' return 13
78
1
"""simple docstring""" import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class _UpperCAmelCase( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase__ = IFInpaintingSuperResolutionPipeline lowercase__ = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'width', 'height'} lowercase__ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({'original_image'} ) lowercase__ = PipelineTesterMixin.required_optional_params - {'latents'} def UpperCAmelCase ( self) -> int: '''simple docstring''' return self._get_superresolution_dummy_components() def UpperCAmelCase ( self , __a , __a=0) -> Optional[Any]: '''simple docstring''' if str(__a).startswith('''mps'''): _UpperCamelCase = torch.manual_seed(__a) else: _UpperCamelCase = torch.Generator(device=__a).manual_seed(__a) _UpperCamelCase = floats_tensor((1, 3, 16, 16) , rng=random.Random(__a)).to(__a) _UpperCamelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(__a)).to(__a) _UpperCamelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(__a)).to(__a) _UpperCamelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''original_image''': original_image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' , reason='''float16 requires CUDA''') def UpperCAmelCase ( self) -> Dict: '''simple docstring''' # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1e-1) def UpperCAmelCase ( self) -> int: '''simple docstring''' self._test_attention_slicing_forward_pass(expected_max_diff=1e-2) def UpperCAmelCase ( self) -> Dict: '''simple docstring''' self._test_save_load_local() def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' self._test_inference_batch_single_identical( expected_max_diff=1e-2 , )
78
"""simple docstring""" import sys from collections import defaultdict class _UpperCAmelCase: def __init__( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = [] def UpperCAmelCase ( self , __a) -> Optional[Any]: '''simple docstring''' return self.node_position[vertex] def UpperCAmelCase ( self , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = pos def UpperCAmelCase ( self , __a , __a , __a , __a) -> Tuple: '''simple docstring''' if start > size // 2 - 1: return else: if 2 * start + 2 >= size: _UpperCamelCase = 2 * start + 1 else: if heap[2 * start + 1] < heap[2 * start + 2]: _UpperCamelCase = 2 * start + 1 else: _UpperCamelCase = 2 * start + 2 if heap[smallest_child] < heap[start]: _UpperCamelCase , _UpperCamelCase = heap[smallest_child], positions[smallest_child] _UpperCamelCase , _UpperCamelCase = ( heap[start], positions[start], ) _UpperCamelCase , _UpperCamelCase = temp, tempa _UpperCamelCase = self.get_position(positions[smallest_child]) self.set_position( positions[smallest_child] , self.get_position(positions[start])) self.set_position(positions[start] , __a) self.top_to_bottom(__a , __a , __a , __a) def UpperCAmelCase ( self , __a , __a , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = position[index] while index != 0: _UpperCamelCase = int((index - 2) / 2) if index % 2 == 0 else int((index - 1) / 2) if val < heap[parent]: _UpperCamelCase = heap[parent] _UpperCamelCase = position[parent] self.set_position(position[parent] , __a) else: _UpperCamelCase = val _UpperCamelCase = temp self.set_position(__a , __a) break _UpperCamelCase = parent else: _UpperCamelCase = val _UpperCamelCase = temp self.set_position(__a , 0) def UpperCAmelCase ( self , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = len(__a) // 2 - 1 for i in range(__a , -1 , -1): self.top_to_bottom(__a , __a , len(__a) , __a) def UpperCAmelCase ( self , __a , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = positions[0] _UpperCamelCase = sys.maxsize self.top_to_bottom(__a , 0 , len(__a) , __a) return temp def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" _UpperCamelCase = Heap() _UpperCamelCase = [0] * len(__snake_case ) _UpperCamelCase = [-1] * len(__snake_case ) # Neighboring Tree Vertex of selected vertex # Minimum Distance of explored vertex with neighboring vertex of partial tree # formed in graph _UpperCamelCase = [] # Heap of Distance of vertices from their neighboring vertex _UpperCamelCase = [] for vertex in range(len(__snake_case ) ): distance_tv.append(sys.maxsize ) positions.append(__snake_case ) heap.node_position.append(__snake_case ) _UpperCamelCase = [] _UpperCamelCase = 1 _UpperCamelCase = sys.maxsize for neighbor, distance in adjacency_list[0]: _UpperCamelCase = 0 _UpperCamelCase = distance heap.heapify(__snake_case, __snake_case ) for _ in range(1, len(__snake_case ) ): _UpperCamelCase = heap.delete_minimum(__snake_case, __snake_case ) if visited[vertex] == 0: tree_edges.append((nbr_tv[vertex], vertex) ) _UpperCamelCase = 1 for neighbor, distance in adjacency_list[vertex]: if ( visited[neighbor] == 0 and distance < distance_tv[heap.get_position(__snake_case )] ): _UpperCamelCase = distance heap.bottom_to_top( __snake_case, heap.get_position(__snake_case ), __snake_case, __snake_case ) _UpperCamelCase = vertex return tree_edges if __name__ == "__main__": # pragma: no cover # < --------- Prims Algorithm --------- > _a = int(input("""Enter number of edges: """).strip()) _a = defaultdict(list) for _ in range(edges_number): _a = [int(x) for x in input().strip().split()] adjacency_list[edge[0]].append([edge[1], edge[2]]) adjacency_list[edge[1]].append([edge[0], edge[2]]) print(prisms_algorithm(adjacency_list))
78
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available, is_vision_available, ) _a = {"""configuration_beit""": ["""BEIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BeitConfig""", """BeitOnnxConfig"""]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ["""BeitFeatureExtractor"""] _a = ["""BeitImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ """BEIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """BeitForImageClassification""", """BeitForMaskedImageModeling""", """BeitForSemanticSegmentation""", """BeitModel""", """BeitPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ """FlaxBeitForImageClassification""", """FlaxBeitForMaskedImageModeling""", """FlaxBeitModel""", """FlaxBeitPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_beit import BEIT_PRETRAINED_CONFIG_ARCHIVE_MAP, BeitConfig, BeitOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_beit import BeitFeatureExtractor from .image_processing_beit import BeitImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_beit import ( BEIT_PRETRAINED_MODEL_ARCHIVE_LIST, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation, BeitModel, BeitPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_beit import ( FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling, FlaxBeitModel, FlaxBeitPreTrainedModel, ) else: import sys _a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
78
"""simple docstring""" import json import sys def lowerCamelCase__ ( __snake_case, __snake_case ) -> Union[str, Any]: """simple docstring""" with open(__snake_case, encoding='''utf-8''' ) as f: _UpperCamelCase = json.load(__snake_case ) _UpperCamelCase = ['''<details>''', '''<summary>Show updated benchmarks!</summary>''', ''' '''] for benchmark_name in sorted(__snake_case ): _UpperCamelCase = results[benchmark_name] _UpperCamelCase = benchmark_name.split('''/''' )[-1] output_md.append(F'''### Benchmark: {benchmark_file_name}''' ) _UpperCamelCase = '''| metric |''' _UpperCamelCase = '''|--------|''' _UpperCamelCase = '''| new / old (diff) |''' for metric_name in sorted(__snake_case ): _UpperCamelCase = benchmark_res[metric_name] _UpperCamelCase = metric_vals['''new'''] _UpperCamelCase = metric_vals.get('''old''', __snake_case ) _UpperCamelCase = metric_vals.get('''diff''', __snake_case ) _UpperCamelCase = F''' {new_val:f}''' if isinstance(__snake_case, (int, float) ) else '''None''' if old_val is not None: val_str += F''' / {old_val:f}''' if isinstance(__snake_case, (int, float) ) else "None" if dif_val is not None: val_str += F''' ({dif_val:f})''' if isinstance(__snake_case, (int, float) ) else "None" title += " " + metric_name + " |" lines += "---|" value += val_str + " |" output_md += [title, lines, value, " "] output_md.append('''</details>''' ) with open(__snake_case, '''w''', encoding='''utf-8''' ) as f: f.writelines('''\n'''.join(__snake_case ) ) if __name__ == "__main__": _a = sys.argv[1] _a = sys.argv[2] format_json_to_md(input_json_file, output_md_file)
78
1
"""simple docstring""" from typing import List, Union import numpy as np from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_DEPTH_ESTIMATION_MAPPING _a = logging.get_logger(__name__) @add_end_docstrings(lowerCamelCase ) class _UpperCAmelCase( lowerCamelCase ): def __init__( self , *__a , **__a) -> str: '''simple docstring''' super().__init__(*__a , **__a) requires_backends(self , '''vision''') self.check_model_type(__a) def __call__( self , __a , **__a) -> Any: '''simple docstring''' return super().__call__(__a , **__a) def UpperCAmelCase ( self , **__a) -> Any: '''simple docstring''' return {}, {}, {} def UpperCAmelCase ( self , __a) -> str: '''simple docstring''' _UpperCamelCase = load_image(__a) _UpperCamelCase = image.size _UpperCamelCase = self.image_processor(images=__a , return_tensors=self.framework) return model_inputs def UpperCAmelCase ( self , __a) -> Dict: '''simple docstring''' _UpperCamelCase = self.model(**__a) return model_outputs def UpperCAmelCase ( self , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = model_outputs.predicted_depth _UpperCamelCase = torch.nn.functional.interpolate( predicted_depth.unsqueeze(1) , size=self.image_size[::-1] , mode='''bicubic''' , align_corners=__a) _UpperCamelCase = prediction.squeeze().cpu().numpy() _UpperCamelCase = (output * 2_55 / np.max(__a)).astype('''uint8''') _UpperCamelCase = Image.fromarray(__a) _UpperCamelCase = {} _UpperCamelCase = predicted_depth _UpperCamelCase = depth return output_dict
78
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTImageProcessor, ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) def lowerCamelCase__ ( __snake_case, __snake_case=False ) -> Tuple: """simple docstring""" _UpperCamelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'''blocks.{i}.norm1.weight''', F'''vit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((F'''blocks.{i}.norm1.bias''', F'''vit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((F'''blocks.{i}.attn.proj.weight''', F'''vit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.attn.proj.bias''', F'''vit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((F'''blocks.{i}.norm2.weight''', F'''vit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((F'''blocks.{i}.norm2.bias''', F'''vit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.weight''', F'''vit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.bias''', F'''vit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.weight''', F'''vit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.bias''', F'''vit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ('''cls_token''', '''vit.embeddings.cls_token'''), ('''patch_embed.proj.weight''', '''vit.embeddings.patch_embeddings.projection.weight'''), ('''patch_embed.proj.bias''', '''vit.embeddings.patch_embeddings.projection.bias'''), ('''pos_embed''', '''vit.embeddings.position_embeddings'''), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ('''pre_logits.fc.weight''', '''pooler.dense.weight'''), ('''pre_logits.fc.bias''', '''pooler.dense.bias'''), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _UpperCamelCase = [(pair[0], pair[1][4:]) if pair[1].startswith('''vit''' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('''norm.weight''', '''vit.layernorm.weight'''), ('''norm.bias''', '''vit.layernorm.bias'''), ('''head.weight''', '''classifier.weight'''), ('''head.bias''', '''classifier.bias'''), ] ) return rename_keys def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case=False ) -> str: """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: _UpperCamelCase = '''''' else: _UpperCamelCase = '''vit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _UpperCamelCase = state_dict.pop(F'''blocks.{i}.attn.qkv.weight''' ) _UpperCamelCase = state_dict.pop(F'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict _UpperCamelCase = in_proj_weight[ : config.hidden_size, : ] _UpperCamelCase = in_proj_bias[: config.hidden_size] _UpperCamelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _UpperCamelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _UpperCamelCase = in_proj_weight[ -config.hidden_size :, : ] _UpperCamelCase = in_proj_bias[-config.hidden_size :] def lowerCamelCase__ ( __snake_case ) -> Dict: """simple docstring""" _UpperCamelCase = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(__snake_case, __snake_case ) def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> Any: """simple docstring""" _UpperCamelCase = dct.pop(__snake_case ) _UpperCamelCase = val def lowerCamelCase__ ( ) -> Dict: """simple docstring""" _UpperCamelCase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _UpperCamelCase = Image.open(requests.get(__snake_case, stream=__snake_case ).raw ) return im @torch.no_grad() def lowerCamelCase__ ( __snake_case, __snake_case ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = ViTConfig() _UpperCamelCase = False # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size if vit_name[-5:] == "in21k": _UpperCamelCase = True _UpperCamelCase = int(vit_name[-12:-10] ) _UpperCamelCase = int(vit_name[-9:-6] ) else: _UpperCamelCase = 10_00 _UpperCamelCase = '''huggingface/label-files''' _UpperCamelCase = '''imagenet-1k-id2label.json''' _UpperCamelCase = json.load(open(hf_hub_download(__snake_case, __snake_case, repo_type='''dataset''' ), '''r''' ) ) _UpperCamelCase = {int(__snake_case ): v for k, v in idalabel.items()} _UpperCamelCase = idalabel _UpperCamelCase = {v: k for k, v in idalabel.items()} _UpperCamelCase = int(vit_name[-6:-4] ) _UpperCamelCase = int(vit_name[-3:] ) # size of the architecture if "deit" in vit_name: if vit_name[9:].startswith('''tiny''' ): _UpperCamelCase = 1_92 _UpperCamelCase = 7_68 _UpperCamelCase = 12 _UpperCamelCase = 3 elif vit_name[9:].startswith('''small''' ): _UpperCamelCase = 3_84 _UpperCamelCase = 15_36 _UpperCamelCase = 12 _UpperCamelCase = 6 else: pass else: if vit_name[4:].startswith('''small''' ): _UpperCamelCase = 7_68 _UpperCamelCase = 23_04 _UpperCamelCase = 8 _UpperCamelCase = 8 elif vit_name[4:].startswith('''base''' ): pass elif vit_name[4:].startswith('''large''' ): _UpperCamelCase = 10_24 _UpperCamelCase = 40_96 _UpperCamelCase = 24 _UpperCamelCase = 16 elif vit_name[4:].startswith('''huge''' ): _UpperCamelCase = 12_80 _UpperCamelCase = 51_20 _UpperCamelCase = 32 _UpperCamelCase = 16 # load original model from timm _UpperCamelCase = timm.create_model(__snake_case, pretrained=__snake_case ) timm_model.eval() # load state_dict of original model, remove and rename some keys _UpperCamelCase = timm_model.state_dict() if base_model: remove_classification_head_(__snake_case ) _UpperCamelCase = create_rename_keys(__snake_case, __snake_case ) for src, dest in rename_keys: rename_key(__snake_case, __snake_case, __snake_case ) read_in_q_k_v(__snake_case, __snake_case, __snake_case ) # load HuggingFace model if vit_name[-5:] == "in21k": _UpperCamelCase = ViTModel(__snake_case ).eval() else: _UpperCamelCase = ViTForImageClassification(__snake_case ).eval() model.load_state_dict(__snake_case ) # Check outputs on an image, prepared by ViTImageProcessor/DeiTImageProcessor if "deit" in vit_name: _UpperCamelCase = DeiTImageProcessor(size=config.image_size ) else: _UpperCamelCase = ViTImageProcessor(size=config.image_size ) _UpperCamelCase = image_processor(images=prepare_img(), return_tensors='''pt''' ) _UpperCamelCase = encoding['''pixel_values'''] _UpperCamelCase = model(__snake_case ) if base_model: _UpperCamelCase = timm_model.forward_features(__snake_case ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(__snake_case, outputs.pooler_output, atol=1e-3 ) else: _UpperCamelCase = timm_model(__snake_case ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__snake_case, outputs.logits, atol=1e-3 ) Path(__snake_case ).mkdir(exist_ok=__snake_case ) print(F'''Saving model {vit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(__snake_case ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(__snake_case ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( """--vit_name""", default="""vit_base_patch16_224""", type=str, help="""Name of the ViT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) _a = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path)
78
1
"""simple docstring""" from typing import Dict, List, Optional, Tuple, Union import torch from ...models import AutoencoderKL, TransformeraDModel from ...schedulers import KarrasDiffusionSchedulers from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class _UpperCAmelCase( lowerCamelCase ): def __init__( self , __a , __a , __a , __a = None , ) -> List[Any]: '''simple docstring''' super().__init__() self.register_modules(transformer=__a , vae=__a , scheduler=__a) # create a imagenet -> id dictionary for easier use _UpperCamelCase = {} if idalabel is not None: for key, value in idalabel.items(): for label in value.split(''','''): _UpperCamelCase = int(__a) _UpperCamelCase = dict(sorted(self.labels.items())) def UpperCAmelCase ( self , __a) -> List[int]: '''simple docstring''' if not isinstance(__a , __a): _UpperCamelCase = list(__a) for l in label: if l not in self.labels: raise ValueError( F'''{l} does not exist. Please make sure to select one of the following labels: \n {self.labels}.''') return [self.labels[l] for l in label] @torch.no_grad() def __call__( self , __a , __a = 4.0 , __a = None , __a = 50 , __a = "pil" , __a = True , ) -> Union[ImagePipelineOutput, Tuple]: '''simple docstring''' _UpperCamelCase = len(__a) _UpperCamelCase = self.transformer.config.sample_size _UpperCamelCase = self.transformer.config.in_channels _UpperCamelCase = randn_tensor( shape=(batch_size, latent_channels, latent_size, latent_size) , generator=__a , device=self.device , dtype=self.transformer.dtype , ) _UpperCamelCase = torch.cat([latents] * 2) if guidance_scale > 1 else latents _UpperCamelCase = torch.tensor(__a , device=self.device).reshape(-1) _UpperCamelCase = torch.tensor([10_00] * batch_size , device=self.device) _UpperCamelCase = torch.cat([class_labels, class_null] , 0) if guidance_scale > 1 else class_labels # set step values self.scheduler.set_timesteps(__a) for t in self.progress_bar(self.scheduler.timesteps): if guidance_scale > 1: _UpperCamelCase = latent_model_input[: len(__a) // 2] _UpperCamelCase = torch.cat([half, half] , dim=0) _UpperCamelCase = self.scheduler.scale_model_input(__a , __a) _UpperCamelCase = t if not torch.is_tensor(__a): # TODO: this requires sync between CPU and GPU. So try to pass timesteps as tensors if you can # This would be a good case for the `match` statement (Python 3.10+) _UpperCamelCase = latent_model_input.device.type == '''mps''' if isinstance(__a , __a): _UpperCamelCase = torch.floataa if is_mps else torch.floataa else: _UpperCamelCase = torch.intaa if is_mps else torch.intaa _UpperCamelCase = torch.tensor([timesteps] , dtype=__a , device=latent_model_input.device) elif len(timesteps.shape) == 0: _UpperCamelCase = timesteps[None].to(latent_model_input.device) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML _UpperCamelCase = timesteps.expand(latent_model_input.shape[0]) # predict noise model_output _UpperCamelCase = self.transformer( __a , timestep=__a , class_labels=__a).sample # perform guidance if guidance_scale > 1: _UpperCamelCase , _UpperCamelCase = noise_pred[:, :latent_channels], noise_pred[:, latent_channels:] _UpperCamelCase , _UpperCamelCase = torch.split(__a , len(__a) // 2 , dim=0) _UpperCamelCase = uncond_eps + guidance_scale * (cond_eps - uncond_eps) _UpperCamelCase = torch.cat([half_eps, half_eps] , dim=0) _UpperCamelCase = torch.cat([eps, rest] , dim=1) # learned sigma if self.transformer.config.out_channels // 2 == latent_channels: _UpperCamelCase , _UpperCamelCase = torch.split(__a , __a , dim=1) else: _UpperCamelCase = noise_pred # compute previous image: x_t -> x_t-1 _UpperCamelCase = self.scheduler.step(__a , __a , __a).prev_sample if guidance_scale > 1: _UpperCamelCase , _UpperCamelCase = latent_model_input.chunk(2 , dim=0) else: _UpperCamelCase = latent_model_input _UpperCamelCase = 1 / self.vae.config.scaling_factor * latents _UpperCamelCase = self.vae.decode(__a).sample _UpperCamelCase = (samples / 2 + 0.5).clamp(0 , 1) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 _UpperCamelCase = samples.cpu().permute(0 , 2 , 3 , 1).float().numpy() if output_type == "pil": _UpperCamelCase = self.numpy_to_pil(__a) if not return_dict: return (samples,) return ImagePipelineOutput(images=__a)
78
"""simple docstring""" import unittest from transformers import AlbertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForPreTraining, AlbertForQuestionAnswering, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertModel, ) from transformers.models.albert.modeling_albert import ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST class _UpperCAmelCase: def __init__( self , __a , __a=13 , __a=7 , __a=True , __a=True , __a=True , __a=True , __a=99 , __a=16 , __a=36 , __a=6 , __a=6 , __a=6 , __a=37 , __a="gelu" , __a=0.1 , __a=0.1 , __a=5_12 , __a=16 , __a=2 , __a=0.02 , __a=3 , __a=4 , __a=None , ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = seq_length _UpperCamelCase = is_training _UpperCamelCase = use_input_mask _UpperCamelCase = use_token_type_ids _UpperCamelCase = use_labels _UpperCamelCase = vocab_size _UpperCamelCase = embedding_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_hidden_groups _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = type_vocab_size _UpperCamelCase = type_sequence_label_size _UpperCamelCase = initializer_range _UpperCamelCase = num_labels _UpperCamelCase = num_choices _UpperCamelCase = scope def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) _UpperCamelCase = None if self.use_input_mask: _UpperCamelCase = random_attention_mask([self.batch_size, self.seq_length]) _UpperCamelCase = None if self.use_token_type_ids: _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size) _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) _UpperCamelCase = ids_tensor([self.batch_size] , self.num_choices) _UpperCamelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase ( self) -> Dict: '''simple docstring''' return AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , num_hidden_groups=self.num_hidden_groups , ) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Optional[int]: '''simple docstring''' _UpperCamelCase = AlbertModel(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , token_type_ids=__a) _UpperCamelCase = model(__a , token_type_ids=__a) _UpperCamelCase = model(__a) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = AlbertForPreTraining(config=__a) model.to(__a) model.eval() _UpperCamelCase = model( __a , attention_mask=__a , token_type_ids=__a , labels=__a , sentence_order_label=__a , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) self.parent.assertEqual(result.sop_logits.shape , (self.batch_size, config.num_labels)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Optional[int]: '''simple docstring''' _UpperCamelCase = AlbertForMaskedLM(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Any: '''simple docstring''' _UpperCamelCase = AlbertForQuestionAnswering(config=__a) model.to(__a) model.eval() _UpperCamelCase = model( __a , attention_mask=__a , token_type_ids=__a , start_positions=__a , end_positions=__a , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> List[Any]: '''simple docstring''' _UpperCamelCase = self.num_labels _UpperCamelCase = AlbertForSequenceClassification(__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> List[str]: '''simple docstring''' _UpperCamelCase = self.num_labels _UpperCamelCase = AlbertForTokenClassification(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.num_choices _UpperCamelCase = AlbertForMultipleChoice(config=__a) model.to(__a) model.eval() _UpperCamelCase = input_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() _UpperCamelCase = token_type_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() _UpperCamelCase = input_mask.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() _UpperCamelCase = model( __a , attention_mask=__a , token_type_ids=__a , labels=__a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices)) def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.prepare_config_and_inputs() ( ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ) = config_and_inputs _UpperCamelCase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class _UpperCAmelCase( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase__ = ( ( AlbertModel, AlbertForPreTraining, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertForQuestionAnswering, ) if is_torch_available() else () ) lowercase__ = ( { 'feature-extraction': AlbertModel, 'fill-mask': AlbertForMaskedLM, 'question-answering': AlbertForQuestionAnswering, 'text-classification': AlbertForSequenceClassification, 'token-classification': AlbertForTokenClassification, 'zero-shot': AlbertForSequenceClassification, } if is_torch_available() else {} ) lowercase__ = True def UpperCAmelCase ( self , __a , __a , __a=False) -> List[str]: '''simple docstring''' _UpperCamelCase = super()._prepare_for_class(__a , __a , return_labels=__a) if return_labels: if model_class in get_values(__a): _UpperCamelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=__a) _UpperCamelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__a) return inputs_dict def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = AlbertModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=__a , hidden_size=37) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*__a) def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__a) def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__a) def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__a) def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__a) def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _UpperCamelCase = type self.model_tester.create_and_check_model(*__a) @slow def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' for model_name in ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase = AlbertModel.from_pretrained(__a) self.assertIsNotNone(__a) @require_torch class _UpperCAmelCase( unittest.TestCase ): @slow def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = AlbertModel.from_pretrained('''albert-base-v2''') _UpperCamelCase = torch.tensor([[0, 3_45, 2_32, 3_28, 7_40, 1_40, 16_95, 69, 60_78, 15_88, 2]]) _UpperCamelCase = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]) with torch.no_grad(): _UpperCamelCase = model(__a , attention_mask=__a)[0] _UpperCamelCase = torch.Size((1, 11, 7_68)) self.assertEqual(output.shape , __a) _UpperCamelCase = torch.tensor( [[[-0.6513, 1.5035, -0.2766], [-0.6515, 1.5046, -0.2780], [-0.6512, 1.5049, -0.2784]]]) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , __a , atol=1e-4))
78
1
"""simple docstring""" from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import tensorflow as tf from transformers import AutoTokenizer, TFAutoModelForSeqaSeqLM @require_tf @require_sentencepiece @require_tokenizers class _UpperCAmelCase( unittest.TestCase ): @slow def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = TFAutoModelForSeqaSeqLM.from_pretrained('''google/mt5-small''') _UpperCamelCase = AutoTokenizer.from_pretrained('''google/mt5-small''') _UpperCamelCase = tokenizer('''Hello there''' , return_tensors='''tf''').input_ids _UpperCamelCase = tokenizer('''Hi I am''' , return_tensors='''tf''').input_ids _UpperCamelCase = model(__a , labels=__a).loss _UpperCamelCase = -tf.math.reduce_mean(__a).numpy() _UpperCamelCase = -21.22_8168 self.assertTrue(abs(mtf_score - EXPECTED_SCORE) < 2e-4)
78
"""simple docstring""" import os from typing import BinaryIO, Optional, Union import numpy as np import pyarrow.parquet as pq from .. import Audio, Dataset, Features, Image, NamedSplit, Value, config from ..features.features import FeatureType, _visit from ..formatting import query_table from ..packaged_modules import _PACKAGED_DATASETS_MODULES from ..packaged_modules.parquet.parquet import Parquet from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" _UpperCamelCase = np.inf def set_batch_size(__snake_case ) -> None: nonlocal batch_size if isinstance(__snake_case, __snake_case ): _UpperCamelCase = min(__snake_case, config.PARQUET_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS ) elif isinstance(__snake_case, __snake_case ): _UpperCamelCase = min(__snake_case, config.PARQUET_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS ) elif isinstance(__snake_case, __snake_case ) and feature.dtype == "binary": _UpperCamelCase = min(__snake_case, config.PARQUET_ROW_GROUP_SIZE_FOR_BINARY_DATASETS ) _visit(__snake_case, __snake_case ) return None if batch_size is np.inf else batch_size class _UpperCAmelCase( lowerCamelCase ): def __init__( self , __a , __a = None , __a = None , __a = None , __a = False , __a = False , __a = None , **__a , ) -> Dict: '''simple docstring''' super().__init__( __a , split=__a , features=__a , cache_dir=__a , keep_in_memory=__a , streaming=__a , num_proc=__a , **__a , ) _UpperCamelCase = path_or_paths if isinstance(__a , __a) else {self.split: path_or_paths} _UpperCamelCase = _PACKAGED_DATASETS_MODULES['''parquet'''][1] _UpperCamelCase = Parquet( cache_dir=__a , data_files=__a , features=__a , hash=__a , **__a , ) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' # Build iterable dataset if self.streaming: _UpperCamelCase = self.builder.as_streaming_dataset(split=self.split) # Build regular (map-style) dataset else: _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None self.builder.download_and_prepare( download_config=__a , download_mode=__a , verification_mode=__a , base_path=__a , num_proc=self.num_proc , ) _UpperCamelCase = self.builder.as_dataset( split=self.split , verification_mode=__a , in_memory=self.keep_in_memory) return dataset class _UpperCAmelCase: def __init__( self , __a , __a , __a = None , **__a , ) -> Dict: '''simple docstring''' _UpperCamelCase = dataset _UpperCamelCase = path_or_buf _UpperCamelCase = batch_size or get_writer_batch_size(dataset.features) _UpperCamelCase = parquet_writer_kwargs def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = self.batch_size if self.batch_size else config.DEFAULT_MAX_BATCH_SIZE if isinstance(self.path_or_buf , (str, bytes, os.PathLike)): with open(self.path_or_buf , '''wb+''') as buffer: _UpperCamelCase = self._write(file_obj=__a , batch_size=__a , **self.parquet_writer_kwargs) else: _UpperCamelCase = self._write(file_obj=self.path_or_buf , batch_size=__a , **self.parquet_writer_kwargs) return written def UpperCAmelCase ( self , __a , __a , **__a) -> int: '''simple docstring''' _UpperCamelCase = 0 _UpperCamelCase = parquet_writer_kwargs.pop('''path_or_buf''' , __a) _UpperCamelCase = self.dataset.features.arrow_schema _UpperCamelCase = pq.ParquetWriter(__a , schema=__a , **__a) for offset in logging.tqdm( range(0 , len(self.dataset) , __a) , unit='''ba''' , disable=not logging.is_progress_bar_enabled() , desc='''Creating parquet from Arrow format''' , ): _UpperCamelCase = query_table( table=self.dataset._data , key=slice(__a , offset + batch_size) , indices=self.dataset._indices if self.dataset._indices is not None else None , ) writer.write_table(__a) written += batch.nbytes writer.close() return written
78
1
"""simple docstring""" import gc import importlib.metadata import tempfile import unittest from packaging import version from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoTokenizer, BitsAndBytesConfig, pipeline, ) from transformers.testing_utils import ( is_torch_available, require_accelerate, require_bitsandbytes, require_torch, require_torch_gpu, require_torch_multi_gpu, slow, ) def lowerCamelCase__ ( __snake_case ) -> List[str]: """simple docstring""" if model.config.model_type == "gpt2": return model.transformer.h[0].mlp.c_fc return model.transformer.h[0].mlp.dense_ah_to_h if is_torch_available(): import torch import torch.nn as nn class _UpperCAmelCase( nn.Module ): def __init__( self , __a , __a) -> List[str]: '''simple docstring''' super().__init__() _UpperCamelCase = module _UpperCamelCase = nn.Sequential( nn.Linear(module.in_features , __a , bias=__a) , nn.Linear(__a , module.out_features , bias=__a) , ) _UpperCamelCase = (2.0 / (5 * min(module.in_features , module.out_features))) ** 0.5 nn.init.normal_(self.adapter[0].weight , std=__a) nn.init.zeros_(self.adapter[1].weight) self.adapter.to(module.weight.device) def UpperCAmelCase ( self , __a , *__a , **__a) -> Optional[Any]: '''simple docstring''' return self.module(__a , *__a , **__a) + self.adapter(__a) @require_bitsandbytes @require_accelerate @require_torch @require_torch_gpu @slow class _UpperCAmelCase( unittest.TestCase ): # We keep the constants inside the init function and model loading inside setUp function # We need to test on relatively large models (aka >1b parameters otherwise the quantiztion may not work as expected) # Therefore here we use only bloom-1b3 to test our module lowercase__ = 'bigscience/bloom-1b7' # Constant values lowercase__ = 2.1_09_65_95_52_69_25_74 lowercase__ = 'Hello my name is' lowercase__ = set() EXPECTED_OUTPUTS.add('Hello my name is John and I am a professional photographer. I' ) EXPECTED_OUTPUTS.add('Hello my name is John.\nI am a friend of your father.\n' ) EXPECTED_OUTPUTS.add('Hello my name is John Doe, I am a student at the University' ) lowercase__ = 10 def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' # Models and tokenizer _UpperCamelCase = AutoTokenizer.from_pretrained(self.model_name) class _UpperCAmelCase( lowerCamelCase ): def UpperCAmelCase ( self) -> int: '''simple docstring''' super().setUp() # Models and tokenizer _UpperCamelCase = AutoModelForCausalLM.from_pretrained( self.model_name , torch_dtype=torch.floataa , device_map='''auto''') _UpperCamelCase = AutoModelForCausalLM.from_pretrained(self.model_name , load_in_abit=__a , device_map='''auto''') def UpperCAmelCase ( self) -> str: '''simple docstring''' del self.model_fpaa del self.model_abit gc.collect() torch.cuda.empty_cache() def UpperCAmelCase ( self) -> int: '''simple docstring''' _UpperCamelCase = self.model_abit.config self.assertTrue(hasattr(__a , '''quantization_config''')) _UpperCamelCase = config.to_dict() _UpperCamelCase = config.to_diff_dict() _UpperCamelCase = config.to_json_string() def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' from bitsandbytes.nn import Paramsabit _UpperCamelCase = self.model_fpaa.get_memory_footprint() _UpperCamelCase = self.model_abit.get_memory_footprint() self.assertAlmostEqual(mem_fpaa / mem_abit , self.EXPECTED_RELATIVE_DIFFERENCE) _UpperCamelCase = get_some_linear_layer(self.model_abit) self.assertTrue(linear.weight.__class__ == Paramsabit) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' from transformers import TaPreTrainedModel self.model_fpaa.get_memory_footprint() self.model_abit.get_memory_footprint() for name, module in self.model_abit.named_modules(): if isinstance(__a , torch.nn.Linear): if name not in ["lm_head"] + TaPreTrainedModel._keep_in_fpaa_modules: # 4-bit parameters are packed in uint8 variables self.assertTrue(module.weight.dtype == torch.uinta) def UpperCAmelCase ( self) -> str: '''simple docstring''' _UpperCamelCase = self.tokenizer(self.input_text , return_tensors='''pt''') _UpperCamelCase = self.model_abit.generate(input_ids=encoded_input['''input_ids'''].to(0) , max_new_tokens=10) self.assertIn(self.tokenizer.decode(output_sequences[0] , skip_special_tokens=__a) , self.EXPECTED_OUTPUTS) def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = BitsAndBytesConfig() _UpperCamelCase = True _UpperCamelCase = AutoModelForCausalLM.from_pretrained( self.model_name , quantization_config=__a , device_map='''auto''') _UpperCamelCase = self.tokenizer(self.input_text , return_tensors='''pt''') _UpperCamelCase = model_abit_from_config.generate( input_ids=encoded_input['''input_ids'''].to(0) , max_new_tokens=10) self.assertIn(self.tokenizer.decode(output_sequences[0] , skip_special_tokens=__a) , self.EXPECTED_OUTPUTS) def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' with self.assertRaises(__a), tempfile.TemporaryDirectory() as tmpdirname: self.model_abit.save_pretrained(__a) def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = BitsAndBytesConfig() with self.assertRaises(__a): _UpperCamelCase = AutoModelForCausalLM.from_pretrained( self.model_name , quantization_config=__a , load_in_abit=__a , device_map='''auto''' , bnb_abit_quant_type='''nf4''' , ) def UpperCAmelCase ( self) -> str: '''simple docstring''' with self.assertRaises(__a): # Tries with `str` self.model_abit.to('''cpu''') with self.assertRaises(__a): # Tries with a `dtype`` self.model_abit.to(torch.floataa) with self.assertRaises(__a): # Tries with a `device` self.model_abit.to(torch.device('''cuda:0''')) with self.assertRaises(__a): # Tries with a `device` self.model_abit.float() with self.assertRaises(__a): # Tries with a `device` self.model_abit.half() # Test if we did not break anything _UpperCamelCase = self.tokenizer(self.input_text , return_tensors='''pt''') _UpperCamelCase = self.model_fpaa.to(torch.floataa) _UpperCamelCase = self.model_fpaa.generate(input_ids=encoded_input['''input_ids'''].to(0) , max_new_tokens=10) # Check this does not throw an error _UpperCamelCase = self.model_fpaa.to('''cpu''') # Check this does not throw an error _UpperCamelCase = self.model_fpaa.half() # Check this does not throw an error _UpperCamelCase = self.model_fpaa.float() def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = AutoModelForSeqaSeqLM.from_pretrained('''t5-small''' , load_in_abit=__a , device_map='''auto''') self.assertTrue(model.decoder.block[0].layer[2].DenseReluDense.wo.weight.dtype == torch.floataa) @require_bitsandbytes @require_accelerate @require_torch @require_torch_gpu @slow class _UpperCAmelCase( unittest.TestCase ): @classmethod def UpperCAmelCase ( cls) -> List[str]: '''simple docstring''' _UpperCamelCase = '''t5-small''' _UpperCamelCase = '''google/flan-t5-small''' # flan-t5 uses dense-act instead of dense-relu-dense _UpperCamelCase = AutoTokenizer.from_pretrained(cls.model_name) _UpperCamelCase = '''Translate in German: Hello, my dog is cute''' def UpperCAmelCase ( self) -> Any: '''simple docstring''' gc.collect() torch.cuda.empty_cache() def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' from transformers import TaForConditionalGeneration _UpperCamelCase = TaForConditionalGeneration._keep_in_fpaa_modules _UpperCamelCase = None # test with `t5-small` _UpperCamelCase = TaForConditionalGeneration.from_pretrained(self.model_name , load_in_abit=__a , device_map='''auto''') _UpperCamelCase = self.tokenizer(self.input_text , return_tensors='''pt''').to(0) _UpperCamelCase = model.generate(**__a) # test with `flan-t5-small` _UpperCamelCase = TaForConditionalGeneration.from_pretrained( self.dense_act_model_name , load_in_abit=__a , device_map='''auto''') _UpperCamelCase = self.tokenizer(self.input_text , return_tensors='''pt''').to(0) _UpperCamelCase = model.generate(**__a) _UpperCamelCase = modules def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' import bitsandbytes as bnb from transformers import TaForConditionalGeneration # test with `t5-small` _UpperCamelCase = TaForConditionalGeneration.from_pretrained(self.model_name , load_in_abit=__a , device_map='''auto''') # there was a bug with decoders - this test checks that it is fixed self.assertTrue(isinstance(model.decoder.block[0].layer[0].SelfAttention.q , bnb.nn.Linearabit)) _UpperCamelCase = self.tokenizer(self.input_text , return_tensors='''pt''').to(0) _UpperCamelCase = model.generate(**__a) # test with `flan-t5-small` _UpperCamelCase = TaForConditionalGeneration.from_pretrained( self.dense_act_model_name , load_in_abit=__a , device_map='''auto''') _UpperCamelCase = self.tokenizer(self.input_text , return_tensors='''pt''').to(0) _UpperCamelCase = model.generate(**__a) class _UpperCAmelCase( lowerCamelCase ): def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' super().setUp() # model_name _UpperCamelCase = '''bigscience/bloom-560m''' _UpperCamelCase = '''t5-small''' # Different types of model _UpperCamelCase = AutoModel.from_pretrained(self.model_name , load_in_abit=__a , device_map='''auto''') # Sequence classification model _UpperCamelCase = AutoModelForSequenceClassification.from_pretrained( self.model_name , load_in_abit=__a , device_map='''auto''') # CausalLM model _UpperCamelCase = AutoModelForCausalLM.from_pretrained(self.model_name , load_in_abit=__a , device_map='''auto''') # Seq2seq model _UpperCamelCase = AutoModelForSeqaSeqLM.from_pretrained( self.seq_to_seq_name , load_in_abit=__a , device_map='''auto''') def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' del self.base_model del self.sequence_model del self.model_abit del self.seq_to_seq_model gc.collect() torch.cuda.empty_cache() def UpperCAmelCase ( self) -> Dict: '''simple docstring''' from bitsandbytes.nn import Paramsabit self.assertTrue(self.base_model.h[-1].mlp.dense_ah_to_h.weight.__class__ == Paramsabit) # Other heads should be nn.Parameter self.assertTrue(self.model_abit.lm_head.weight.__class__ == torch.nn.Parameter) self.assertTrue(self.sequence_model.score.weight.__class__ == torch.nn.Parameter) self.assertTrue(self.seq_to_seq_model.lm_head.weight.__class__ == torch.nn.Parameter) class _UpperCAmelCase( lowerCamelCase ): def UpperCAmelCase ( self) -> int: '''simple docstring''' super().setUp() def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' del self.pipe gc.collect() torch.cuda.empty_cache() def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = pipeline( '''text-generation''' , model=self.model_name , model_kwargs={'''device_map''': '''auto''', '''load_in_4bit''': True, '''torch_dtype''': torch.floataa} , max_new_tokens=self.MAX_NEW_TOKENS , ) # Real second forward pass _UpperCamelCase = self.pipe(self.input_text) self.assertIn(pipeline_output[0]['''generated_text'''] , self.EXPECTED_OUTPUTS) @require_torch_multi_gpu class _UpperCAmelCase( lowerCamelCase ): def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' super().setUp() def UpperCAmelCase ( self) -> str: '''simple docstring''' _UpperCamelCase = AutoModelForCausalLM.from_pretrained( self.model_name , load_in_abit=__a , device_map='''balanced''') # Check correct device map self.assertEqual(set(model_parallel.hf_device_map.values()) , {0, 1}) # Check that inference pass works on the model _UpperCamelCase = self.tokenizer(self.input_text , return_tensors='''pt''') # Second real batch _UpperCamelCase = model_parallel.generate(input_ids=encoded_input['''input_ids'''].to(0) , max_new_tokens=10) self.assertIn(self.tokenizer.decode(output_parallel[0] , skip_special_tokens=__a) , self.EXPECTED_OUTPUTS) class _UpperCAmelCase( lowerCamelCase ): def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = '''facebook/opt-350m''' super().setUp() def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' if version.parse(importlib.metadata.version('''bitsandbytes''')) < version.parse('''0.37.0'''): return # Step 1: freeze all parameters _UpperCamelCase = AutoModelForCausalLM.from_pretrained(self.model_name , load_in_abit=__a) self.assertEqual(set(model.hf_device_map.values()) , {torch.cuda.current_device()}) for param in model.parameters(): _UpperCamelCase = False # freeze the model - train adapters later if param.ndim == 1: # cast the small parameters (e.g. layernorm) to fp32 for stability _UpperCamelCase = param.data.to(torch.floataa) # Step 2: add adapters for _, module in model.named_modules(): if "OPTAttention" in repr(type(__a)): _UpperCamelCase = LoRALayer(module.q_proj , rank=16) _UpperCamelCase = LoRALayer(module.k_proj , rank=16) _UpperCamelCase = LoRALayer(module.v_proj , rank=16) # Step 3: dummy batch _UpperCamelCase = self.tokenizer('''Test batch ''' , return_tensors='''pt''').to(0) # Step 4: Check if the gradient is not None with torch.cuda.amp.autocast(): _UpperCamelCase = model.forward(**__a) out.logits.norm().backward() for module in model.modules(): if isinstance(__a , __a): self.assertTrue(module.adapter[1].weight.grad is not None) self.assertTrue(module.adapter[1].weight.grad.norm().item() > 0) elif isinstance(__a , nn.Embedding): self.assertTrue(module.weight.grad is None) class _UpperCAmelCase( lowerCamelCase ): lowercase__ = 'gpt2-xl' lowercase__ = 3.31_91_85_48_54_15_21_87
78
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class _UpperCAmelCase( unittest.TestCase ): def __init__( self , __a , __a=7 , __a=3 , __a=18 , __a=30 , __a=4_00 , __a=True , __a=None , __a=True , __a=None , __a=True , ) -> int: '''simple docstring''' _UpperCamelCase = size if size is not None else {'''shortest_edge''': 20} _UpperCamelCase = crop_size if crop_size is not None else {'''height''': 18, '''width''': 18} _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = num_channels _UpperCamelCase = image_size _UpperCamelCase = min_resolution _UpperCamelCase = max_resolution _UpperCamelCase = do_resize _UpperCamelCase = size _UpperCamelCase = do_center_crop _UpperCamelCase = crop_size _UpperCamelCase = do_flip_channel_order def UpperCAmelCase ( self) -> str: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_flip_channel_order": self.do_flip_channel_order, } @require_torch @require_vision class _UpperCAmelCase( lowerCamelCase , unittest.TestCase ): lowercase__ = MobileViTImageProcessor if is_vision_available() else None def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = MobileViTImageProcessingTester(self) @property def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(__a , '''do_resize''')) self.assertTrue(hasattr(__a , '''size''')) self.assertTrue(hasattr(__a , '''do_center_crop''')) self.assertTrue(hasattr(__a , '''center_crop''')) self.assertTrue(hasattr(__a , '''do_flip_channel_order''')) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size , {'''shortest_edge''': 20}) self.assertEqual(image_processor.crop_size , {'''height''': 18, '''width''': 18}) _UpperCamelCase = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84) self.assertEqual(image_processor.size , {'''shortest_edge''': 42}) self.assertEqual(image_processor.crop_size , {'''height''': 84, '''width''': 84}) def UpperCAmelCase ( self) -> Dict: '''simple docstring''' pass def UpperCAmelCase ( self) -> str: '''simple docstring''' # Initialize image_processing _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random PIL images _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__a) for image in image_inputs: self.assertIsInstance(__a , Image.Image) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(__a , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' # Initialize image_processing _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__a , numpify=__a) for image in image_inputs: self.assertIsInstance(__a , np.ndarray) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(__a , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def UpperCAmelCase ( self) -> int: '''simple docstring''' # Initialize image_processing _UpperCamelCase = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=__a , torchify=__a) for image in image_inputs: self.assertIsInstance(__a , torch.Tensor) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(__a , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
78
1
"""simple docstring""" import collections import json import os import re from typing import TYPE_CHECKING, List, Optional, Tuple import numpy as np from ...tokenization_utils_fast import PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _a = logging.get_logger(__name__) _a = {"""vocab_file""": """vocab.txt""", """emoji_file""": """emoji.json"""} _a = { """vocab_file""": { """abeja/gpt-neox-japanese-2.7b""": """https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/vocab.txt""", }, """emoji_file""": { """abeja/gpt-neox-japanese-2.7b""": """https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/emoji.json""", }, } _a = { """abeja/gpt-neox-japanese-2.7b""": 2048, } def lowerCamelCase__ ( __snake_case, __snake_case ) -> str: """simple docstring""" with open(__snake_case, '''r''', encoding='''utf-8''' ) as f: _UpperCamelCase = json.loads(f.read() ) _UpperCamelCase = collections.OrderedDict() _UpperCamelCase = collections.OrderedDict() _UpperCamelCase = collections.OrderedDict() with open(__snake_case, '''r''', encoding='''utf-8''' ) as f: _UpperCamelCase = f.readlines() _UpperCamelCase = [[t.rstrip('''\n''' )] if (t == ''',''' or ''',''' not in t) else t.rstrip('''\n''' ).split(''',''' ) for t in token] for idx, b in enumerate(__snake_case ): _UpperCamelCase = b _UpperCamelCase = idx for wd in b: _UpperCamelCase = idx return vocab, raw_vocab, ids_to_tokens, emoji class _UpperCAmelCase( lowerCamelCase ): lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = ['input_ids', 'attention_mask'] def __init__( self , __a , __a , __a="<|endoftext|>" , __a="<|endoftext|>" , __a="<|startoftext|>" , __a="<|endoftext|>" , __a=False , **__a , ) -> Any: '''simple docstring''' super().__init__( unk_token=__a , pad_token=__a , bos_token=__a , eos_token=__a , do_clean_text=__a , **__a , ) if not os.path.isfile(__a): raise ValueError( F'''Can\'t find a vocabulary file at path \'{vocab_file}\'. To load the vocabulary from a Google pretrained''' ''' model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`''') if not os.path.isfile(__a): raise ValueError( F'''Can\'t find a emoji file at path \'{emoji_file}\'. To load the emoji information from a Google''' ''' pretrained model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`''') _UpperCamelCase = do_clean_text _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = load_vocab_and_emoji(__a , __a) _UpperCamelCase = SubWordJapaneseTokenizer( vocab=self.vocab , ids_to_tokens=self.ids_to_tokens , emoji=self.emoji) @property def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' # self.vocab contains support for character fluctuation unique to Japanese, and has a large number of vocab return len(self.raw_vocab) def UpperCAmelCase ( self) -> str: '''simple docstring''' return dict(self.raw_vocab , **self.added_tokens_encoder) def UpperCAmelCase ( self , __a) -> Tuple: '''simple docstring''' return self.subword_tokenizer.tokenize(__a , clean=self.do_clean_text) def UpperCAmelCase ( self , __a) -> Any: '''simple docstring''' return self.vocab.get(__a , self.vocab.get(self.unk_token)) def UpperCAmelCase ( self , __a) -> Union[str, Any]: '''simple docstring''' return self.subword_tokenizer.convert_id_to_token(__a) def UpperCAmelCase ( self , __a) -> int: '''simple docstring''' _UpperCamelCase = ''''''.join(__a).strip() return out_string def UpperCAmelCase ( self , __a) -> List[int]: '''simple docstring''' _UpperCamelCase = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(__a , add_special_tokens=__a) + [self.eos_token_id]) if len(__a) > self.model_max_length: _UpperCamelCase = input_ids[-self.model_max_length :] return input_ids def UpperCAmelCase ( self , __a , __a = None) -> Tuple[str]: '''simple docstring''' _UpperCamelCase = 0 if os.path.isdir(__a): _UpperCamelCase = os.path.join( __a , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file''']) _UpperCamelCase = os.path.join( __a , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''emoji_file''']) else: _UpperCamelCase = ( (filename_prefix + '''-''' if filename_prefix else '''''') + save_directory + VOCAB_FILES_NAMES['''vocab_file'''] ) _UpperCamelCase = ( (filename_prefix + '''-''' if filename_prefix else '''''') + save_directory + VOCAB_FILES_NAMES['''emoji_file'''] ) with open(__a , '''w''' , encoding='''utf-8''') as writer: for token_index, token in self.ids_to_tokens.items(): if index != token_index: logger.warning( F'''Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.''' ''' Please check that the vocabulary is not corrupted!''') _UpperCamelCase = token_index writer.write(''','''.join(__a) + '''\n''') index += 1 with open(__a , '''w''' , encoding='''utf-8''') as writer: json.dump(self.emoji , __a) return vocab_file, emoji_file class _UpperCAmelCase( lowerCamelCase ): def __init__( self , __a , __a , __a) -> Any: '''simple docstring''' _UpperCamelCase = vocab # same as swe _UpperCamelCase = ids_to_tokens # same as bpe _UpperCamelCase = emoji _UpperCamelCase = np.max([len(__a) for w in self.vocab.keys()]) _UpperCamelCase = re.compile(R'''(https?|ftp)(:\/\/[-_\.!~*\'()a-zA-Z0-9;\/?:\@&=\+$,%#]+)''') _UpperCamelCase = re.compile(R'''[A-Za-z0-9\._+]*@[\-_0-9A-Za-z]+(\.[A-Za-z]+)*''') _UpperCamelCase = re.compile(R'''[\(]{0,1}[0-9]{2,4}[\)\-\(]{0,1}[0-9]{2,4}[\)\-]{0,1}[0-9]{3,4}''') _UpperCamelCase = re.compile( R'''([12]\d{3}[/\-年])*(0?[1-9]|1[0-2])[/\-月]((0?[1-9]|[12][0-9]|3[01])日?)*(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*''') _UpperCamelCase = re.compile( R'''(明治|大正|昭和|平成|令和|㍾|㍽|㍼|㍻|\u32ff)\d{1,2}年(0?[1-9]|1[0-2])月(0?[1-9]|[12][0-9]|3[01])日(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*''') _UpperCamelCase = re.compile( R'''((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*億)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*万)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*千)*(0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*(千円|万円|千万円|円|千ドル|万ドル|千万ドル|ドル|千ユーロ|万ユーロ|千万ユーロ|ユーロ)+(\(税込\)|\(税抜\)|\+tax)*''') _UpperCamelCase = '''─━│┃┄┅┆┇┈┉┊┋┌┍┎┏┐┑┒┓└┕┖┗┘┙┚┛├┝┞┟┠┡┢┣┤┥┦┧┨┩┪┫┬┭┮┯┰┱┲┳┴┵┶┷┸┹┺┻┼┽┾┿╀╁╂╃╄╅╆╇╈╉╊╋╌╍╎╏═║╒╓╔╕╖╗╘╙╚╛╜╝╞╟╠╡╢╣╤╥╦╧╨╩╪╫╬╭╮╯╰╱╲╳╴╵╶╷╸╹╺╻╼╽╾╿''' _UpperCamelCase = '''▀▁▂▃▄▅▆▇█▉▊▋▌▍▎▏▐░▒▓▔▕▖▗▘▙▚▛▜▝▞▟''' _UpperCamelCase = str.maketrans({k: '''<BLOCK>''' for k in keisen + blocks}) def __len__( self) -> int: '''simple docstring''' return len(self.ids_to_tokens) def UpperCAmelCase ( self , __a) -> List[str]: '''simple docstring''' _UpperCamelCase = self.content_repattera.sub('''<URL>''' , __a) _UpperCamelCase = self.content_repattera.sub('''<EMAIL>''' , __a) _UpperCamelCase = self.content_repattera.sub('''<TEL>''' , __a) _UpperCamelCase = self.content_repattera.sub('''<DATE>''' , __a) _UpperCamelCase = self.content_repattera.sub('''<DATE>''' , __a) _UpperCamelCase = self.content_repattera.sub('''<PRICE>''' , __a) _UpperCamelCase = content.translate(self.content_transa) while "<BLOCK><BLOCK>" in content: _UpperCamelCase = content.replace('''<BLOCK><BLOCK>''' , '''<BLOCK>''') return content def UpperCAmelCase ( self , __a , __a=False) -> List[Any]: '''simple docstring''' _UpperCamelCase = text.replace(''' ''' , '''<SP>''') _UpperCamelCase = text.replace(''' ''' , '''<SP>''') _UpperCamelCase = text.replace('''\r\n''' , '''<BR>''') _UpperCamelCase = text.replace('''\n''' , '''<BR>''') _UpperCamelCase = text.replace('''\r''' , '''<BR>''') _UpperCamelCase = text.replace('''\t''' , '''<TAB>''') _UpperCamelCase = text.replace('''—''' , '''ー''') _UpperCamelCase = text.replace('''−''' , '''ー''') for k, v in self.emoji["emoji"].items(): if k in text: _UpperCamelCase = text.replace(__a , __a) if clean: _UpperCamelCase = self.clean_text(__a) def check_simbol(__a): _UpperCamelCase = x.encode() if len(__a) == 1 and len(__a) == 2: _UpperCamelCase = (int(e[0]) << 8) + int(e[1]) if ( (c >= 0XC2A1 and c <= 0XC2BF) or (c >= 0XC780 and c <= 0XC783) or (c >= 0XCAB9 and c <= 0XCBBF) or (c >= 0XCC80 and c <= 0XCDA2) ): return True return False def checkuae(__a): _UpperCamelCase = x.encode() if len(__a) == 1 and len(__a) == 3: _UpperCamelCase = (int(e[0]) << 16) + (int(e[1]) << 8) + int(e[2]) if c >= 0XE28080 and c <= 0XE2B07F: return True return False _UpperCamelCase = 0 _UpperCamelCase = [] while pos < len(__a): _UpperCamelCase = min(len(__a) , pos + self.maxlen + 1) if text[pos] == '''<''' else pos + 3 _UpperCamelCase = [] # (token_id, token, pos) for e in range(__a , __a , -1): _UpperCamelCase = text[pos:e] if wd in self.vocab: if wd[0] == "<" and len(__a) > 2: _UpperCamelCase = [(self.vocab[wd], wd, e)] break else: candidates.append((self.vocab[wd], wd, e)) if len(__a) > 0: # the smallest token_id is adopted _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = sorted(__a , key=lambda __a: x[0])[0] result.append(__a) _UpperCamelCase = e else: _UpperCamelCase = pos + 1 _UpperCamelCase = text[pos:end] if check_simbol(__a): result.append('''<KIGOU>''') elif checkuae(__a): result.append('''<U2000U2BFF>''') else: for i in wd.encode('''utf-8'''): result.append('''<|byte%d|>''' % i) _UpperCamelCase = end return result def UpperCAmelCase ( self , __a , __a="\n") -> str: '''simple docstring''' _UpperCamelCase = [] _UpperCamelCase = [] _UpperCamelCase = self.ids_to_tokens[index][0] if word[:6] == "<|byte" and word[-2:] == "|>": byte_tokens.append(int(word[6:-2])) else: if len(__a) > 0: words.append(bytearray(__a).decode('''utf-8''' , errors='''replace''')) _UpperCamelCase = [] if word[:7] == "<|emoji" and word[-2:] == "|>": words.append(self.emoji['''emoji_inv'''][word]) elif word == "<SP>": words.append(''' ''') elif word == "<BR>": words.append(__a) elif word == "<TAB>": words.append('''\t''') elif word == "<BLOCK>": words.append('''▀''') elif word == "<KIGOU>": words.append('''ǀ''') elif word == "<U2000U2BFF>": words.append('''‖''') else: words.append(__a) if len(__a) > 0: words.append(bytearray(__a).decode('''utf-8''' , errors='''replace''')) _UpperCamelCase = ''''''.join(__a) return text
78
"""simple docstring""" import warnings from typing import List import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import is_flax_available, is_tf_available, is_torch_available class _UpperCAmelCase( lowerCamelCase ): lowercase__ = ['image_processor', 'tokenizer'] lowercase__ = 'OwlViTImageProcessor' lowercase__ = ('CLIPTokenizer', 'CLIPTokenizerFast') def __init__( self , __a=None , __a=None , **__a) -> List[Any]: '''simple docstring''' _UpperCamelCase = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , __a , ) _UpperCamelCase = kwargs.pop('''feature_extractor''') _UpperCamelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''') if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''') super().__init__(__a , __a) def __call__( self , __a=None , __a=None , __a=None , __a="max_length" , __a="np" , **__a) -> List[str]: '''simple docstring''' if text is None and query_images is None and images is None: raise ValueError( '''You have to specify at least one text or query image or image. All three cannot be none.''') if text is not None: if isinstance(__a , __a) or (isinstance(__a , __a) and not isinstance(text[0] , __a)): _UpperCamelCase = [self.tokenizer(__a , padding=__a , return_tensors=__a , **__a)] elif isinstance(__a , __a) and isinstance(text[0] , __a): _UpperCamelCase = [] # Maximum number of queries across batch _UpperCamelCase = max([len(__a) for t in text]) # Pad all batch samples to max number of text queries for t in text: if len(__a) != max_num_queries: _UpperCamelCase = t + [''' '''] * (max_num_queries - len(__a)) _UpperCamelCase = self.tokenizer(__a , padding=__a , return_tensors=__a , **__a) encodings.append(__a) else: raise TypeError('''Input text should be a string, a list of strings or a nested list of strings''') if return_tensors == "np": _UpperCamelCase = np.concatenate([encoding['''input_ids'''] for encoding in encodings] , axis=0) _UpperCamelCase = np.concatenate([encoding['''attention_mask'''] for encoding in encodings] , axis=0) elif return_tensors == "jax" and is_flax_available(): import jax.numpy as jnp _UpperCamelCase = jnp.concatenate([encoding['''input_ids'''] for encoding in encodings] , axis=0) _UpperCamelCase = jnp.concatenate([encoding['''attention_mask'''] for encoding in encodings] , axis=0) elif return_tensors == "pt" and is_torch_available(): import torch _UpperCamelCase = torch.cat([encoding['''input_ids'''] for encoding in encodings] , dim=0) _UpperCamelCase = torch.cat([encoding['''attention_mask'''] for encoding in encodings] , dim=0) elif return_tensors == "tf" and is_tf_available(): import tensorflow as tf _UpperCamelCase = tf.stack([encoding['''input_ids'''] for encoding in encodings] , axis=0) _UpperCamelCase = tf.stack([encoding['''attention_mask'''] for encoding in encodings] , axis=0) else: raise ValueError('''Target return tensor type could not be returned''') _UpperCamelCase = BatchEncoding() _UpperCamelCase = input_ids _UpperCamelCase = attention_mask if query_images is not None: _UpperCamelCase = BatchEncoding() _UpperCamelCase = self.image_processor( __a , return_tensors=__a , **__a).pixel_values _UpperCamelCase = query_pixel_values if images is not None: _UpperCamelCase = self.image_processor(__a , return_tensors=__a , **__a) if text is not None and images is not None: _UpperCamelCase = image_features.pixel_values return encoding elif query_images is not None and images is not None: _UpperCamelCase = image_features.pixel_values return encoding elif text is not None or query_images is not None: return encoding else: return BatchEncoding(data=dict(**__a) , tensor_type=__a) def UpperCAmelCase ( self , *__a , **__a) -> str: '''simple docstring''' return self.image_processor.post_process(*__a , **__a) def UpperCAmelCase ( self , *__a , **__a) -> Dict: '''simple docstring''' return self.image_processor.post_process_object_detection(*__a , **__a) def UpperCAmelCase ( self , *__a , **__a) -> Optional[Any]: '''simple docstring''' return self.image_processor.post_process_image_guided_detection(*__a , **__a) def UpperCAmelCase ( self , *__a , **__a) -> Optional[int]: '''simple docstring''' return self.tokenizer.batch_decode(*__a , **__a) def UpperCAmelCase ( self , *__a , **__a) -> Optional[Any]: '''simple docstring''' return self.tokenizer.decode(*__a , **__a) @property def UpperCAmelCase ( self) -> str: '''simple docstring''' warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , __a , ) return self.image_processor_class @property def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , __a , ) return self.image_processor
78
1
"""simple docstring""" import numpy as np from numpy import ndarray from scipy.optimize import Bounds, LinearConstraint, minimize def lowerCamelCase__ ( __snake_case ) -> float: """simple docstring""" return np.dot(__snake_case, __snake_case ) class _UpperCAmelCase: def __init__( self , *, __a = np.inf , __a = "linear" , __a = 0.0 , ) -> None: '''simple docstring''' _UpperCamelCase = regularization _UpperCamelCase = gamma if kernel == "linear": _UpperCamelCase = self.__linear elif kernel == "rbf": if self.gamma == 0: raise ValueError('''rbf kernel requires gamma''') if not isinstance(self.gamma , (float, int)): raise ValueError('''gamma must be float or int''') if not self.gamma > 0: raise ValueError('''gamma must be > 0''') _UpperCamelCase = self.__rbf # in the future, there could be a default value like in sklearn # sklear: def_gamma = 1/(n_features * X.var()) (wiki) # previously it was 1/(n_features) else: _UpperCamelCase = F'''Unknown kernel: {kernel}''' raise ValueError(__a) def UpperCAmelCase ( self , __a , __a) -> float: '''simple docstring''' return np.dot(__a , __a) def UpperCAmelCase ( self , __a , __a) -> float: '''simple docstring''' return np.exp(-(self.gamma * norm_squared(vectora - vectora))) def UpperCAmelCase ( self , __a , __a) -> None: '''simple docstring''' _UpperCamelCase = observations _UpperCamelCase = classes # using Wolfe's Dual to calculate w. # Primal problem: minimize 1/2*norm_squared(w) # constraint: yn(w . xn + b) >= 1 # # With l a vector # Dual problem: maximize sum_n(ln) - # 1/2 * sum_n(sum_m(ln*lm*yn*ym*xn . xm)) # constraint: self.C >= ln >= 0 # and sum_n(ln*yn) = 0 # Then we get w using w = sum_n(ln*yn*xn) # At the end we can get b ~= mean(yn - w . xn) # # Since we use kernels, we only need l_star to calculate b # and to classify observations ((_UpperCamelCase) , ) = np.shape(__a) def to_minimize(__a) -> float: _UpperCamelCase = 0 ((_UpperCamelCase) , ) = np.shape(__a) for i in range(__a): for j in range(__a): s += ( candidate[i] * candidate[j] * classes[i] * classes[j] * self.kernel(observations[i] , observations[j]) ) return 1 / 2 * s - sum(__a) _UpperCamelCase = LinearConstraint(__a , 0 , 0) _UpperCamelCase = Bounds(0 , self.regularization) _UpperCamelCase = minimize( __a , np.ones(__a) , bounds=__a , constraints=[ly_contraint]).x _UpperCamelCase = l_star # calculating mean offset of separation plane to points _UpperCamelCase = 0 for i in range(__a): for j in range(__a): s += classes[i] - classes[i] * self.optimum[i] * self.kernel( observations[i] , observations[j]) _UpperCamelCase = s / n def UpperCAmelCase ( self , __a) -> int: '''simple docstring''' _UpperCamelCase = sum( self.optimum[n] * self.classes[n] * self.kernel(self.observations[n] , __a) for n in range(len(self.classes))) return 1 if s + self.offset >= 0 else -1 if __name__ == "__main__": import doctest doctest.testmod()
78
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) _a = { """configuration_perceiver""": ["""PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """PerceiverConfig""", """PerceiverOnnxConfig"""], """tokenization_perceiver""": ["""PerceiverTokenizer"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ["""PerceiverFeatureExtractor"""] _a = ["""PerceiverImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ """PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST""", """PerceiverForImageClassificationConvProcessing""", """PerceiverForImageClassificationFourier""", """PerceiverForImageClassificationLearned""", """PerceiverForMaskedLM""", """PerceiverForMultimodalAutoencoding""", """PerceiverForOpticalFlow""", """PerceiverForSequenceClassification""", """PerceiverLayer""", """PerceiverModel""", """PerceiverPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys _a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
78
1
"""simple docstring""" import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class _UpperCAmelCase( tf.keras.layers.Layer ): def __init__( self , __a , __a , __a = None , __a = None) -> List[str]: '''simple docstring''' super().__init__() _UpperCamelCase = pad_token_id _UpperCamelCase = max_length _UpperCamelCase = vocab _UpperCamelCase = merges _UpperCamelCase = BytePairTokenizer(__a , __a , sequence_length=__a) @classmethod def UpperCAmelCase ( cls , __a , *__a , **__a) -> Optional[int]: '''simple docstring''' _UpperCamelCase = [''' '''.join(__a) for m in tokenizer.bpe_ranks.keys()] _UpperCamelCase = tokenizer.get_vocab() return cls(__a , __a , *__a , **__a) @classmethod def UpperCAmelCase ( cls , __a , *__a , **__a) -> int: '''simple docstring''' _UpperCamelCase = GPTaTokenizer.from_pretrained(__a , *__a , **__a) return cls.from_tokenizer(__a , *__a , **__a) @classmethod def UpperCAmelCase ( cls , __a) -> str: '''simple docstring''' return cls(**__a) def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' return { "vocab": self.vocab, "merges": self.merges, "max_length": self.max_length, "pad_token_id": self.pad_token_id, } def UpperCAmelCase ( self , __a , __a = None) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.tf_tokenizer(__a) _UpperCamelCase = tf.ones_like(__a) if self.pad_token_id is not None: # pad the tokens up to max length _UpperCamelCase = max_length if max_length is not None else self.max_length if max_length is not None: _UpperCamelCase , _UpperCamelCase = pad_model_inputs( __a , max_seq_length=__a , pad_value=self.pad_token_id) return {"attention_mask": attention_mask, "input_ids": input_ids}
78
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class _UpperCAmelCase: def __init__( self , __a , __a=13 , __a=2 , __a=24 , __a=16 , __a=True , __a=True , __a=32 , __a=5 , __a=4 , __a=37 , __a="gelu" , __a=0.1 , __a=0.1 , __a=10 , __a=0.02 , __a=None , __a=2 , __a=2 , ) -> List[str]: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = patch_size _UpperCamelCase = max_length _UpperCamelCase = num_mel_bins _UpperCamelCase = is_training _UpperCamelCase = use_labels _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = type_sequence_label_size _UpperCamelCase = initializer_range _UpperCamelCase = scope _UpperCamelCase = frequency_stride _UpperCamelCase = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) _UpperCamelCase = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 _UpperCamelCase = (self.max_length - self.patch_size) // self.time_stride + 1 _UpperCamelCase = frequency_out_dimension * time_out_dimension _UpperCamelCase = num_patches + 2 def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins]) _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size) _UpperCamelCase = self.get_config() return config, input_values, labels def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__a , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def UpperCAmelCase ( self , __a , __a , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = ASTModel(config=__a) model.to(__a) model.eval() _UpperCamelCase = model(__a) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.prepare_config_and_inputs() ( ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ) = config_and_inputs _UpperCamelCase = {'''input_values''': input_values} return config, inputs_dict @require_torch class _UpperCAmelCase( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase__ = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) lowercase__ = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) lowercase__ = False lowercase__ = False lowercase__ = False lowercase__ = False def UpperCAmelCase ( self , __a , __a , __a , __a , __a) -> Optional[Any]: '''simple docstring''' if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = ASTModelTester(self) _UpperCamelCase = ConfigTester(self , config_class=__a , has_text_modality=__a , hidden_size=37) def UpperCAmelCase ( self) -> int: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='''AST does not use inputs_embeds''') def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' pass def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = model_class(__a) self.assertIsInstance(model.get_input_embeddings() , (nn.Module)) _UpperCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__a , nn.Linear)) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = model_class(__a) _UpperCamelCase = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCamelCase = [*signature.parameters.keys()] _UpperCamelCase = ['''input_values'''] self.assertListEqual(arg_names[:1] , __a) def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a) @slow def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase = ASTModel.from_pretrained(__a) self.assertIsNotNone(__a) def lowerCamelCase__ ( ) -> List[str]: """simple docstring""" _UpperCamelCase = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''', filename='''sample_audio.flac''', repo_type='''dataset''' ) _UpperCamelCase , _UpperCamelCase = torchaudio.load(__snake_case ) return audio, sampling_rate @require_torch @require_torchaudio class _UpperCAmelCase( unittest.TestCase ): @cached_property def UpperCAmelCase ( self) -> Dict: '''simple docstring''' return ( ASTFeatureExtractor.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''') if is_torchaudio_available() else None ) @slow def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.default_feature_extractor _UpperCamelCase = ASTForAudioClassification.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''').to(__a) _UpperCamelCase = self.default_feature_extractor _UpperCamelCase , _UpperCamelCase = prepare_audio() _UpperCamelCase = audio.squeeze().numpy() _UpperCamelCase = feature_extractor(__a , sampling_rate=__a , return_tensors='''pt''').to(__a) # forward pass with torch.no_grad(): _UpperCamelCase = model(**__a) # verify the logits _UpperCamelCase = torch.Size((1, 5_27)) self.assertEqual(outputs.logits.shape , __a) _UpperCamelCase = torch.tensor([-0.8760, -7.0042, -8.6602]).to(__a) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __a , atol=1e-4))
78
1
"""simple docstring""" import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class _UpperCAmelCase( unittest.TestCase ): @slow def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = FlaxMTaForConditionalGeneration.from_pretrained('''google/mt5-small''') _UpperCamelCase = AutoTokenizer.from_pretrained('''google/mt5-small''') _UpperCamelCase = tokenizer('''Hello there''' , return_tensors='''np''').input_ids _UpperCamelCase = tokenizer('''Hi I am''' , return_tensors='''np''').input_ids _UpperCamelCase = shift_tokens_right(__a , model.config.pad_token_id , model.config.decoder_start_token_id) _UpperCamelCase = model(__a , decoder_input_ids=__a).logits _UpperCamelCase = optax.softmax_cross_entropy(__a , onehot(__a , logits.shape[-1])).mean() _UpperCamelCase = -(labels.shape[-1] * loss.item()) _UpperCamelCase = -84.9127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE) < 1e-4)
78
"""simple docstring""" def lowerCamelCase__ ( ) -> list[list[int]]: """simple docstring""" return [list(range(10_00 - i, -10_00 - i, -1 ) ) for i in range(10_00 )] _a = generate_large_matrix() _a = ( [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]], [[3, 2], [1, 0]], [[7, 7, 6]], [[7, 7, 6], [-1, -2, -3]], grid, ) def lowerCamelCase__ ( __snake_case ) -> None: """simple docstring""" assert all(row == sorted(__snake_case, reverse=__snake_case ) for row in grid ) assert all(list(__snake_case ) == sorted(__snake_case, reverse=__snake_case ) for col in zip(*__snake_case ) ) def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" _UpperCamelCase = 0 _UpperCamelCase = len(__snake_case ) - 1 # Edge cases such as no values or all numbers are negative. if not array or array[0] < 0: return 0 while right + 1 > left: _UpperCamelCase = (left + right) // 2 _UpperCamelCase = array[mid] # Num must be negative and the index must be greater than or equal to 0. if num < 0 and array[mid - 1] >= 0: return mid if num >= 0: _UpperCamelCase = mid + 1 else: _UpperCamelCase = mid - 1 # No negative numbers so return the last index of the array + 1 which is the length. return len(__snake_case ) def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" _UpperCamelCase = 0 _UpperCamelCase = len(grid[0] ) for i in range(len(__snake_case ) ): _UpperCamelCase = find_negative_index(grid[i][:bound] ) total += bound return (len(__snake_case ) * len(grid[0] )) - total def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" return len([number for row in grid for number in row if number < 0] ) def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" _UpperCamelCase = 0 for row in grid: for i, number in enumerate(__snake_case ): if number < 0: total += len(__snake_case ) - i break return total def lowerCamelCase__ ( ) -> None: """simple docstring""" from timeit import timeit print('''Running benchmarks''' ) _UpperCamelCase = ( '''from __main__ import count_negatives_binary_search, ''' '''count_negatives_brute_force, count_negatives_brute_force_with_break, grid''' ) for func in ( "count_negatives_binary_search", # took 0.7727 seconds "count_negatives_brute_force_with_break", # took 4.6505 seconds "count_negatives_brute_force", # took 12.8160 seconds ): _UpperCamelCase = timeit(F'''{func}(grid=grid)''', setup=__snake_case, number=5_00 ) print(F'''{func}() took {time:0.4f} seconds''' ) if __name__ == "__main__": import doctest doctest.testmod() benchmark()
78
1
"""simple docstring""" def lowerCamelCase__ ( ) -> list[list[int]]: """simple docstring""" return [list(range(10_00 - i, -10_00 - i, -1 ) ) for i in range(10_00 )] _a = generate_large_matrix() _a = ( [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]], [[3, 2], [1, 0]], [[7, 7, 6]], [[7, 7, 6], [-1, -2, -3]], grid, ) def lowerCamelCase__ ( __snake_case ) -> None: """simple docstring""" assert all(row == sorted(__snake_case, reverse=__snake_case ) for row in grid ) assert all(list(__snake_case ) == sorted(__snake_case, reverse=__snake_case ) for col in zip(*__snake_case ) ) def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" _UpperCamelCase = 0 _UpperCamelCase = len(__snake_case ) - 1 # Edge cases such as no values or all numbers are negative. if not array or array[0] < 0: return 0 while right + 1 > left: _UpperCamelCase = (left + right) // 2 _UpperCamelCase = array[mid] # Num must be negative and the index must be greater than or equal to 0. if num < 0 and array[mid - 1] >= 0: return mid if num >= 0: _UpperCamelCase = mid + 1 else: _UpperCamelCase = mid - 1 # No negative numbers so return the last index of the array + 1 which is the length. return len(__snake_case ) def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" _UpperCamelCase = 0 _UpperCamelCase = len(grid[0] ) for i in range(len(__snake_case ) ): _UpperCamelCase = find_negative_index(grid[i][:bound] ) total += bound return (len(__snake_case ) * len(grid[0] )) - total def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" return len([number for row in grid for number in row if number < 0] ) def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" _UpperCamelCase = 0 for row in grid: for i, number in enumerate(__snake_case ): if number < 0: total += len(__snake_case ) - i break return total def lowerCamelCase__ ( ) -> None: """simple docstring""" from timeit import timeit print('''Running benchmarks''' ) _UpperCamelCase = ( '''from __main__ import count_negatives_binary_search, ''' '''count_negatives_brute_force, count_negatives_brute_force_with_break, grid''' ) for func in ( "count_negatives_binary_search", # took 0.7727 seconds "count_negatives_brute_force_with_break", # took 4.6505 seconds "count_negatives_brute_force", # took 12.8160 seconds ): _UpperCamelCase = timeit(F'''{func}(grid=grid)''', setup=__snake_case, number=5_00 ) print(F'''{func}() took {time:0.4f} seconds''' ) if __name__ == "__main__": import doctest doctest.testmod() benchmark()
78
"""simple docstring""" import copy import re class _UpperCAmelCase: lowercase__ = 'hp' lowercase__ = {} lowercase__ = None @classmethod def UpperCAmelCase ( cls , __a , __a) -> Dict: '''simple docstring''' _UpperCamelCase = prefix _UpperCamelCase = defaults cls.build_naming_info() @staticmethod def UpperCAmelCase ( __a , __a) -> Union[str, Any]: '''simple docstring''' if len(__a) == 0: return "" _UpperCamelCase = None if any(char.isdigit() for char in word): raise Exception(F'''Parameters should not contain numbers: \'{word}\' contains a number''') if word in info["short_word"]: return info["short_word"][word] for prefix_len in range(1 , len(__a) + 1): _UpperCamelCase = word[:prefix_len] if prefix in info["reverse_short_word"]: continue else: _UpperCamelCase = prefix break if short_word is None: # Paranoid fallback def int_to_alphabetic(__a): _UpperCamelCase = '''''' while integer != 0: _UpperCamelCase = chr(ord('''A''') + integer % 10) + s integer //= 10 return s _UpperCamelCase = 0 while True: _UpperCamelCase = word + '''#''' + int_to_alphabetic(__a) if sword in info["reverse_short_word"]: continue else: _UpperCamelCase = sword break _UpperCamelCase = short_word _UpperCamelCase = word return short_word @staticmethod def UpperCAmelCase ( __a , __a) -> Any: '''simple docstring''' _UpperCamelCase = param_name.split('''_''') _UpperCamelCase = [TrialShortNamer.shortname_for_word(__a , __a) for word in words] # We try to create a separatorless short name, but if there is a collision we have to fallback # to a separated short name _UpperCamelCase = ['''''', '''_'''] for separator in separators: _UpperCamelCase = separator.join(__a) if shortname not in info["reverse_short_param"]: _UpperCamelCase = shortname _UpperCamelCase = param_name return shortname return param_name @staticmethod def UpperCAmelCase ( __a , __a) -> List[str]: '''simple docstring''' _UpperCamelCase = TrialShortNamer.shortname_for_key(__a , __a) _UpperCamelCase = short_name _UpperCamelCase = param_name @classmethod def UpperCAmelCase ( cls) -> Any: '''simple docstring''' if cls.NAMING_INFO is not None: return _UpperCamelCase = { '''short_word''': {}, '''reverse_short_word''': {}, '''short_param''': {}, '''reverse_short_param''': {}, } _UpperCamelCase = list(cls.DEFAULTS.keys()) for k in field_keys: cls.add_new_param_name(__a , __a) _UpperCamelCase = info @classmethod def UpperCAmelCase ( cls , __a) -> Optional[Any]: '''simple docstring''' cls.build_naming_info() assert cls.PREFIX is not None _UpperCamelCase = [copy.copy(cls.PREFIX)] for k, v in params.items(): if k not in cls.DEFAULTS: raise Exception(F'''You should provide a default value for the param name {k} with value {v}''') if v == cls.DEFAULTS[k]: # The default value is not added to the name continue _UpperCamelCase = cls.NAMING_INFO['''short_param'''][k] if isinstance(__a , __a): _UpperCamelCase = 1 if v else 0 _UpperCamelCase = '''''' if isinstance(__a , (int, float)) else '''-''' _UpperCamelCase = F'''{key}{sep}{v}''' name.append(__a) return "_".join(__a) @classmethod def UpperCAmelCase ( cls , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = repr[len(cls.PREFIX) + 1 :] if repr == "": _UpperCamelCase = [] else: _UpperCamelCase = repr.split('''_''') _UpperCamelCase = {} for value in values: if "-" in value: _UpperCamelCase , _UpperCamelCase = value.split('''-''') else: _UpperCamelCase = re.sub('''[0-9.]''' , '''''' , __a) _UpperCamelCase = float(re.sub('''[^0-9.]''' , '''''' , __a)) _UpperCamelCase = cls.NAMING_INFO['''reverse_short_param'''][p_k] _UpperCamelCase = p_v for k in cls.DEFAULTS: if k not in parameters: _UpperCamelCase = cls.DEFAULTS[k] return parameters
78
1
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTImageProcessor, ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) def lowerCamelCase__ ( __snake_case, __snake_case=False ) -> Tuple: """simple docstring""" _UpperCamelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'''blocks.{i}.norm1.weight''', F'''vit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((F'''blocks.{i}.norm1.bias''', F'''vit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((F'''blocks.{i}.attn.proj.weight''', F'''vit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.attn.proj.bias''', F'''vit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((F'''blocks.{i}.norm2.weight''', F'''vit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((F'''blocks.{i}.norm2.bias''', F'''vit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.weight''', F'''vit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.bias''', F'''vit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.weight''', F'''vit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.bias''', F'''vit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ('''cls_token''', '''vit.embeddings.cls_token'''), ('''patch_embed.proj.weight''', '''vit.embeddings.patch_embeddings.projection.weight'''), ('''patch_embed.proj.bias''', '''vit.embeddings.patch_embeddings.projection.bias'''), ('''pos_embed''', '''vit.embeddings.position_embeddings'''), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ('''pre_logits.fc.weight''', '''pooler.dense.weight'''), ('''pre_logits.fc.bias''', '''pooler.dense.bias'''), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _UpperCamelCase = [(pair[0], pair[1][4:]) if pair[1].startswith('''vit''' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('''norm.weight''', '''vit.layernorm.weight'''), ('''norm.bias''', '''vit.layernorm.bias'''), ('''head.weight''', '''classifier.weight'''), ('''head.bias''', '''classifier.bias'''), ] ) return rename_keys def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case=False ) -> str: """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: _UpperCamelCase = '''''' else: _UpperCamelCase = '''vit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _UpperCamelCase = state_dict.pop(F'''blocks.{i}.attn.qkv.weight''' ) _UpperCamelCase = state_dict.pop(F'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict _UpperCamelCase = in_proj_weight[ : config.hidden_size, : ] _UpperCamelCase = in_proj_bias[: config.hidden_size] _UpperCamelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _UpperCamelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _UpperCamelCase = in_proj_weight[ -config.hidden_size :, : ] _UpperCamelCase = in_proj_bias[-config.hidden_size :] def lowerCamelCase__ ( __snake_case ) -> Dict: """simple docstring""" _UpperCamelCase = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(__snake_case, __snake_case ) def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> Any: """simple docstring""" _UpperCamelCase = dct.pop(__snake_case ) _UpperCamelCase = val def lowerCamelCase__ ( ) -> Dict: """simple docstring""" _UpperCamelCase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _UpperCamelCase = Image.open(requests.get(__snake_case, stream=__snake_case ).raw ) return im @torch.no_grad() def lowerCamelCase__ ( __snake_case, __snake_case ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = ViTConfig() _UpperCamelCase = False # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size if vit_name[-5:] == "in21k": _UpperCamelCase = True _UpperCamelCase = int(vit_name[-12:-10] ) _UpperCamelCase = int(vit_name[-9:-6] ) else: _UpperCamelCase = 10_00 _UpperCamelCase = '''huggingface/label-files''' _UpperCamelCase = '''imagenet-1k-id2label.json''' _UpperCamelCase = json.load(open(hf_hub_download(__snake_case, __snake_case, repo_type='''dataset''' ), '''r''' ) ) _UpperCamelCase = {int(__snake_case ): v for k, v in idalabel.items()} _UpperCamelCase = idalabel _UpperCamelCase = {v: k for k, v in idalabel.items()} _UpperCamelCase = int(vit_name[-6:-4] ) _UpperCamelCase = int(vit_name[-3:] ) # size of the architecture if "deit" in vit_name: if vit_name[9:].startswith('''tiny''' ): _UpperCamelCase = 1_92 _UpperCamelCase = 7_68 _UpperCamelCase = 12 _UpperCamelCase = 3 elif vit_name[9:].startswith('''small''' ): _UpperCamelCase = 3_84 _UpperCamelCase = 15_36 _UpperCamelCase = 12 _UpperCamelCase = 6 else: pass else: if vit_name[4:].startswith('''small''' ): _UpperCamelCase = 7_68 _UpperCamelCase = 23_04 _UpperCamelCase = 8 _UpperCamelCase = 8 elif vit_name[4:].startswith('''base''' ): pass elif vit_name[4:].startswith('''large''' ): _UpperCamelCase = 10_24 _UpperCamelCase = 40_96 _UpperCamelCase = 24 _UpperCamelCase = 16 elif vit_name[4:].startswith('''huge''' ): _UpperCamelCase = 12_80 _UpperCamelCase = 51_20 _UpperCamelCase = 32 _UpperCamelCase = 16 # load original model from timm _UpperCamelCase = timm.create_model(__snake_case, pretrained=__snake_case ) timm_model.eval() # load state_dict of original model, remove and rename some keys _UpperCamelCase = timm_model.state_dict() if base_model: remove_classification_head_(__snake_case ) _UpperCamelCase = create_rename_keys(__snake_case, __snake_case ) for src, dest in rename_keys: rename_key(__snake_case, __snake_case, __snake_case ) read_in_q_k_v(__snake_case, __snake_case, __snake_case ) # load HuggingFace model if vit_name[-5:] == "in21k": _UpperCamelCase = ViTModel(__snake_case ).eval() else: _UpperCamelCase = ViTForImageClassification(__snake_case ).eval() model.load_state_dict(__snake_case ) # Check outputs on an image, prepared by ViTImageProcessor/DeiTImageProcessor if "deit" in vit_name: _UpperCamelCase = DeiTImageProcessor(size=config.image_size ) else: _UpperCamelCase = ViTImageProcessor(size=config.image_size ) _UpperCamelCase = image_processor(images=prepare_img(), return_tensors='''pt''' ) _UpperCamelCase = encoding['''pixel_values'''] _UpperCamelCase = model(__snake_case ) if base_model: _UpperCamelCase = timm_model.forward_features(__snake_case ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(__snake_case, outputs.pooler_output, atol=1e-3 ) else: _UpperCamelCase = timm_model(__snake_case ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__snake_case, outputs.logits, atol=1e-3 ) Path(__snake_case ).mkdir(exist_ok=__snake_case ) print(F'''Saving model {vit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(__snake_case ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(__snake_case ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( """--vit_name""", default="""vit_base_patch16_224""", type=str, help="""Name of the ViT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) _a = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path)
78
"""simple docstring""" import os import time import pytest from datasets.utils.filelock import FileLock, Timeout def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = FileLock(str(tmpdir / '''foo.lock''' ) ) _UpperCamelCase = FileLock(str(tmpdir / '''foo.lock''' ) ) _UpperCamelCase = 0.01 with locka.acquire(): with pytest.raises(__snake_case ): _UpperCamelCase = time.time() locka.acquire(__snake_case ) assert time.time() - _start > timeout def lowerCamelCase__ ( __snake_case ) -> List[Any]: """simple docstring""" _UpperCamelCase = '''a''' * 10_00 + '''.lock''' _UpperCamelCase = FileLock(str(tmpdir / filename ) ) assert locka._lock_file.endswith('''.lock''' ) assert not locka._lock_file.endswith(__snake_case ) assert len(os.path.basename(locka._lock_file ) ) <= 2_55 _UpperCamelCase = FileLock(tmpdir / filename ) with locka.acquire(): with pytest.raises(__snake_case ): locka.acquire(0 )
78
1
"""simple docstring""" from ....configuration_utils import PretrainedConfig from ....utils import logging _a = logging.get_logger(__name__) _a = { """Visual-Attention-Network/van-base""": ( """https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json""" ), } class _UpperCAmelCase( lowerCamelCase ): lowercase__ = 'van' def __init__( self , __a=2_24 , __a=3 , __a=[7, 3, 3, 3] , __a=[4, 2, 2, 2] , __a=[64, 1_28, 3_20, 5_12] , __a=[3, 3, 12, 3] , __a=[8, 8, 4, 4] , __a="gelu" , __a=0.02 , __a=1e-6 , __a=1e-2 , __a=0.0 , __a=0.0 , **__a , ) -> Dict: '''simple docstring''' super().__init__(**__a) _UpperCamelCase = image_size _UpperCamelCase = num_channels _UpperCamelCase = patch_sizes _UpperCamelCase = strides _UpperCamelCase = hidden_sizes _UpperCamelCase = depths _UpperCamelCase = mlp_ratios _UpperCamelCase = hidden_act _UpperCamelCase = initializer_range _UpperCamelCase = layer_norm_eps _UpperCamelCase = layer_scale_init_value _UpperCamelCase = drop_path_rate _UpperCamelCase = dropout_rate
78
"""simple docstring""" from math import sqrt def lowerCamelCase__ ( __snake_case ) -> bool: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and ( number >= 0 ), "'number' must been an int and positive" _UpperCamelCase = True # 0 and 1 are none primes. if number <= 1: _UpperCamelCase = False for divisor in range(2, int(round(sqrt(__snake_case ) ) ) + 1 ): # if 'number' divisible by 'divisor' then sets 'status' # of false and break up the loop. if number % divisor == 0: _UpperCamelCase = False break # precondition assert isinstance(__snake_case, __snake_case ), "'status' must been from type bool" return status def lowerCamelCase__ ( __snake_case ) -> Dict: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n > 2), "'N' must been an int and > 2" # beginList: contains all natural numbers from 2 up to N _UpperCamelCase = list(range(2, n + 1 ) ) _UpperCamelCase = [] # this list will be returns. # actual sieve of erathostenes for i in range(len(__snake_case ) ): for j in range(i + 1, len(__snake_case ) ): if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0): _UpperCamelCase = 0 # filters actual prime numbers. _UpperCamelCase = [x for x in begin_list if x != 0] # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type list" return ans def lowerCamelCase__ ( __snake_case ) -> List[str]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n > 2), "'N' must been an int and > 2" _UpperCamelCase = [] # iterates over all numbers between 2 up to N+1 # if a number is prime then appends to list 'ans' for number in range(2, n + 1 ): if is_prime(__snake_case ): ans.append(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type list" return ans def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and number >= 0, "'number' must been an int and >= 0" _UpperCamelCase = [] # this list will be returns of the function. # potential prime number factors. _UpperCamelCase = 2 _UpperCamelCase = number if number == 0 or number == 1: ans.append(__snake_case ) # if 'number' not prime then builds the prime factorization of 'number' elif not is_prime(__snake_case ): while quotient != 1: if is_prime(__snake_case ) and (quotient % factor == 0): ans.append(__snake_case ) quotient /= factor else: factor += 1 else: ans.append(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type list" return ans def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and ( number >= 0 ), "'number' bust been an int and >= 0" _UpperCamelCase = 0 # prime factorization of 'number' _UpperCamelCase = prime_factorization(__snake_case ) _UpperCamelCase = max(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type int" return ans def lowerCamelCase__ ( __snake_case ) -> int: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and ( number >= 0 ), "'number' bust been an int and >= 0" _UpperCamelCase = 0 # prime factorization of 'number' _UpperCamelCase = prime_factorization(__snake_case ) _UpperCamelCase = min(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ), "'ans' must been from type int" return ans def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ), "'number' must been an int" assert isinstance(number % 2 == 0, __snake_case ), "compare bust been from type bool" return number % 2 == 0 def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ), "'number' must been an int" assert isinstance(number % 2 != 0, __snake_case ), "compare bust been from type bool" return number % 2 != 0 def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" assert ( isinstance(__snake_case, __snake_case ) and (number > 2) and is_even(__snake_case ) ), "'number' must been an int, even and > 2" _UpperCamelCase = [] # this list will returned # creates a list of prime numbers between 2 up to 'number' _UpperCamelCase = get_prime_numbers(__snake_case ) _UpperCamelCase = len(__snake_case ) # run variable for while-loops. _UpperCamelCase = 0 _UpperCamelCase = None # exit variable. for break up the loops _UpperCamelCase = True while i < len_pn and loop: _UpperCamelCase = i + 1 while j < len_pn and loop: if prime_numbers[i] + prime_numbers[j] == number: _UpperCamelCase = False ans.append(prime_numbers[i] ) ans.append(prime_numbers[j] ) j += 1 i += 1 # precondition assert ( isinstance(__snake_case, __snake_case ) and (len(__snake_case ) == 2) and (ans[0] + ans[1] == number) and is_prime(ans[0] ) and is_prime(ans[1] ) ), "'ans' must contains two primes. And sum of elements must been eq 'number'" return ans def lowerCamelCase__ ( __snake_case, __snake_case ) -> str: """simple docstring""" assert ( isinstance(__snake_case, __snake_case ) and isinstance(__snake_case, __snake_case ) and (numbera >= 0) and (numbera >= 0) ), "'number1' and 'number2' must been positive integer." _UpperCamelCase = 0 while numbera != 0: _UpperCamelCase = numbera % numbera _UpperCamelCase = numbera _UpperCamelCase = rest # precondition assert isinstance(__snake_case, __snake_case ) and ( numbera >= 0 ), "'number' must been from type int and positive" return numbera def lowerCamelCase__ ( __snake_case, __snake_case ) -> List[str]: """simple docstring""" assert ( isinstance(__snake_case, __snake_case ) and isinstance(__snake_case, __snake_case ) and (numbera >= 1) and (numbera >= 1) ), "'number1' and 'number2' must been positive integer." _UpperCamelCase = 1 # actual answer that will be return. # for kgV (x,1) if numbera > 1 and numbera > 1: # builds the prime factorization of 'number1' and 'number2' _UpperCamelCase = prime_factorization(__snake_case ) _UpperCamelCase = prime_factorization(__snake_case ) elif numbera == 1 or numbera == 1: _UpperCamelCase = [] _UpperCamelCase = [] _UpperCamelCase = max(__snake_case, __snake_case ) _UpperCamelCase = 0 _UpperCamelCase = 0 _UpperCamelCase = [] # captured numbers int both 'primeFac1' and 'primeFac2' # iterates through primeFac1 for n in prime_fac_a: if n not in done: if n in prime_fac_a: _UpperCamelCase = prime_fac_a.count(__snake_case ) _UpperCamelCase = prime_fac_a.count(__snake_case ) for _ in range(max(__snake_case, __snake_case ) ): ans *= n else: _UpperCamelCase = prime_fac_a.count(__snake_case ) for _ in range(__snake_case ): ans *= n done.append(__snake_case ) # iterates through primeFac2 for n in prime_fac_a: if n not in done: _UpperCamelCase = prime_fac_a.count(__snake_case ) for _ in range(__snake_case ): ans *= n done.append(__snake_case ) # precondition assert isinstance(__snake_case, __snake_case ) and ( ans >= 0 ), "'ans' must been from type int and positive" return ans def lowerCamelCase__ ( __snake_case ) -> Tuple: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n >= 0), "'number' must been a positive int" _UpperCamelCase = 0 _UpperCamelCase = 2 # this variable holds the answer while index < n: index += 1 ans += 1 # counts to the next number # if ans not prime then # runs to the next prime number. while not is_prime(__snake_case ): ans += 1 # precondition assert isinstance(__snake_case, __snake_case ) and is_prime( __snake_case ), "'ans' must been a prime number and from type int" return ans def lowerCamelCase__ ( __snake_case, __snake_case ) -> Tuple: """simple docstring""" assert ( is_prime(__snake_case ) and is_prime(__snake_case ) and (p_number_a < p_number_a) ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'" _UpperCamelCase = p_number_a + 1 # jump to the next number _UpperCamelCase = [] # this list will be returns. # if number is not prime then # fetch the next prime number. while not is_prime(__snake_case ): number += 1 while number < p_number_a: ans.append(__snake_case ) number += 1 # fetch the next prime number. while not is_prime(__snake_case ): number += 1 # precondition assert ( isinstance(__snake_case, __snake_case ) and ans[0] != p_number_a and ans[len(__snake_case ) - 1] != p_number_a ), "'ans' must been a list without the arguments" # 'ans' contains not 'pNumber1' and 'pNumber2' ! return ans def lowerCamelCase__ ( __snake_case ) -> List[Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n >= 1), "'n' must been int and >= 1" _UpperCamelCase = [] # will be returned. for divisor in range(1, n + 1 ): if n % divisor == 0: ans.append(__snake_case ) # precondition assert ans[0] == 1 and ans[len(__snake_case ) - 1] == n, "Error in function getDivisiors(...)" return ans def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and ( number > 1 ), "'number' must been an int and >= 1" _UpperCamelCase = get_divisors(__snake_case ) # precondition assert ( isinstance(__snake_case, __snake_case ) and (divisors[0] == 1) and (divisors[len(__snake_case ) - 1] == number) ), "Error in help-function getDivisiors(...)" # summed all divisors up to 'number' (exclusive), hence [:-1] return sum(divisors[:-1] ) == number def lowerCamelCase__ ( __snake_case, __snake_case ) -> Optional[Any]: """simple docstring""" assert ( isinstance(__snake_case, __snake_case ) and isinstance(__snake_case, __snake_case ) and (denominator != 0) ), "The arguments must been from type int and 'denominator' != 0" # build the greatest common divisor of numerator and denominator. _UpperCamelCase = gcd(abs(__snake_case ), abs(__snake_case ) ) # precondition assert ( isinstance(__snake_case, __snake_case ) and (numerator % gcd_of_fraction == 0) and (denominator % gcd_of_fraction == 0) ), "Error in function gcd(...,...)" return (numerator // gcd_of_fraction, denominator // gcd_of_fraction) def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n >= 0), "'n' must been a int and >= 0" _UpperCamelCase = 1 # this will be return. for factor in range(1, n + 1 ): ans *= factor return ans def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" assert isinstance(__snake_case, __snake_case ) and (n >= 0), "'n' must been an int and >= 0" _UpperCamelCase = 0 _UpperCamelCase = 1 _UpperCamelCase = 1 # this will be return for _ in range(n - 1 ): _UpperCamelCase = ans ans += fiba _UpperCamelCase = tmp return ans
78
1
"""simple docstring""" from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_OBJECT_DETECTION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING _a = logging.get_logger(__name__) _a = Dict[str, Any] _a = List[Prediction] @add_end_docstrings(lowerCamelCase ) class _UpperCAmelCase( lowerCamelCase ): def __init__( self , *__a , **__a) -> Union[str, Any]: '''simple docstring''' super().__init__(*__a , **__a) if self.framework == "tf": raise ValueError(F'''The {self.__class__} is only available in PyTorch.''') requires_backends(self , '''vision''') self.check_model_type( dict(MODEL_FOR_OBJECT_DETECTION_MAPPING.items() + MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING.items())) def UpperCAmelCase ( self , **__a) -> List[Any]: '''simple docstring''' _UpperCamelCase = {} if "threshold" in kwargs: _UpperCamelCase = kwargs['''threshold'''] return {}, {}, postprocess_kwargs def __call__( self , *__a , **__a) -> Union[Predictions, List[Prediction]]: '''simple docstring''' return super().__call__(*__a , **__a) def UpperCAmelCase ( self , __a) -> Any: '''simple docstring''' _UpperCamelCase = load_image(__a) _UpperCamelCase = torch.IntTensor([[image.height, image.width]]) _UpperCamelCase = self.image_processor(images=[image] , return_tensors='''pt''') if self.tokenizer is not None: _UpperCamelCase = self.tokenizer(text=inputs['''words'''] , boxes=inputs['''boxes'''] , return_tensors='''pt''') _UpperCamelCase = target_size return inputs def UpperCAmelCase ( self , __a) -> int: '''simple docstring''' _UpperCamelCase = model_inputs.pop('''target_size''') _UpperCamelCase = self.model(**__a) _UpperCamelCase = outputs.__class__({'''target_size''': target_size, **outputs}) if self.tokenizer is not None: _UpperCamelCase = model_inputs['''bbox'''] return model_outputs def UpperCAmelCase ( self , __a , __a=0.9) -> List[Any]: '''simple docstring''' _UpperCamelCase = model_outputs['''target_size'''] if self.tokenizer is not None: # This is a LayoutLMForTokenClassification variant. # The OCR got the boxes and the model classified the words. _UpperCamelCase , _UpperCamelCase = target_size[0].tolist() def unnormalize(__a): return self._get_bounding_box( torch.Tensor( [ (width * bbox[0] / 10_00), (height * bbox[1] / 10_00), (width * bbox[2] / 10_00), (height * bbox[3] / 10_00), ])) _UpperCamelCase , _UpperCamelCase = model_outputs['''logits'''].squeeze(0).softmax(dim=-1).max(dim=-1) _UpperCamelCase = [self.model.config.idalabel[prediction] for prediction in classes.tolist()] _UpperCamelCase = [unnormalize(__a) for bbox in model_outputs['''bbox'''].squeeze(0)] _UpperCamelCase = ['''score''', '''label''', '''box'''] _UpperCamelCase = [dict(zip(__a , __a)) for vals in zip(scores.tolist() , __a , __a) if vals[0] > threshold] else: # This is a regular ForObjectDetectionModel _UpperCamelCase = self.image_processor.post_process_object_detection(__a , __a , __a) _UpperCamelCase = raw_annotations[0] _UpperCamelCase = raw_annotation['''scores'''] _UpperCamelCase = raw_annotation['''labels'''] _UpperCamelCase = raw_annotation['''boxes'''] _UpperCamelCase = scores.tolist() _UpperCamelCase = [self.model.config.idalabel[label.item()] for label in labels] _UpperCamelCase = [self._get_bounding_box(__a) for box in boxes] # {"scores": [...], ...} --> [{"score":x, ...}, ...] _UpperCamelCase = ['''score''', '''label''', '''box'''] _UpperCamelCase = [ dict(zip(__a , __a)) for vals in zip(raw_annotation['''scores'''] , raw_annotation['''labels'''] , raw_annotation['''boxes''']) ] return annotation def UpperCAmelCase ( self , __a) -> Dict[str, int]: '''simple docstring''' if self.framework != "pt": raise ValueError('''The ObjectDetectionPipeline is only available in PyTorch.''') _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = box.int().tolist() _UpperCamelCase = { '''xmin''': xmin, '''ymin''': ymin, '''xmax''': xmax, '''ymax''': ymax, } return bbox
78
"""simple docstring""" import logging from dataclasses import dataclass, field from pathlib import Path from typing import Optional, Union from .generation.configuration_utils import GenerationConfig from .training_args import TrainingArguments from .utils import add_start_docstrings _a = logging.getLogger(__name__) @dataclass @add_start_docstrings(TrainingArguments.__doc__ ) class _UpperCAmelCase( lowerCamelCase ): lowercase__ = field(default=lowerCamelCase , metadata={'help': 'Whether to use SortishSampler or not.'} ) lowercase__ = field( default=lowerCamelCase , metadata={'help': 'Whether to use generate to calculate generative metrics (ROUGE, BLEU).'} ) lowercase__ = field( default=lowerCamelCase , metadata={ 'help': ( 'The `max_length` to use on each evaluation loop when `predict_with_generate=True`. Will default ' 'to the `max_length` value of the model configuration.' ) } , ) lowercase__ = field( default=lowerCamelCase , metadata={ 'help': ( 'The `num_beams` to use on each evaluation loop when `predict_with_generate=True`. Will default ' 'to the `num_beams` value of the model configuration.' ) } , ) lowercase__ = field( default=lowerCamelCase , metadata={ 'help': 'Model id, file path or url pointing to a GenerationConfig json file, to use during prediction.' } , ) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = super().to_dict() for k, v in d.items(): if isinstance(__a , __a): _UpperCamelCase = v.to_dict() return d
78
1
"""simple docstring""" from collections import OrderedDict from typing import Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...feature_extraction_utils import FeatureExtractionMixin from ...onnx import OnnxConfig from ...onnx.utils import compute_effective_axis_dimension from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType, logging _a = logging.get_logger(__name__) _a = { """deepmind/language-perceiver""": """https://huggingface.co/deepmind/language-perceiver/resolve/main/config.json""", # See all Perceiver models at https://huggingface.co/models?filter=perceiver } class _UpperCAmelCase( lowerCamelCase ): lowercase__ = 'perceiver' def __init__( self , __a=2_56 , __a=12_80 , __a=7_68 , __a=1 , __a=26 , __a=8 , __a=8 , __a=None , __a=None , __a="kv" , __a=1 , __a=1 , __a="gelu" , __a=0.1 , __a=0.02 , __a=1e-12 , __a=True , __a=2_62 , __a=20_48 , __a=56 , __a=[3_68, 4_96] , __a=16 , __a=19_20 , __a=16 , __a=[1, 16, 2_24, 2_24] , **__a , ) -> int: '''simple docstring''' super().__init__(**__a) _UpperCamelCase = num_latents _UpperCamelCase = d_latents _UpperCamelCase = d_model _UpperCamelCase = num_blocks _UpperCamelCase = num_self_attends_per_block _UpperCamelCase = num_self_attention_heads _UpperCamelCase = num_cross_attention_heads _UpperCamelCase = qk_channels _UpperCamelCase = v_channels _UpperCamelCase = cross_attention_shape_for_attention _UpperCamelCase = self_attention_widening_factor _UpperCamelCase = cross_attention_widening_factor _UpperCamelCase = hidden_act _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = initializer_range _UpperCamelCase = layer_norm_eps _UpperCamelCase = use_query_residual # masked language modeling attributes _UpperCamelCase = vocab_size _UpperCamelCase = max_position_embeddings # image classification attributes _UpperCamelCase = image_size # flow attributes _UpperCamelCase = train_size # multimodal autoencoding attributes _UpperCamelCase = num_frames _UpperCamelCase = audio_samples_per_frame _UpperCamelCase = samples_per_patch _UpperCamelCase = output_shape class _UpperCAmelCase( lowerCamelCase ): @property def UpperCAmelCase ( self) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": _UpperCamelCase = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: _UpperCamelCase = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''inputs''', dynamic_axis), ('''attention_mask''', dynamic_axis), ]) @property def UpperCAmelCase ( self) -> float: '''simple docstring''' return 1e-4 def UpperCAmelCase ( self , __a , __a = -1 , __a = -1 , __a = -1 , __a = False , __a = None , __a = 3 , __a = 40 , __a = 40 , ) -> Mapping[str, Any]: '''simple docstring''' # copied from `transformers.onnx.config.OnnxConfig` and slightly altered/simplified if isinstance(__a , __a): # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX _UpperCamelCase = compute_effective_axis_dimension( __a , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX _UpperCamelCase = preprocessor.num_special_tokens_to_add(__a) _UpperCamelCase = compute_effective_axis_dimension( __a , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=__a) # Generate dummy inputs according to compute batch and sequence _UpperCamelCase = [''' '''.join(['''a''']) * seq_length] * batch_size _UpperCamelCase = dict(preprocessor(__a , return_tensors=__a)) _UpperCamelCase = inputs.pop('''input_ids''') return inputs elif isinstance(__a , __a) and preprocessor.model_input_names[0] == "pixel_values": # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX _UpperCamelCase = compute_effective_axis_dimension(__a , fixed_dimension=OnnxConfig.default_fixed_batch) _UpperCamelCase = self._generate_dummy_images(__a , __a , __a , __a) _UpperCamelCase = dict(preprocessor(images=__a , return_tensors=__a)) _UpperCamelCase = inputs.pop('''pixel_values''') return inputs else: raise ValueError( '''Unable to generate dummy inputs for the model. Please provide a tokenizer or a preprocessor.''')
78
"""simple docstring""" import argparse import torch from transformers import BlenderbotConfig, BlenderbotForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) _a = [ ["""attention""", """attn"""], ["""encoder_attention""", """encoder_attn"""], ["""q_lin""", """q_proj"""], ["""k_lin""", """k_proj"""], ["""v_lin""", """v_proj"""], ["""out_lin""", """out_proj"""], ["""norm_embeddings""", """layernorm_embedding"""], ["""position_embeddings""", """embed_positions"""], ["""embeddings""", """embed_tokens"""], ["""ffn.lin""", """fc"""], ] def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" if k == "embeddings.weight": return "shared.weight" for parlai_name, hf_name in PATTERNS: _UpperCamelCase = k.replace(__snake_case, __snake_case ) if k.startswith('''encoder''' ): _UpperCamelCase = k.replace('''.attn''', '''.self_attn''' ) _UpperCamelCase = k.replace('''norm1''', '''self_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm2''', '''final_layer_norm''' ) elif k.startswith('''decoder''' ): _UpperCamelCase = k.replace('''norm1''', '''self_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm2''', '''encoder_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm3''', '''final_layer_norm''' ) return k def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" _UpperCamelCase = [ '''model.encoder.layernorm_embedding.weight''', '''model.encoder.layernorm_embedding.bias''', '''model.decoder.layernorm_embedding.weight''', '''model.decoder.layernorm_embedding.bias''', ] for k in keys: _UpperCamelCase = sd.pop(__snake_case ) _UpperCamelCase = k.replace('''layernorm_embedding''', '''layer_norm''' ) assert new_k not in sd _UpperCamelCase = v _a = ["""START"""] @torch.no_grad() def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> int: """simple docstring""" _UpperCamelCase = torch.load(__snake_case, map_location='''cpu''' ) _UpperCamelCase = model['''model'''] _UpperCamelCase = BlenderbotConfig.from_json_file(__snake_case ) _UpperCamelCase = BlenderbotForConditionalGeneration(__snake_case ) _UpperCamelCase = m.model.state_dict().keys() _UpperCamelCase = [] _UpperCamelCase = {} for k, v in sd.items(): if k in IGNORE_KEYS: continue _UpperCamelCase = rename_state_dict_key(__snake_case ) if new_k not in valid_keys: failures.append([k, new_k] ) else: _UpperCamelCase = v if cfg.normalize_before: # Blenderbot-3B checkpoints. Rename layernorm_embedding -> layer_norm rename_layernorm_keys(__snake_case ) m.model.load_state_dict(__snake_case, strict=__snake_case ) m.half() m.save_pretrained(__snake_case ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument("""--src_path""", type=str, help="""like blenderbot-model.bin""") parser.add_argument("""--save_dir""", default="""hf_blenderbot""", type=str, help="""Where to save converted model.""") parser.add_argument( """--hf_config_json""", default="""blenderbot-3b-config.json""", type=str, help="""Path to config to use""" ) _a = parser.parse_args() convert_parlai_checkpoint(args.src_path, args.save_dir, args.hf_config_json)
78
1
"""simple docstring""" import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal _a = datasets.utils.logging.get_logger(__name__) _a = ["""names""", """prefix"""] _a = ["""warn_bad_lines""", """error_bad_lines""", """mangle_dupe_cols"""] _a = ["""encoding_errors""", """on_bad_lines"""] _a = ["""date_format"""] @dataclass class _UpperCAmelCase( datasets.BuilderConfig ): lowercase__ = "," lowercase__ = None lowercase__ = "infer" lowercase__ = None lowercase__ = None lowercase__ = None lowercase__ = None lowercase__ = None lowercase__ = True lowercase__ = None lowercase__ = None lowercase__ = None lowercase__ = None lowercase__ = False lowercase__ = None lowercase__ = None lowercase__ = None lowercase__ = True lowercase__ = True lowercase__ = False lowercase__ = True lowercase__ = None lowercase__ = "." lowercase__ = None lowercase__ = '"' lowercase__ = 0 lowercase__ = None lowercase__ = None lowercase__ = None lowercase__ = None lowercase__ = True lowercase__ = True lowercase__ = 0 lowercase__ = True lowercase__ = False lowercase__ = None lowercase__ = 1_00_00 lowercase__ = None lowercase__ = "strict" lowercase__ = "error" lowercase__ = None def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' if self.delimiter is not None: _UpperCamelCase = self.delimiter if self.column_names is not None: _UpperCamelCase = self.column_names @property def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = { '''sep''': self.sep, '''header''': self.header, '''names''': self.names, '''index_col''': self.index_col, '''usecols''': self.usecols, '''prefix''': self.prefix, '''mangle_dupe_cols''': self.mangle_dupe_cols, '''engine''': self.engine, '''converters''': self.converters, '''true_values''': self.true_values, '''false_values''': self.false_values, '''skipinitialspace''': self.skipinitialspace, '''skiprows''': self.skiprows, '''nrows''': self.nrows, '''na_values''': self.na_values, '''keep_default_na''': self.keep_default_na, '''na_filter''': self.na_filter, '''verbose''': self.verbose, '''skip_blank_lines''': self.skip_blank_lines, '''thousands''': self.thousands, '''decimal''': self.decimal, '''lineterminator''': self.lineterminator, '''quotechar''': self.quotechar, '''quoting''': self.quoting, '''escapechar''': self.escapechar, '''comment''': self.comment, '''encoding''': self.encoding, '''dialect''': self.dialect, '''error_bad_lines''': self.error_bad_lines, '''warn_bad_lines''': self.warn_bad_lines, '''skipfooter''': self.skipfooter, '''doublequote''': self.doublequote, '''memory_map''': self.memory_map, '''float_precision''': self.float_precision, '''chunksize''': self.chunksize, '''encoding_errors''': self.encoding_errors, '''on_bad_lines''': self.on_bad_lines, '''date_format''': self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , __a): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class _UpperCAmelCase( datasets.ArrowBasedBuilder ): lowercase__ = CsvConfig def UpperCAmelCase ( self) -> str: '''simple docstring''' return datasets.DatasetInfo(features=self.config.features) def UpperCAmelCase ( self , __a) -> List[str]: '''simple docstring''' if not self.config.data_files: raise ValueError(F'''At least one data file must be specified, but got data_files={self.config.data_files}''') _UpperCamelCase = dl_manager.download_and_extract(self.config.data_files) if isinstance(__a , (str, list, tuple)): _UpperCamelCase = data_files if isinstance(__a , __a): _UpperCamelCase = [files] _UpperCamelCase = [dl_manager.iter_files(__a) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'''files''': files})] _UpperCamelCase = [] for split_name, files in data_files.items(): if isinstance(__a , __a): _UpperCamelCase = [files] _UpperCamelCase = [dl_manager.iter_files(__a) for file in files] splits.append(datasets.SplitGenerator(name=__a , gen_kwargs={'''files''': files})) return splits def UpperCAmelCase ( self , __a) -> pa.Table: '''simple docstring''' if self.config.features is not None: _UpperCamelCase = self.config.features.arrow_schema if all(not require_storage_cast(__a) for feature in self.config.features.values()): # cheaper cast _UpperCamelCase = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=__a) else: # more expensive cast; allows str <-> int/float or str to Audio for example _UpperCamelCase = table_cast(__a , __a) return pa_table def UpperCAmelCase ( self , __a) -> List[str]: '''simple docstring''' _UpperCamelCase = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str _UpperCamelCase = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(__a) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values()) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(__a)): _UpperCamelCase = pd.read_csv(__a , iterator=__a , dtype=__a , **self.config.pd_read_csv_kwargs) try: for batch_idx, df in enumerate(__a): _UpperCamelCase = pa.Table.from_pandas(__a) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(__a) except ValueError as e: logger.error(F'''Failed to read file \'{file}\' with error {type(__a)}: {e}''') raise
78
"""simple docstring""" import argparse import os.path as osp import re import torch from safetensors.torch import load_file, save_file # =================# # UNet Conversion # # =================# _a = [ # (stable-diffusion, HF Diffusers) ("""time_embed.0.weight""", """time_embedding.linear_1.weight"""), ("""time_embed.0.bias""", """time_embedding.linear_1.bias"""), ("""time_embed.2.weight""", """time_embedding.linear_2.weight"""), ("""time_embed.2.bias""", """time_embedding.linear_2.bias"""), ("""input_blocks.0.0.weight""", """conv_in.weight"""), ("""input_blocks.0.0.bias""", """conv_in.bias"""), ("""out.0.weight""", """conv_norm_out.weight"""), ("""out.0.bias""", """conv_norm_out.bias"""), ("""out.2.weight""", """conv_out.weight"""), ("""out.2.bias""", """conv_out.bias"""), ] _a = [ # (stable-diffusion, HF Diffusers) ("""in_layers.0""", """norm1"""), ("""in_layers.2""", """conv1"""), ("""out_layers.0""", """norm2"""), ("""out_layers.3""", """conv2"""), ("""emb_layers.1""", """time_emb_proj"""), ("""skip_connection""", """conv_shortcut"""), ] _a = [] # hardcoded number of downblocks and resnets/attentions... # would need smarter logic for other networks. for i in range(4): # loop over downblocks/upblocks for j in range(2): # loop over resnets/attentions for downblocks _a = F"""down_blocks.{i}.resnets.{j}.""" _a = F"""input_blocks.{3*i + j + 1}.0.""" unet_conversion_map_layer.append((sd_down_res_prefix, hf_down_res_prefix)) if i < 3: # no attention layers in down_blocks.3 _a = F"""down_blocks.{i}.attentions.{j}.""" _a = F"""input_blocks.{3*i + j + 1}.1.""" unet_conversion_map_layer.append((sd_down_atn_prefix, hf_down_atn_prefix)) for j in range(3): # loop over resnets/attentions for upblocks _a = F"""up_blocks.{i}.resnets.{j}.""" _a = F"""output_blocks.{3*i + j}.0.""" unet_conversion_map_layer.append((sd_up_res_prefix, hf_up_res_prefix)) if i > 0: # no attention layers in up_blocks.0 _a = F"""up_blocks.{i}.attentions.{j}.""" _a = F"""output_blocks.{3*i + j}.1.""" unet_conversion_map_layer.append((sd_up_atn_prefix, hf_up_atn_prefix)) if i < 3: # no downsample in down_blocks.3 _a = F"""down_blocks.{i}.downsamplers.0.conv.""" _a = F"""input_blocks.{3*(i+1)}.0.op.""" unet_conversion_map_layer.append((sd_downsample_prefix, hf_downsample_prefix)) # no upsample in up_blocks.3 _a = F"""up_blocks.{i}.upsamplers.0.""" _a = F"""output_blocks.{3*i + 2}.{1 if i == 0 else 2}.""" unet_conversion_map_layer.append((sd_upsample_prefix, hf_upsample_prefix)) _a = """mid_block.attentions.0.""" _a = """middle_block.1.""" unet_conversion_map_layer.append((sd_mid_atn_prefix, hf_mid_atn_prefix)) for j in range(2): _a = F"""mid_block.resnets.{j}.""" _a = F"""middle_block.{2*j}.""" unet_conversion_map_layer.append((sd_mid_res_prefix, hf_mid_res_prefix)) def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" _UpperCamelCase = {k: k for k in unet_state_dict.keys()} for sd_name, hf_name in unet_conversion_map: _UpperCamelCase = sd_name for k, v in mapping.items(): if "resnets" in k: for sd_part, hf_part in unet_conversion_map_resnet: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v for k, v in mapping.items(): for sd_part, hf_part in unet_conversion_map_layer: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v _UpperCamelCase = {v: unet_state_dict[k] for k, v in mapping.items()} return new_state_dict # ================# # VAE Conversion # # ================# _a = [ # (stable-diffusion, HF Diffusers) ("""nin_shortcut""", """conv_shortcut"""), ("""norm_out""", """conv_norm_out"""), ("""mid.attn_1.""", """mid_block.attentions.0."""), ] for i in range(4): # down_blocks have two resnets for j in range(2): _a = F"""encoder.down_blocks.{i}.resnets.{j}.""" _a = F"""encoder.down.{i}.block.{j}.""" vae_conversion_map.append((sd_down_prefix, hf_down_prefix)) if i < 3: _a = F"""down_blocks.{i}.downsamplers.0.""" _a = F"""down.{i}.downsample.""" vae_conversion_map.append((sd_downsample_prefix, hf_downsample_prefix)) _a = F"""up_blocks.{i}.upsamplers.0.""" _a = F"""up.{3-i}.upsample.""" vae_conversion_map.append((sd_upsample_prefix, hf_upsample_prefix)) # up_blocks have three resnets # also, up blocks in hf are numbered in reverse from sd for j in range(3): _a = F"""decoder.up_blocks.{i}.resnets.{j}.""" _a = F"""decoder.up.{3-i}.block.{j}.""" vae_conversion_map.append((sd_up_prefix, hf_up_prefix)) # this part accounts for mid blocks in both the encoder and the decoder for i in range(2): _a = F"""mid_block.resnets.{i}.""" _a = F"""mid.block_{i+1}.""" vae_conversion_map.append((sd_mid_res_prefix, hf_mid_res_prefix)) _a = [ # (stable-diffusion, HF Diffusers) ("""norm.""", """group_norm."""), ("""q.""", """query."""), ("""k.""", """key."""), ("""v.""", """value."""), ("""proj_out.""", """proj_attn."""), ] def lowerCamelCase__ ( __snake_case ) -> List[str]: """simple docstring""" return w.reshape(*w.shape, 1, 1 ) def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" _UpperCamelCase = {k: k for k in vae_state_dict.keys()} for k, v in mapping.items(): for sd_part, hf_part in vae_conversion_map: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v for k, v in mapping.items(): if "attentions" in k: for sd_part, hf_part in vae_conversion_map_attn: _UpperCamelCase = v.replace(__snake_case, __snake_case ) _UpperCamelCase = v _UpperCamelCase = {v: vae_state_dict[k] for k, v in mapping.items()} _UpperCamelCase = ['''q''', '''k''', '''v''', '''proj_out'''] for k, v in new_state_dict.items(): for weight_name in weights_to_convert: if F'''mid.attn_1.{weight_name}.weight''' in k: print(F'''Reshaping {k} for SD format''' ) _UpperCamelCase = reshape_weight_for_sd(__snake_case ) return new_state_dict # =========================# # Text Encoder Conversion # # =========================# _a = [ # (stable-diffusion, HF Diffusers) ("""resblocks.""", """text_model.encoder.layers."""), ("""ln_1""", """layer_norm1"""), ("""ln_2""", """layer_norm2"""), (""".c_fc.""", """.fc1."""), (""".c_proj.""", """.fc2."""), (""".attn""", """.self_attn"""), ("""ln_final.""", """transformer.text_model.final_layer_norm."""), ("""token_embedding.weight""", """transformer.text_model.embeddings.token_embedding.weight"""), ("""positional_embedding""", """transformer.text_model.embeddings.position_embedding.weight"""), ] _a = {re.escape(x[1]): x[0] for x in textenc_conversion_lst} _a = re.compile("""|""".join(protected.keys())) # Ordering is from https://github.com/pytorch/pytorch/blob/master/test/cpp/api/modules.cpp _a = {"""q""": 0, """k""": 1, """v""": 2} def lowerCamelCase__ ( __snake_case ) -> Any: """simple docstring""" _UpperCamelCase = {} _UpperCamelCase = {} _UpperCamelCase = {} for k, v in text_enc_dict.items(): if ( k.endswith('''.self_attn.q_proj.weight''' ) or k.endswith('''.self_attn.k_proj.weight''' ) or k.endswith('''.self_attn.v_proj.weight''' ) ): _UpperCamelCase = k[: -len('''.q_proj.weight''' )] _UpperCamelCase = k[-len('''q_proj.weight''' )] if k_pre not in capture_qkv_weight: _UpperCamelCase = [None, None, None] _UpperCamelCase = v continue if ( k.endswith('''.self_attn.q_proj.bias''' ) or k.endswith('''.self_attn.k_proj.bias''' ) or k.endswith('''.self_attn.v_proj.bias''' ) ): _UpperCamelCase = k[: -len('''.q_proj.bias''' )] _UpperCamelCase = k[-len('''q_proj.bias''' )] if k_pre not in capture_qkv_bias: _UpperCamelCase = [None, None, None] _UpperCamelCase = v continue _UpperCamelCase = textenc_pattern.sub(lambda __snake_case : protected[re.escape(m.group(0 ) )], __snake_case ) _UpperCamelCase = v for k_pre, tensors in capture_qkv_weight.items(): if None in tensors: raise Exception('''CORRUPTED MODEL: one of the q-k-v values for the text encoder was missing''' ) _UpperCamelCase = textenc_pattern.sub(lambda __snake_case : protected[re.escape(m.group(0 ) )], __snake_case ) _UpperCamelCase = torch.cat(__snake_case ) for k_pre, tensors in capture_qkv_bias.items(): if None in tensors: raise Exception('''CORRUPTED MODEL: one of the q-k-v values for the text encoder was missing''' ) _UpperCamelCase = textenc_pattern.sub(lambda __snake_case : protected[re.escape(m.group(0 ) )], __snake_case ) _UpperCamelCase = torch.cat(__snake_case ) return new_state_dict def lowerCamelCase__ ( __snake_case ) -> Tuple: """simple docstring""" return text_enc_dict if __name__ == "__main__": _a = argparse.ArgumentParser() parser.add_argument("""--model_path""", default=None, type=str, required=True, help="""Path to the model to convert.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, required=True, help="""Path to the output model.""") parser.add_argument("""--half""", action="""store_true""", help="""Save weights in half precision.""") parser.add_argument( """--use_safetensors""", action="""store_true""", help="""Save weights use safetensors, default is ckpt.""" ) _a = parser.parse_args() assert args.model_path is not None, "Must provide a model path!" assert args.checkpoint_path is not None, "Must provide a checkpoint path!" # Path for safetensors _a = osp.join(args.model_path, """unet""", """diffusion_pytorch_model.safetensors""") _a = osp.join(args.model_path, """vae""", """diffusion_pytorch_model.safetensors""") _a = osp.join(args.model_path, """text_encoder""", """model.safetensors""") # Load models from safetensors if it exists, if it doesn't pytorch if osp.exists(unet_path): _a = load_file(unet_path, device="""cpu""") else: _a = osp.join(args.model_path, """unet""", """diffusion_pytorch_model.bin""") _a = torch.load(unet_path, map_location="""cpu""") if osp.exists(vae_path): _a = load_file(vae_path, device="""cpu""") else: _a = osp.join(args.model_path, """vae""", """diffusion_pytorch_model.bin""") _a = torch.load(vae_path, map_location="""cpu""") if osp.exists(text_enc_path): _a = load_file(text_enc_path, device="""cpu""") else: _a = osp.join(args.model_path, """text_encoder""", """pytorch_model.bin""") _a = torch.load(text_enc_path, map_location="""cpu""") # Convert the UNet model _a = convert_unet_state_dict(unet_state_dict) _a = {"""model.diffusion_model.""" + k: v for k, v in unet_state_dict.items()} # Convert the VAE model _a = convert_vae_state_dict(vae_state_dict) _a = {"""first_stage_model.""" + k: v for k, v in vae_state_dict.items()} # Easiest way to identify v2.0 model seems to be that the text encoder (OpenCLIP) is deeper _a = """text_model.encoder.layers.22.layer_norm2.bias""" in text_enc_dict if is_vaa_model: # Need to add the tag 'transformer' in advance so we can knock it out from the final layer-norm _a = {"""transformer.""" + k: v for k, v in text_enc_dict.items()} _a = convert_text_enc_state_dict_vaa(text_enc_dict) _a = {"""cond_stage_model.model.""" + k: v for k, v in text_enc_dict.items()} else: _a = convert_text_enc_state_dict(text_enc_dict) _a = {"""cond_stage_model.transformer.""" + k: v for k, v in text_enc_dict.items()} # Put together new checkpoint _a = {**unet_state_dict, **vae_state_dict, **text_enc_dict} if args.half: _a = {k: v.half() for k, v in state_dict.items()} if args.use_safetensors: save_file(state_dict, args.checkpoint_path) else: _a = {"""state_dict""": state_dict} torch.save(state_dict, args.checkpoint_path)
78
1
"""simple docstring""" def lowerCamelCase__ ( __snake_case = 1_00_00_00 ) -> int: """simple docstring""" _UpperCamelCase = limit + 1 _UpperCamelCase = [0] * limit for first_term in range(1, __snake_case ): for n in range(__snake_case, __snake_case, __snake_case ): _UpperCamelCase = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a _UpperCamelCase = sum(1 for x in frequency[1:limit] if x == 10 ) return count if __name__ == "__main__": print(F"""{solution() = }""")
78
"""simple docstring""" import argparse import torch from transformers import OpenAIGPTConfig, OpenAIGPTModel, load_tf_weights_in_openai_gpt from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> Union[str, Any]: """simple docstring""" if openai_config_file == "": _UpperCamelCase = OpenAIGPTConfig() else: _UpperCamelCase = OpenAIGPTConfig.from_json_file(__snake_case ) _UpperCamelCase = OpenAIGPTModel(__snake_case ) # Load weights from numpy load_tf_weights_in_openai_gpt(__snake_case, __snake_case, __snake_case ) # Save pytorch-model _UpperCamelCase = pytorch_dump_folder_path + '''/''' + WEIGHTS_NAME _UpperCamelCase = pytorch_dump_folder_path + '''/''' + CONFIG_NAME print(F'''Save PyTorch model to {pytorch_weights_dump_path}''' ) torch.save(model.state_dict(), __snake_case ) print(F'''Save configuration file to {pytorch_config_dump_path}''' ) with open(__snake_case, '''w''', encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( """--openai_checkpoint_folder_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--openai_config_file""", default="""""", type=str, help=( """An optional config json file corresponding to the pre-trained OpenAI model. \n""" """This specifies the model architecture.""" ), ) _a = parser.parse_args() convert_openai_checkpoint_to_pytorch( args.openai_checkpoint_folder_path, args.openai_config_file, args.pytorch_dump_folder_path )
78
1