code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case : Tuple = logging.get_logger(__name__) _snake_case : Optional[int] = { "RUCAIBox/mvp": "https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json", } class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Optional[int] = "mvp" __UpperCAmelCase : Optional[int] = ["past_key_values"] __UpperCAmelCase : Optional[Any] = {"num_attention_heads": "encoder_attention_heads", "hidden_size": "d_model"} def __init__( self : Union[str, Any] , lowerCamelCase : str=50267 , lowerCamelCase : str=1024 , lowerCamelCase : Any=12 , lowerCamelCase : List[str]=4096 , lowerCamelCase : Dict=16 , lowerCamelCase : Optional[Any]=12 , lowerCamelCase : List[Any]=4096 , lowerCamelCase : Dict=16 , lowerCamelCase : Any=0.0 , lowerCamelCase : List[str]=0.0 , lowerCamelCase : Dict="gelu" , lowerCamelCase : Optional[Any]=1024 , lowerCamelCase : Optional[int]=0.1 , lowerCamelCase : List[Any]=0.0 , lowerCamelCase : str=0.0 , lowerCamelCase : Any=0.02 , lowerCamelCase : Any=0.0 , lowerCamelCase : Dict=False , lowerCamelCase : Optional[Any]=True , lowerCamelCase : List[Any]=1 , lowerCamelCase : Tuple=0 , lowerCamelCase : Optional[int]=2 , lowerCamelCase : Union[str, Any]=True , lowerCamelCase : Any=2 , lowerCamelCase : Union[str, Any]=2 , lowerCamelCase : Tuple=False , lowerCamelCase : str=100 , lowerCamelCase : int=800 , **lowerCamelCase : Optional[int] , ) -> Any: __snake_case : Tuple = vocab_size __snake_case : Union[str, Any] = max_position_embeddings __snake_case : Union[str, Any] = d_model __snake_case : List[str] = encoder_ffn_dim __snake_case : Tuple = encoder_layers __snake_case : Dict = encoder_attention_heads __snake_case : str = decoder_ffn_dim __snake_case : Optional[int] = decoder_layers __snake_case : Optional[int] = decoder_attention_heads __snake_case : List[str] = dropout __snake_case : Dict = attention_dropout __snake_case : int = activation_dropout __snake_case : Optional[int] = activation_function __snake_case : Any = init_std __snake_case : Any = encoder_layerdrop __snake_case : Optional[int] = decoder_layerdrop __snake_case : Dict = classifier_dropout __snake_case : Union[str, Any] = use_cache __snake_case : List[Any] = encoder_layers __snake_case : List[Any] = scale_embedding # scale factor will be sqrt(d_model) if True __snake_case : List[str] = use_prompt __snake_case : Union[str, Any] = prompt_length __snake_case : int = prompt_mid_dim super().__init__( pad_token_id=lowerCamelCase , bos_token_id=lowerCamelCase , eos_token_id=lowerCamelCase , is_encoder_decoder=lowerCamelCase , decoder_start_token_id=lowerCamelCase , forced_eos_token_id=lowerCamelCase , **lowerCamelCase , ) if self.forced_bos_token_id is None and kwargs.get("force_bos_token_to_be_generated" , lowerCamelCase ): __snake_case : Any = self.bos_token_id warnings.warn( F'Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. ' "The config can simply be saved and uploaded again to be fixed." )
81
from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class A: '''simple docstring''' def __init__( self : List[Any] , A_ : Dict , A_ : Union[str, Any]=13 , A_ : List[Any]=30 , A_ : Optional[Any]=2 , A_ : List[str]=3 , A_ : List[str]=True , A_ : Dict=True , A_ : List[Any]=32 , A_ : Any=2 , A_ : Any=4 , A_ : Optional[int]=37 , A_ : Dict="gelu" , A_ : List[Any]=0.1 , A_ : Optional[int]=0.1 , A_ : Union[str, Any]=10 , A_ : Optional[Any]=0.02 , A_ : List[Any]=3 , A_ : str=None , ) -> str: """simple docstring""" lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = image_size lowerCamelCase_ = patch_size lowerCamelCase_ = num_channels lowerCamelCase_ = is_training lowerCamelCase_ = use_labels lowerCamelCase_ = hidden_size lowerCamelCase_ = num_hidden_layers lowerCamelCase_ = num_attention_heads lowerCamelCase_ = intermediate_size lowerCamelCase_ = hidden_act lowerCamelCase_ = hidden_dropout_prob lowerCamelCase_ = attention_probs_dropout_prob lowerCamelCase_ = type_sequence_label_size lowerCamelCase_ = initializer_range lowerCamelCase_ = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) lowerCamelCase_ = (image_size // patch_size) ** 2 lowerCamelCase_ = num_patches + 1 def a__ ( self : List[str] ) -> Dict: """simple docstring""" lowerCamelCase_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase_ = None if self.use_labels: lowerCamelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase_ = self.get_config() return config, pixel_values, labels def a__ ( self : List[Any] ) -> Any: """simple docstring""" return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=A_ , initializer_range=self.initializer_range , ) def a__ ( self : Any , A_ : int , A_ : int , A_ : int ) -> List[Any]: """simple docstring""" lowerCamelCase_ = TFViTModel(config=A_ ) lowerCamelCase_ = model(A_ , training=A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # Test with an image with different size than the one specified in config. lowerCamelCase_ = self.image_size // 2 lowerCamelCase_ = pixel_values[:, :, :image_size, :image_size] lowerCamelCase_ = model(A_ , interpolate_pos_encoding=A_ , training=A_ ) lowerCamelCase_ = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, seq_length, self.hidden_size) ) def a__ ( self : List[Any] , A_ : List[Any] , A_ : Any , A_ : Any ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.type_sequence_label_size lowerCamelCase_ = TFViTForImageClassification(A_ ) lowerCamelCase_ = model(A_ , labels=A_ , training=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # Test with an image with different size than the one specified in config. lowerCamelCase_ = self.image_size // 2 lowerCamelCase_ = pixel_values[:, :, :image_size, :image_size] lowerCamelCase_ = model(A_ , interpolate_pos_encoding=A_ , training=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowerCamelCase_ = 1 lowerCamelCase_ = TFViTForImageClassification(A_ ) lowerCamelCase_ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCamelCase_ = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def a__ ( self : List[Any] ) -> Any: """simple docstring""" lowerCamelCase_ = self.prepare_config_and_inputs() lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = config_and_inputs lowerCamelCase_ = {'pixel_values': pixel_values} return config, inputs_dict @require_tf class A( UpperCamelCase , UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () UpperCamelCase = ( {'''feature-extraction''': TFViTModel, '''image-classification''': TFViTForImageClassification} if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def a__ ( self : int ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = TFViTModelTester(self ) lowerCamelCase_ = ConfigTester(self , config_class=A_ , has_text_modality=A_ , hidden_size=37 ) def a__ ( self : int ) -> List[str]: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='ViT does not use inputs_embeds' ) def a__ ( self : List[str] ) -> Tuple: """simple docstring""" pass @unittest.skip(reason='ViT does not use inputs_embeds' ) def a__ ( self : int ) -> Optional[Any]: """simple docstring""" pass def a__ ( self : str ) -> List[str]: """simple docstring""" lowerCamelCase_ , lowerCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase_ = model_class(A_ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) lowerCamelCase_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(A_ , tf.keras.layers.Layer ) ) def a__ ( self : List[Any] ) -> Dict: """simple docstring""" lowerCamelCase_ , lowerCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase_ = model_class(A_ ) lowerCamelCase_ = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase_ = [*signature.parameters.keys()] lowerCamelCase_ = ['pixel_values'] self.assertListEqual(arg_names[:1] , A_ ) def a__ ( self : Tuple ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def a__ ( self : Tuple ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A_ ) @slow def a__ ( self : Any ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = TFViTModel.from_pretrained('google/vit-base-patch16-224' ) self.assertIsNotNone(A_ ) def _SCREAMING_SNAKE_CASE ( ): '''simple docstring''' lowerCamelCase_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class A( unittest.TestCase ): '''simple docstring''' @cached_property def a__ ( self : int ) -> Tuple: """simple docstring""" return ViTImageProcessor.from_pretrained('google/vit-base-patch16-224' ) if is_vision_available() else None @slow def a__ ( self : List[str] ) -> str: """simple docstring""" lowerCamelCase_ = TFViTForImageClassification.from_pretrained('google/vit-base-patch16-224' ) lowerCamelCase_ = self.default_image_processor lowerCamelCase_ = prepare_img() lowerCamelCase_ = image_processor(images=A_ , return_tensors='tf' ) # forward pass lowerCamelCase_ = model(**A_ ) # verify the logits lowerCamelCase_ = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , A_ ) lowerCamelCase_ = tf.constant([-0.2744, 0.8215, -0.0836] ) tf.debugging.assert_near(outputs.logits[0, :3] , A_ , atol=1E-4 )
70
0
"""simple docstring""" import baseaa def a__ ( lowerCAmelCase__ ): return baseaa.baaencode(string.encode("utf-8" ) ) def a__ ( lowerCAmelCase__ ): return baseaa.baadecode(lowerCAmelCase__ ).decode("utf-8" ) if __name__ == "__main__": lowerCamelCase = """Hello World!""" lowerCamelCase = baseaa_encode(test) print(encoded) lowerCamelCase = baseaa_decode(encoded) print(decoded)
82
import itertools import random import unittest import numpy as np from transformers import is_speech_available from transformers.testing_utils import require_torch, require_torchaudio from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import SpeechaTextFeatureExtractor lowerCamelCase : Any = random.Random() def _SCREAMING_SNAKE_CASE ( lowercase : Union[str, Any] , lowercase : int=1.0 , lowercase : List[str]=None , lowercase : str=None ): '''simple docstring''' if rng is None: lowerCamelCase_ = global_rng lowerCamelCase_ = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class A( unittest.TestCase ): '''simple docstring''' def __init__( self : List[Any] , A_ : Dict , A_ : int=7 , A_ : str=400 , A_ : Dict=2000 , A_ : List[Any]=24 , A_ : List[Any]=24 , A_ : int=0.0 , A_ : Dict=16000 , A_ : List[Any]=True , A_ : str=True , ) -> Dict: """simple docstring""" lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = min_seq_length lowerCamelCase_ = max_seq_length lowerCamelCase_ = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) lowerCamelCase_ = feature_size lowerCamelCase_ = num_mel_bins lowerCamelCase_ = padding_value lowerCamelCase_ = sampling_rate lowerCamelCase_ = return_attention_mask lowerCamelCase_ = do_normalize def a__ ( self : List[str] ) -> Optional[int]: """simple docstring""" return { "feature_size": self.feature_size, "num_mel_bins": self.num_mel_bins, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def a__ ( self : List[Any] , A_ : str=False , A_ : Union[str, Any]=False ) -> str: """simple docstring""" def _flatten(A_ : List[Any] ): return list(itertools.chain(*A_ ) ) if equal_length: lowerCamelCase_ = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size lowerCamelCase_ = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: lowerCamelCase_ = [np.asarray(A_ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class A( UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = SpeechaTextFeatureExtractor if is_speech_available() else None def a__ ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = SpeechaTextFeatureExtractionTester(self ) def a__ ( self : str , A_ : Dict ) -> Dict: """simple docstring""" self.assertTrue(np.all(np.mean(A_ , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(A_ , axis=0 ) - 1 ) < 1E-3 ) ) def a__ ( self : int ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = [np.asarray(A_ ) for speech_input in speech_inputs] # Test feature size lowerCamelCase_ = feature_extractor(A_ , padding=A_ , return_tensors='np' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.feature_size ) # Test not batched input lowerCamelCase_ = feature_extractor(speech_inputs[0] , return_tensors='np' ).input_features lowerCamelCase_ = feature_extractor(np_speech_inputs[0] , return_tensors='np' ).input_features self.assertTrue(np.allclose(A_ , A_ , atol=1E-3 ) ) # Test batched lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. lowerCamelCase_ = [floats_list((1, x) )[0] for x in (800, 800, 800)] lowerCamelCase_ = np.asarray(A_ ) lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1E-3 ) ) def a__ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = ['longest', 'max_length', 'do_not_pad'] lowerCamelCase_ = [None, 16, None] for max_length, padding in zip(A_ , A_ ): lowerCamelCase_ = feature_extractor( A_ , padding=A_ , max_length=A_ , return_attention_mask=A_ ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = [np.sum(A_ ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def a__ ( self : List[Any] ) -> Any: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = ['longest', 'max_length', 'do_not_pad'] lowerCamelCase_ = [None, 16, None] for max_length, padding in zip(A_ , A_ ): lowerCamelCase_ = feature_extractor( A_ , max_length=A_ , padding=A_ , return_tensors='np' , return_attention_mask=A_ ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = [np.sum(A_ ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self.assertTrue(input_features[0][fbank_feat_lengths[0] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self.assertTrue(input_features[0][fbank_feat_lengths[1] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def a__ ( self : List[str] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = feature_extractor( A_ , padding='max_length' , max_length=4 , truncation=A_ , return_tensors='np' , return_attention_mask=A_ , ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1] ) self._check_zero_mean_unit_variance(input_features[2] ) def a__ ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = feature_extractor( A_ , padding='longest' , max_length=4 , truncation=A_ , return_tensors='np' , return_attention_mask=A_ , ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 4, 24) ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = feature_extractor( A_ , padding='longest' , max_length=16 , truncation=A_ , return_tensors='np' , return_attention_mask=A_ , ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 6, 24) ) def a__ ( self : List[Any] ) -> List[str]: """simple docstring""" import torch lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = np.random.rand(100 , 32 ).astype(np.floataa ) lowerCamelCase_ = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: lowerCamelCase_ = feature_extractor.pad([{'input_features': inputs}] , return_tensors='np' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) lowerCamelCase_ = feature_extractor.pad([{'input_features': inputs}] , return_tensors='pt' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def a__ ( self : List[str] , A_ : Union[str, Any] ) -> List[Any]: """simple docstring""" from datasets import load_dataset lowerCamelCase_ = load_dataset('hf-internal-testing/librispeech_asr_dummy' , 'clean' , split='validation' ) # automatic decoding with librispeech lowerCamelCase_ = ds.sort('id' ).select(range(A_ ) )[:num_samples]['audio'] return [x["array"] for x in speech_samples] def a__ ( self : str ) -> Tuple: """simple docstring""" lowerCamelCase_ = np.array([ -1.5745, -1.7713, -1.7020, -1.6069, -1.2250, -1.1105, -0.9072, -0.8241, -1.2310, -0.8098, -0.3320, -0.4101, -0.7985, -0.4996, -0.8213, -0.9128, -1.0420, -1.1286, -1.0440, -0.7999, -0.8405, -1.2275, -1.5443, -1.4625, ] ) # fmt: on lowerCamelCase_ = self._load_datasamples(1 ) lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = feature_extractor(A_ , return_tensors='pt' ).input_features self.assertEquals(input_features.shape , (1, 584, 24) ) self.assertTrue(np.allclose(input_features[0, 0, :30] , A_ , atol=1E-4 ) )
70
0
"""simple docstring""" def snake_case_ ( A_ : list, A_ : list, A_ : int, A_ : int, A_ : int ): '''simple docstring''' if index == number_of_items: return 0 _lowerCamelCase : int = 0 _lowerCamelCase : str = 0 _lowerCamelCase : Dict = knapsack(A_, A_, A_, A_, index + 1 ) if weights[index] <= max_weight: _lowerCamelCase : Tuple = values[index] + knapsack( A_, A_, A_, max_weight - weights[index], index + 1 ) return max(A_, A_ ) if __name__ == "__main__": import doctest doctest.testmod()
83
import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class A( UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = TransfoXLTokenizer UpperCamelCase = False UpperCamelCase = False def a__ ( self : Optional[Any] ) -> int: """simple docstring""" super().setUp() lowerCamelCase_ = [ '<unk>', '[CLS]', '[SEP]', 'want', 'unwanted', 'wa', 'un', 'running', ',', 'low', 'l', ] lowerCamelCase_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def a__ ( self : Optional[Any] , **A_ : Tuple ) -> Any: """simple docstring""" lowerCamelCase_ = True return TransfoXLTokenizer.from_pretrained(self.tmpdirname , **A_ ) def a__ ( self : List[str] , A_ : Dict ) -> Any: """simple docstring""" lowerCamelCase_ = '<unk> UNwanted , running' lowerCamelCase_ = '<unk> unwanted, running' return input_text, output_text def a__ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = TransfoXLTokenizer(vocab_file=self.vocab_file , lower_case=A_ ) lowerCamelCase_ = tokenizer.tokenize('<unk> UNwanted , running' ) self.assertListEqual(A_ , ['<unk>', 'unwanted', ',', 'running'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A_ ) , [0, 4, 8, 7] ) def a__ ( self : Any ) -> str: """simple docstring""" lowerCamelCase_ = TransfoXLTokenizer(lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['hello', '!', 'how', 'are', 'you', '?'] ) def a__ ( self : int ) -> Dict: """simple docstring""" lowerCamelCase_ = TransfoXLTokenizer(lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?'] ) def a__ ( self : List[Any] ) -> Tuple: """simple docstring""" lowerCamelCase_ = TransfoXLTokenizer(lower_case=A_ ) lowerCamelCase_ = 'Hello (bracket) and side-scrolled [and] Henry\'s $5,000 with 3.34 m. What\'s up!?' lowerCamelCase_ = [ 'Hello', '(', 'bracket', ')', 'and', 'side', '@-@', 'scrolled', '[', 'and', ']', 'Henry', '\'s', '$', '5', '@,@', '000', 'with', '3', '@.@', '34', 'm', '.', 'What', '\'s', 'up', '!', '?', ] self.assertListEqual(tokenizer.tokenize(A_ ) , A_ ) self.assertEqual(tokenizer.convert_tokens_to_string(A_ ) , A_ ) def a__ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = len(A_ ) tokenizer.add_tokens(['new1', 'new2'] ) tokenizer.move_added_token('new1' , 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(A_ ) , original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode('new1' ) , [1] ) self.assertEqual(tokenizer.decode([1] ) , 'new1' )
70
0
import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class A_ : '''simple docstring''' def __init__( self , snake_case , snake_case=99 , snake_case=13 , snake_case=7 , snake_case=9 , snake_case=True , snake_case=True , snake_case=False , snake_case=32 , snake_case=5 , snake_case=4 , snake_case=37 , snake_case=8 , snake_case=0.1 , snake_case=0.002 , snake_case=1 , snake_case=0 , snake_case=0 , snake_case=None , snake_case=None , ): lowercase = parent lowercase = batch_size lowercase = encoder_seq_length lowercase = decoder_seq_length # For common tests lowercase = self.decoder_seq_length lowercase = is_training lowercase = use_attention_mask lowercase = use_labels lowercase = vocab_size lowercase = hidden_size lowercase = num_hidden_layers lowercase = num_attention_heads lowercase = d_ff lowercase = relative_attention_num_buckets lowercase = dropout_rate lowercase = initializer_factor lowercase = eos_token_id lowercase = pad_token_id lowercase = decoder_start_token_id lowercase = None lowercase = decoder_layers def SCREAMING_SNAKE_CASE__ ( self ): return TaConfig.from_pretrained('google/umt5-base' ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case=None , snake_case=None , snake_case=None , snake_case=None , snake_case=None , ): if attention_mask is None: lowercase = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: lowercase = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: lowercase = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=snake_case ) if decoder_head_mask is None: lowercase = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=snake_case ) if cross_attn_head_mask is None: lowercase = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=snake_case ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def SCREAMING_SNAKE_CASE__ ( self ): lowercase = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) lowercase = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input lowercase = input_ids.clamp(self.pad_token_id + 1 ) lowercase = decoder_input_ids.clamp(self.pad_token_id + 1 ) lowercase = self.get_config() lowercase = config.num_attention_heads lowercase = self.prepare_inputs_dict(snake_case , snake_case , snake_case ) return config, input_dict def SCREAMING_SNAKE_CASE__ ( self ): lowercase , lowercase = self.prepare_config_and_inputs() return config, inputs_dict def SCREAMING_SNAKE_CASE__ ( self ): return TaConfig( vocab_size=166 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def SCREAMING_SNAKE_CASE__ ( self ): return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , ): lowercase = UMTaModel(config=snake_case ) model.to(snake_case ) model.eval() lowercase = model( input_ids=snake_case , decoder_input_ids=snake_case , attention_mask=snake_case , decoder_attention_mask=snake_case , ) lowercase = model(input_ids=snake_case , decoder_input_ids=snake_case ) lowercase = result.last_hidden_state lowercase = result.past_key_values lowercase = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(snake_case ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , ): lowercase = UMTaModel(config=snake_case ).get_decoder().to(snake_case ).eval() # first forward pass lowercase = model(snake_case , use_cache=snake_case ) lowercase = model(snake_case ) lowercase = model(snake_case , use_cache=snake_case ) self.parent.assertTrue(len(snake_case ) == len(snake_case ) ) self.parent.assertTrue(len(snake_case ) == len(snake_case ) + 1 ) lowercase , lowercase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids lowercase = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and lowercase = torch.cat([input_ids, next_tokens] , dim=-1 ) lowercase = model(snake_case )['last_hidden_state'] lowercase = model(snake_case , past_key_values=snake_case )['last_hidden_state'] # select random slice lowercase = ids_tensor((1,) , output_from_past.shape[-1] ).item() lowercase = output_from_no_past[:, -1, random_slice_idx].detach() lowercase = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case , snake_case , atol=1E-3 ) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , ): lowercase = UMTaModel(config=snake_case ).to(snake_case ).half().eval() lowercase = model(**snake_case )['last_hidden_state'] self.parent.assertFalse(torch.isnan(snake_case ).any().item() ) @require_torch class A_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): '''simple docstring''' _UpperCamelCase : Any = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) _UpperCamelCase : int = (UMTaForConditionalGeneration,) if is_torch_available() else () _UpperCamelCase : Tuple = ( { """conversational""": UMTaForConditionalGeneration, """feature-extraction""": UMTaModel, """summarization""": UMTaForConditionalGeneration, """text2text-generation""": UMTaForConditionalGeneration, """translation""": UMTaForConditionalGeneration, """question-answering""": UMTaForQuestionAnswering, } if is_torch_available() else {} ) _UpperCamelCase : List[Any] = True _UpperCamelCase : Optional[int] = False _UpperCamelCase : Union[str, Any] = False _UpperCamelCase : Tuple = True _UpperCamelCase : Optional[Any] = True # The small UMT5 model needs higher percentages for CPU/MP tests _UpperCamelCase : str = [0.8, 0.9] def SCREAMING_SNAKE_CASE__ ( self ): lowercase = UMTaModelTester(self ) @unittest.skip('Test has a segmentation fault on torch 1.8.0' ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() lowercase = UMTaModel(config_and_inputs[0] ).to(snake_case ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( snake_case , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , F'''{tmpdirname}/t5_test.onnx''' , export_params=snake_case , opset_version=9 , input_names=['input_ids', 'decoder_input_ids'] , ) @unittest.skipIf(torch_device == 'cpu' , 'Cant do half precision' ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = ['encoder_attentions', 'decoder_attentions', 'cross_attentions'] lowercase = self.model_tester.prepare_config_and_inputs() lowercase = config_and_inputs[0] lowercase = UMTaForConditionalGeneration(snake_case ).eval() model.to(snake_case ) lowercase = { 'head_mask': torch.zeros(config.num_layers , config.num_heads , device=snake_case ), 'decoder_head_mask': torch.zeros(config.num_decoder_layers , config.num_heads , device=snake_case ), 'cross_attn_head_mask': torch.zeros(config.num_decoder_layers , config.num_heads , device=snake_case ), } for attn_name, (name, mask) in zip(snake_case , head_masking.items() ): lowercase = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": lowercase = torch.ones( config.num_decoder_layers , config.num_heads , device=snake_case ) lowercase = model.generate( config_and_inputs[1]['input_ids'] , num_beams=1 , max_length=3 , output_attentions=snake_case , return_dict_in_generate=snake_case , **snake_case , ) # We check the state of decoder_attentions and cross_attentions just from the last step lowercase = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip('Does not work on the tiny model as we keep hitting edge cases.' ) def SCREAMING_SNAKE_CASE__ ( self ): pass @require_torch @require_sentencepiece @require_tokenizers class A_ ( unittest.TestCase ): '''simple docstring''' @slow @unittest.skip( 'Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged' ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = UMTaForConditionalGeneration.from_pretrained('google/umt5-small' , return_dict=snake_case ).to(snake_case ) lowercase = AutoTokenizer.from_pretrained('google/umt5-small' , use_fast=snake_case , legacy=snake_case ) lowercase = [ 'Bonjour monsieur <extra_id_0> bien <extra_id_1>.', 'No se como puedo <extra_id_0>.', 'This is the reason why we <extra_id_0> them.', 'The <extra_id_0> walks in <extra_id_1>, seats', 'A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.', ] lowercase = tokenizer(snake_case , return_tensors='pt' , padding=snake_case ).input_ids # fmt: off lowercase = torch.tensor( [ [ 3_8530, 21_0703, 25_6299, 1410, 25_6298, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 826, 321, 671, 2_5922, 25_6299, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1460, 339, 312, 1_9014, 1_0620, 758, 25_6299, 2355,274, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 517, 25_6299, 1_4869, 281, 301, 25_6298, 275, 11_9983,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 320, 25_6299, 1_4869, 281, 2234, 289, 2275, 333,6_1391, 289, 25_6298, 543, 25_6297, 16_8714, 329, 25_6296,274, 1], ] ) # fmt: on torch.testing.assert_allclose(snake_case , snake_case ) lowercase = model.generate(input_ids.to(snake_case ) ) lowercase = [ '<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>', '<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', ] lowercase = tokenizer.batch_decode(snake_case ) self.assertEqual(snake_case , snake_case )
84
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetrImageProcessor class A( unittest.TestCase ): '''simple docstring''' def __init__( self : Optional[Any] , A_ : Dict , A_ : int=7 , A_ : Any=3 , A_ : List[str]=30 , A_ : Union[str, Any]=400 , A_ : List[str]=True , A_ : int=None , A_ : Any=True , A_ : str=1 / 255 , A_ : int=True , A_ : List[Any]=[0.5, 0.5, 0.5] , A_ : Union[str, Any]=[0.5, 0.5, 0.5] , A_ : Union[str, Any]=True , ) -> List[str]: """simple docstring""" lowerCamelCase_ = size if size is not None else {'shortest_edge': 18, 'longest_edge': 1333} lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = num_channels lowerCamelCase_ = min_resolution lowerCamelCase_ = max_resolution lowerCamelCase_ = do_resize lowerCamelCase_ = size lowerCamelCase_ = do_rescale lowerCamelCase_ = rescale_factor lowerCamelCase_ = do_normalize lowerCamelCase_ = image_mean lowerCamelCase_ = image_std lowerCamelCase_ = do_pad def a__ ( self : Tuple ) -> Dict: """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_pad": self.do_pad, } def a__ ( self : Union[str, Any] , A_ : Dict , A_ : Any=False ) -> Union[str, Any]: """simple docstring""" if not batched: lowerCamelCase_ = image_inputs[0] if isinstance(A_ , Image.Image ): lowerCamelCase_ , lowerCamelCase_ = image.size else: lowerCamelCase_ , lowerCamelCase_ = image.shape[1], image.shape[2] if w < h: lowerCamelCase_ = int(self.size['shortest_edge'] * h / w ) lowerCamelCase_ = self.size['shortest_edge'] elif w > h: lowerCamelCase_ = self.size['shortest_edge'] lowerCamelCase_ = int(self.size['shortest_edge'] * w / h ) else: lowerCamelCase_ = self.size['shortest_edge'] lowerCamelCase_ = self.size['shortest_edge'] else: lowerCamelCase_ = [] for image in image_inputs: lowerCamelCase_ , lowerCamelCase_ = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) lowerCamelCase_ = max(A_ , key=lambda A_ : item[0] )[0] lowerCamelCase_ = max(A_ , key=lambda A_ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class A( UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = DetrImageProcessor if is_vision_available() else None def a__ ( self : List[Any] ) -> str: """simple docstring""" lowerCamelCase_ = DetrImageProcessingTester(self ) @property def a__ ( self : List[str] ) -> List[Any]: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def a__ ( self : Dict ) -> int: """simple docstring""" lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , 'image_mean' ) ) self.assertTrue(hasattr(A_ , 'image_std' ) ) self.assertTrue(hasattr(A_ , 'do_normalize' ) ) self.assertTrue(hasattr(A_ , 'do_rescale' ) ) self.assertTrue(hasattr(A_ , 'rescale_factor' ) ) self.assertTrue(hasattr(A_ , 'do_resize' ) ) self.assertTrue(hasattr(A_ , 'size' ) ) self.assertTrue(hasattr(A_ , 'do_pad' ) ) def a__ ( self : List[str] ) -> str: """simple docstring""" lowerCamelCase_ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 18, 'longest_edge': 1333} ) self.assertEqual(image_processor.do_pad , A_ ) lowerCamelCase_ = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A_ ) self.assertEqual(image_processor.size , {'shortest_edge': 42, 'longest_edge': 84} ) self.assertEqual(image_processor.do_pad , A_ ) def a__ ( self : Dict ) -> Any: """simple docstring""" pass def a__ ( self : Tuple ) -> Tuple: """simple docstring""" lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCamelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ ) for image in image_inputs: self.assertIsInstance(A_ , Image.Image ) # Test not batched input lowerCamelCase_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) lowerCamelCase_ = image_processing(A_ , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def a__ ( self : List[str] ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCamelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , np.ndarray ) # Test not batched input lowerCamelCase_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase_ = image_processing(A_ , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def a__ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCamelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test not batched input lowerCamelCase_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase_ = image_processing(A_ , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def a__ ( self : Tuple ) -> List[Any]: """simple docstring""" lowerCamelCase_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_annotations.txt' , 'r' ) as f: lowerCamelCase_ = json.loads(f.read() ) lowerCamelCase_ = {'image_id': 39769, 'annotations': target} # encode them lowerCamelCase_ = DetrImageProcessor.from_pretrained('facebook/detr-resnet-50' ) lowerCamelCase_ = image_processing(images=A_ , annotations=A_ , return_tensors='pt' ) # verify pixel values lowerCamelCase_ = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , A_ ) lowerCamelCase_ = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area lowerCamelCase_ = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , A_ ) ) # verify boxes lowerCamelCase_ = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , A_ ) lowerCamelCase_ = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , A_ , atol=1E-3 ) ) # verify image_id lowerCamelCase_ = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , A_ ) ) # verify is_crowd lowerCamelCase_ = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , A_ ) ) # verify class_labels lowerCamelCase_ = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , A_ ) ) # verify orig_size lowerCamelCase_ = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , A_ ) ) # verify size lowerCamelCase_ = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , A_ ) ) @slow def a__ ( self : str ) -> Any: """simple docstring""" lowerCamelCase_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt' , 'r' ) as f: lowerCamelCase_ = json.loads(f.read() ) lowerCamelCase_ = {'file_name': '000000039769.png', 'image_id': 39769, 'segments_info': target} lowerCamelCase_ = pathlib.Path('./tests/fixtures/tests_samples/COCO/coco_panoptic' ) # encode them lowerCamelCase_ = DetrImageProcessor.from_pretrained('facebook/detr-resnet-50-panoptic' ) lowerCamelCase_ = image_processing(images=A_ , annotations=A_ , masks_path=A_ , return_tensors='pt' ) # verify pixel values lowerCamelCase_ = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , A_ ) lowerCamelCase_ = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area lowerCamelCase_ = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , A_ ) ) # verify boxes lowerCamelCase_ = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , A_ ) lowerCamelCase_ = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , A_ , atol=1E-3 ) ) # verify image_id lowerCamelCase_ = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , A_ ) ) # verify is_crowd lowerCamelCase_ = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , A_ ) ) # verify class_labels lowerCamelCase_ = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , A_ ) ) # verify masks lowerCamelCase_ = 822873 self.assertEqual(encoding['labels'][0]['masks'].sum().item() , A_ ) # verify orig_size lowerCamelCase_ = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , A_ ) ) # verify size lowerCamelCase_ = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , A_ ) )
70
0
from __future__ import annotations import pandas as pd def _a ( lowercase__ : list[int] , lowercase__ : list[int] , lowercase__ : int ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Any = [0] * no_of_processes SCREAMING_SNAKE_CASE__ : int = [0] * no_of_processes # Copy the burst time into remaining_time[] for i in range(lowercase__ ): SCREAMING_SNAKE_CASE__ : Optional[int] = burst_time[i] SCREAMING_SNAKE_CASE__ : Optional[int] = 0 SCREAMING_SNAKE_CASE__ : Optional[int] = 0 SCREAMING_SNAKE_CASE__ : Any = 9_99_99_99_99 SCREAMING_SNAKE_CASE__ : List[str] = 0 SCREAMING_SNAKE_CASE__ : Any = False # Process until all processes are completed while complete != no_of_processes: for j in range(lowercase__ ): if arrival_time[j] <= increment_time and remaining_time[j] > 0: if remaining_time[j] < minm: SCREAMING_SNAKE_CASE__ : Dict = remaining_time[j] SCREAMING_SNAKE_CASE__ : List[str] = j SCREAMING_SNAKE_CASE__ : List[str] = True if not check: increment_time += 1 continue remaining_time[short] -= 1 SCREAMING_SNAKE_CASE__ : Any = remaining_time[short] if minm == 0: SCREAMING_SNAKE_CASE__ : int = 9_99_99_99_99 if remaining_time[short] == 0: complete += 1 SCREAMING_SNAKE_CASE__ : str = False # Find finish time of current process SCREAMING_SNAKE_CASE__ : str = increment_time + 1 # Calculate waiting time SCREAMING_SNAKE_CASE__ : int = finish_time - arrival_time[short] SCREAMING_SNAKE_CASE__ : int = finar - burst_time[short] if waiting_time[short] < 0: SCREAMING_SNAKE_CASE__ : int = 0 # Increment time increment_time += 1 return waiting_time def _a ( lowercase__ : list[int] , lowercase__ : int , lowercase__ : list[int] ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[Any] = [0] * no_of_processes for i in range(lowercase__ ): SCREAMING_SNAKE_CASE__ : Optional[int] = burst_time[i] + waiting_time[i] return turn_around_time def _a ( lowercase__ : list[int] , lowercase__ : list[int] , lowercase__ : int ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[Any] = 0 SCREAMING_SNAKE_CASE__ : List[str] = 0 for i in range(lowercase__ ): SCREAMING_SNAKE_CASE__ : List[str] = total_waiting_time + waiting_time[i] SCREAMING_SNAKE_CASE__ : List[str] = total_turn_around_time + turn_around_time[i] print(f'''Average waiting time = {total_waiting_time / no_of_processes:.5f}''' ) print('Average turn around time =' , total_turn_around_time / no_of_processes ) if __name__ == "__main__": print("Enter how many process you want to analyze") SCREAMING_SNAKE_CASE__ : str = int(input()) SCREAMING_SNAKE_CASE__ : Optional[int] = [0] * no_of_processes SCREAMING_SNAKE_CASE__ : str = [0] * no_of_processes SCREAMING_SNAKE_CASE__ : int = list(range(1, no_of_processes + 1)) for i in range(no_of_processes): print("Enter the arrival time and burst time for process:--" + str(i + 1)) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : str = map(int, input().split()) SCREAMING_SNAKE_CASE__ : int = calculate_waitingtime(arrival_time, burst_time, no_of_processes) SCREAMING_SNAKE_CASE__ : Any = burst_time SCREAMING_SNAKE_CASE__ : Union[str, Any] = no_of_processes SCREAMING_SNAKE_CASE__ : str = waiting_time SCREAMING_SNAKE_CASE__ : int = calculate_turnaroundtime(bt, n, wt) calculate_average_times(waiting_time, turn_around_time, no_of_processes) SCREAMING_SNAKE_CASE__ : Any = pd.DataFrame( list(zip(processes, burst_time, arrival_time, waiting_time, turn_around_time)), columns=[ "Process", "BurstTime", "ArrivalTime", "WaitingTime", "TurnAroundTime", ], ) # Printing the dataFrame pd.set_option("display.max_rows", fcfs.shape[0] + 1) print(fcfs)
85
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase : Dict = logging.get_logger(__name__) lowerCamelCase : int = { "microsoft/swinv2-tiny-patch4-window8-256": ( "https://huggingface.co/microsoft/swinv2-tiny-patch4-window8-256/resolve/main/config.json" ), } class A( UpperCamelCase ): '''simple docstring''' UpperCamelCase = '''swinv2''' UpperCamelCase = { '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers''', } def __init__( self : Dict , A_ : List[Any]=224 , A_ : Optional[Any]=4 , A_ : int=3 , A_ : Dict=96 , A_ : Any=[2, 2, 6, 2] , A_ : Optional[Any]=[3, 6, 12, 24] , A_ : Tuple=7 , A_ : Tuple=4.0 , A_ : str=True , A_ : str=0.0 , A_ : Union[str, Any]=0.0 , A_ : Optional[Any]=0.1 , A_ : str="gelu" , A_ : int=False , A_ : str=0.02 , A_ : List[Any]=1E-5 , A_ : Any=32 , **A_ : Tuple , ) -> Any: """simple docstring""" super().__init__(**A_ ) lowerCamelCase_ = image_size lowerCamelCase_ = patch_size lowerCamelCase_ = num_channels lowerCamelCase_ = embed_dim lowerCamelCase_ = depths lowerCamelCase_ = len(A_ ) lowerCamelCase_ = num_heads lowerCamelCase_ = window_size lowerCamelCase_ = mlp_ratio lowerCamelCase_ = qkv_bias lowerCamelCase_ = hidden_dropout_prob lowerCamelCase_ = attention_probs_dropout_prob lowerCamelCase_ = drop_path_rate lowerCamelCase_ = hidden_act lowerCamelCase_ = use_absolute_embeddings lowerCamelCase_ = layer_norm_eps lowerCamelCase_ = initializer_range lowerCamelCase_ = encoder_stride # we set the hidden_size attribute in order to make Swinv2 work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowerCamelCase_ = int(embed_dim * 2 ** (len(A_ ) - 1) ) lowerCamelCase_ = (0, 0, 0, 0)
70
0
import torch from diffusers import DDPMParallelScheduler from .test_schedulers import SchedulerCommonTest class _a ( snake_case_ ): """simple docstring""" _lowerCamelCase : Optional[Any] = (DDPMParallelScheduler,) def __A ( self : List[Any] , **UpperCAmelCase : Optional[int] ): A_ = { "num_train_timesteps": 1000, "beta_start": 0.0_001, "beta_end": 0.02, "beta_schedule": "linear", "variance_type": "fixed_small", "clip_sample": True, } config.update(**UpperCAmelCase ) return config def __A ( self : Optional[Any] ): for timesteps in [1, 5, 100, 1000]: self.check_over_configs(num_train_timesteps=UpperCAmelCase ) def __A ( self : Dict ): for beta_start, beta_end in zip([0.0_001, 0.001, 0.01, 0.1] , [0.002, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=UpperCAmelCase , beta_end=UpperCAmelCase ) def __A ( self : int ): for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=UpperCAmelCase ) def __A ( self : Tuple ): for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=UpperCAmelCase ) def __A ( self : int ): for clip_sample in [True, False]: self.check_over_configs(clip_sample=UpperCAmelCase ) def __A ( self : Union[str, Any] ): self.check_over_configs(thresholding=UpperCAmelCase ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=UpperCAmelCase , prediction_type=UpperCAmelCase , sample_max_value=UpperCAmelCase , ) def __A ( self : Optional[int] ): for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=UpperCAmelCase ) def __A ( self : Tuple ): for t in [0, 500, 999]: self.check_over_forward(time_step=UpperCAmelCase ) def __A ( self : Tuple ): A_ = self.scheduler_classes[0] A_ = self.get_scheduler_config() A_ = scheduler_class(**UpperCAmelCase ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.00_979 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.02 ) ) < 1E-5 def __A ( self : List[Any] ): A_ = self.scheduler_classes[0] A_ = self.get_scheduler_config() A_ = scheduler_class(**UpperCAmelCase ) A_ = len(UpperCAmelCase ) A_ = self.dummy_model() A_ = self.dummy_sample_deter A_ = self.dummy_sample_deter + 0.1 A_ = self.dummy_sample_deter - 0.1 A_ = samplea.shape[0] A_ = torch.stack([samplea, samplea, samplea] , dim=0 ) A_ = torch.arange(UpperCAmelCase )[0:3, None].repeat(1 , UpperCAmelCase ) A_ = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) A_ = scheduler.batch_step_no_noise(UpperCAmelCase , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) ) A_ = torch.sum(torch.abs(UpperCAmelCase ) ) A_ = torch.mean(torch.abs(UpperCAmelCase ) ) assert abs(result_sum.item() - 1_153.1_833 ) < 1E-2 assert abs(result_mean.item() - 0.5_005 ) < 1E-3 def __A ( self : Tuple ): A_ = self.scheduler_classes[0] A_ = self.get_scheduler_config() A_ = scheduler_class(**UpperCAmelCase ) A_ = len(UpperCAmelCase ) A_ = self.dummy_model() A_ = self.dummy_sample_deter A_ = torch.manual_seed(0 ) for t in reversed(range(UpperCAmelCase ) ): # 1. predict noise residual A_ = model(UpperCAmelCase , UpperCAmelCase ) # 2. predict previous mean of sample x_t-1 A_ = scheduler.step(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , generator=UpperCAmelCase ).prev_sample A_ = pred_prev_sample A_ = torch.sum(torch.abs(UpperCAmelCase ) ) A_ = torch.mean(torch.abs(UpperCAmelCase ) ) assert abs(result_sum.item() - 258.9_606 ) < 1E-2 assert abs(result_mean.item() - 0.3_372 ) < 1E-3 def __A ( self : Tuple ): A_ = self.scheduler_classes[0] A_ = self.get_scheduler_config(prediction_type="v_prediction" ) A_ = scheduler_class(**UpperCAmelCase ) A_ = len(UpperCAmelCase ) A_ = self.dummy_model() A_ = self.dummy_sample_deter A_ = torch.manual_seed(0 ) for t in reversed(range(UpperCAmelCase ) ): # 1. predict noise residual A_ = model(UpperCAmelCase , UpperCAmelCase ) # 2. predict previous mean of sample x_t-1 A_ = scheduler.step(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , generator=UpperCAmelCase ).prev_sample A_ = pred_prev_sample A_ = torch.sum(torch.abs(UpperCAmelCase ) ) A_ = torch.mean(torch.abs(UpperCAmelCase ) ) assert abs(result_sum.item() - 202.0_296 ) < 1E-2 assert abs(result_mean.item() - 0.2_631 ) < 1E-3 def __A ( self : Union[str, Any] ): A_ = self.scheduler_classes[0] A_ = self.get_scheduler_config() A_ = scheduler_class(**UpperCAmelCase ) A_ = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=UpperCAmelCase ) A_ = scheduler.timesteps for i, timestep in enumerate(UpperCAmelCase ): if i == len(UpperCAmelCase ) - 1: A_ = -1 else: A_ = timesteps[i + 1] A_ = scheduler.previous_timestep(UpperCAmelCase ) A_ = prev_t.item() self.assertEqual(UpperCAmelCase , UpperCAmelCase ) def __A ( self : List[Any] ): A_ = self.scheduler_classes[0] A_ = self.get_scheduler_config() A_ = scheduler_class(**UpperCAmelCase ) A_ = [100, 87, 50, 51, 0] with self.assertRaises(UpperCAmelCase , msg="`custom_timesteps` must be in descending order." ): scheduler.set_timesteps(timesteps=UpperCAmelCase ) def __A ( self : List[Any] ): A_ = self.scheduler_classes[0] A_ = self.get_scheduler_config() A_ = scheduler_class(**UpperCAmelCase ) A_ = [100, 87, 50, 1, 0] A_ = len(UpperCAmelCase ) with self.assertRaises(UpperCAmelCase , msg="Can only pass one of `num_inference_steps` or `custom_timesteps`." ): scheduler.set_timesteps(num_inference_steps=UpperCAmelCase , timesteps=UpperCAmelCase ) def __A ( self : Optional[Any] ): A_ = self.scheduler_classes[0] A_ = self.get_scheduler_config() A_ = scheduler_class(**UpperCAmelCase ) A_ = [scheduler.config.num_train_timesteps] with self.assertRaises( UpperCAmelCase , msg="`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}" , ): scheduler.set_timesteps(timesteps=UpperCAmelCase )
86
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import ( AutoProcessor, BertTokenizerFast, BlipImageProcessor, GPTaTokenizer, InstructBlipProcessor, PreTrainedTokenizerFast, ) @require_vision class A( unittest.TestCase ): '''simple docstring''' def a__ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = tempfile.mkdtemp() lowerCamelCase_ = BlipImageProcessor() lowerCamelCase_ = GPTaTokenizer.from_pretrained('hf-internal-testing/tiny-random-GPT2Model' ) lowerCamelCase_ = BertTokenizerFast.from_pretrained('hf-internal-testing/tiny-random-bert' ) lowerCamelCase_ = InstructBlipProcessor(A_ , A_ , A_ ) processor.save_pretrained(self.tmpdirname ) def a__ ( self : Optional[int] , **A_ : Optional[int] ) -> Dict: """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **A_ ).tokenizer def a__ ( self : List[str] , **A_ : str ) -> Optional[Any]: """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **A_ ).image_processor def a__ ( self : Tuple , **A_ : Any ) -> Optional[int]: """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **A_ ).qformer_tokenizer def a__ ( self : str ) -> str: """simple docstring""" shutil.rmtree(self.tmpdirname ) def a__ ( self : Dict ) -> Dict: """simple docstring""" lowerCamelCase_ = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] lowerCamelCase_ = [Image.fromarray(np.moveaxis(A_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def a__ ( self : Tuple ) -> Tuple: """simple docstring""" lowerCamelCase_ = InstructBlipProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() , qformer_tokenizer=self.get_qformer_tokenizer() , ) processor.save_pretrained(self.tmpdirname ) lowerCamelCase_ = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) lowerCamelCase_ = self.get_image_processor(do_normalize=A_ , padding_value=1.0 ) lowerCamelCase_ = InstructBlipProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=A_ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , A_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , A_ ) self.assertIsInstance(processor.qformer_tokenizer , A_ ) def a__ ( self : List[str] ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = self.prepare_image_inputs() lowerCamelCase_ = image_processor(A_ , return_tensors='np' ) lowerCamelCase_ = processor(images=A_ , return_tensors='np' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def a__ ( self : Union[str, Any] ) -> Tuple: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = 'lower newer' lowerCamelCase_ = processor(text=A_ ) lowerCamelCase_ = tokenizer(A_ , return_token_type_ids=A_ ) lowerCamelCase_ = qformer_tokenizer(A_ , return_token_type_ids=A_ ) for key in encoded_tokens.keys(): self.assertListEqual(encoded_tokens[key] , encoded_processor[key] ) for key in encoded_tokens_qformer.keys(): self.assertListEqual(encoded_tokens_qformer[key] , encoded_processor['qformer_' + key] ) def a__ ( self : Tuple ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = 'lower newer' lowerCamelCase_ = self.prepare_image_inputs() lowerCamelCase_ = processor(text=A_ , images=A_ ) self.assertListEqual( list(inputs.keys() ) , ['input_ids', 'attention_mask', 'qformer_input_ids', 'qformer_attention_mask', 'pixel_values'] , ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def a__ ( self : Optional[int] ) -> int: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowerCamelCase_ = processor.batch_decode(A_ ) lowerCamelCase_ = tokenizer.batch_decode(A_ ) self.assertListEqual(A_ , A_ ) def a__ ( self : Tuple ) -> List[str]: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = 'lower newer' lowerCamelCase_ = self.prepare_image_inputs() lowerCamelCase_ = processor(text=A_ , images=A_ ) self.assertListEqual( list(inputs.keys() ) , ['input_ids', 'attention_mask', 'qformer_input_ids', 'qformer_attention_mask', 'pixel_values'] , )
70
0
from typing import TYPE_CHECKING from ...utils import _LazyModule _lowerCamelCase : Union[str, Any] = {"""tokenization_bertweet""": ["""BertweetTokenizer"""]} if TYPE_CHECKING: from .tokenization_bertweet import BertweetTokenizer else: import sys _lowerCamelCase : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
87
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_OBJECT_DETECTION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING lowerCamelCase : Tuple = logging.get_logger(__name__) lowerCamelCase : List[Any] = Dict[str, Any] lowerCamelCase : Dict = List[Prediction] @add_end_docstrings(UpperCamelCase ) class A( UpperCamelCase ): '''simple docstring''' def __init__( self : Tuple , *A_ : int , **A_ : int ) -> Optional[int]: """simple docstring""" super().__init__(*A_ , **A_ ) if self.framework == "tf": raise ValueError(f"""The {self.__class__} is only available in PyTorch.""" ) requires_backends(self , 'vision' ) self.check_model_type( dict(MODEL_FOR_OBJECT_DETECTION_MAPPING.items() + MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING.items() ) ) def a__ ( self : Union[str, Any] , **A_ : Union[str, Any] ) -> List[str]: """simple docstring""" lowerCamelCase_ = {} if "threshold" in kwargs: lowerCamelCase_ = kwargs['threshold'] return {}, {}, postprocess_kwargs def __call__( self : str , *A_ : Optional[int] , **A_ : Tuple ) -> Union[Predictions, List[Prediction]]: """simple docstring""" return super().__call__(*A_ , **A_ ) def a__ ( self : Union[str, Any] , A_ : Tuple ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = load_image(A_ ) lowerCamelCase_ = torch.IntTensor([[image.height, image.width]] ) lowerCamelCase_ = self.image_processor(images=[image] , return_tensors='pt' ) if self.tokenizer is not None: lowerCamelCase_ = self.tokenizer(text=inputs['words'] , boxes=inputs['boxes'] , return_tensors='pt' ) lowerCamelCase_ = target_size return inputs def a__ ( self : Union[str, Any] , A_ : List[str] ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = model_inputs.pop('target_size' ) lowerCamelCase_ = self.model(**A_ ) lowerCamelCase_ = outputs.__class__({'target_size': target_size, **outputs} ) if self.tokenizer is not None: lowerCamelCase_ = model_inputs['bbox'] return model_outputs def a__ ( self : str , A_ : Any , A_ : Tuple=0.9 ) -> str: """simple docstring""" lowerCamelCase_ = model_outputs['target_size'] if self.tokenizer is not None: # This is a LayoutLMForTokenClassification variant. # The OCR got the boxes and the model classified the words. lowerCamelCase_ , lowerCamelCase_ = target_size[0].tolist() def unnormalize(A_ : Dict ): return self._get_bounding_box( torch.Tensor( [ (width * bbox[0] / 1000), (height * bbox[1] / 1000), (width * bbox[2] / 1000), (height * bbox[3] / 1000), ] ) ) lowerCamelCase_ , lowerCamelCase_ = model_outputs['logits'].squeeze(0 ).softmax(dim=-1 ).max(dim=-1 ) lowerCamelCase_ = [self.model.config.idalabel[prediction] for prediction in classes.tolist()] lowerCamelCase_ = [unnormalize(A_ ) for bbox in model_outputs['bbox'].squeeze(0 )] lowerCamelCase_ = ['score', 'label', 'box'] lowerCamelCase_ = [dict(zip(A_ , A_ ) ) for vals in zip(scores.tolist() , A_ , A_ ) if vals[0] > threshold] else: # This is a regular ForObjectDetectionModel lowerCamelCase_ = self.image_processor.post_process_object_detection(A_ , A_ , A_ ) lowerCamelCase_ = raw_annotations[0] lowerCamelCase_ = raw_annotation['scores'] lowerCamelCase_ = raw_annotation['labels'] lowerCamelCase_ = raw_annotation['boxes'] lowerCamelCase_ = scores.tolist() lowerCamelCase_ = [self.model.config.idalabel[label.item()] for label in labels] lowerCamelCase_ = [self._get_bounding_box(A_ ) for box in boxes] # {"scores": [...], ...} --> [{"score":x, ...}, ...] lowerCamelCase_ = ['score', 'label', 'box'] lowerCamelCase_ = [ dict(zip(A_ , A_ ) ) for vals in zip(raw_annotation['scores'] , raw_annotation['labels'] , raw_annotation['boxes'] ) ] return annotation def a__ ( self : Union[str, Any] , A_ : "torch.Tensor" ) -> Dict[str, int]: """simple docstring""" if self.framework != "pt": raise ValueError('The ObjectDetectionPipeline is only available in PyTorch.' ) lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = box.int().tolist() lowerCamelCase_ = { 'xmin': xmin, 'ymin': ymin, 'xmax': xmax, 'ymax': ymax, } return bbox
70
0
"""simple docstring""" import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class lowercase__ ( A_ ): __UpperCAmelCase = ['''image_processor''', '''tokenizer'''] __UpperCAmelCase = '''BlipImageProcessor''' __UpperCAmelCase = '''AutoTokenizer''' def __init__( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) -> Union[str, Any]: super().__init__(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) # add QFormer tokenizer _lowerCamelCase : Dict = qformer_tokenizer def __call__( self , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = True , SCREAMING_SNAKE_CASE = False , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = 0 , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = False , SCREAMING_SNAKE_CASE = False , SCREAMING_SNAKE_CASE = False , SCREAMING_SNAKE_CASE = False , SCREAMING_SNAKE_CASE = False , SCREAMING_SNAKE_CASE = True , SCREAMING_SNAKE_CASE = None , **SCREAMING_SNAKE_CASE , ) -> BatchFeature: if images is None and text is None: raise ValueError("""You have to specify at least images or text.""") _lowerCamelCase : Tuple = BatchFeature() if text is not None: _lowerCamelCase : List[Any] = self.tokenizer( text=SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE , padding=SCREAMING_SNAKE_CASE , truncation=SCREAMING_SNAKE_CASE , max_length=SCREAMING_SNAKE_CASE , stride=SCREAMING_SNAKE_CASE , pad_to_multiple_of=SCREAMING_SNAKE_CASE , return_attention_mask=SCREAMING_SNAKE_CASE , return_overflowing_tokens=SCREAMING_SNAKE_CASE , return_special_tokens_mask=SCREAMING_SNAKE_CASE , return_offsets_mapping=SCREAMING_SNAKE_CASE , return_token_type_ids=SCREAMING_SNAKE_CASE , return_length=SCREAMING_SNAKE_CASE , verbose=SCREAMING_SNAKE_CASE , return_tensors=SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE , ) encoding.update(SCREAMING_SNAKE_CASE) _lowerCamelCase : Tuple = self.qformer_tokenizer( text=SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE , padding=SCREAMING_SNAKE_CASE , truncation=SCREAMING_SNAKE_CASE , max_length=SCREAMING_SNAKE_CASE , stride=SCREAMING_SNAKE_CASE , pad_to_multiple_of=SCREAMING_SNAKE_CASE , return_attention_mask=SCREAMING_SNAKE_CASE , return_overflowing_tokens=SCREAMING_SNAKE_CASE , return_special_tokens_mask=SCREAMING_SNAKE_CASE , return_offsets_mapping=SCREAMING_SNAKE_CASE , return_token_type_ids=SCREAMING_SNAKE_CASE , return_length=SCREAMING_SNAKE_CASE , verbose=SCREAMING_SNAKE_CASE , return_tensors=SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE , ) _lowerCamelCase : str = qformer_text_encoding.pop("""input_ids""") _lowerCamelCase : str = qformer_text_encoding.pop("""attention_mask""") if images is not None: _lowerCamelCase : Tuple = self.image_processor(SCREAMING_SNAKE_CASE , return_tensors=SCREAMING_SNAKE_CASE) encoding.update(SCREAMING_SNAKE_CASE) return encoding def UpperCamelCase_ ( self , *SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE) -> Optional[int]: return self.tokenizer.batch_decode(*SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE) def UpperCamelCase_ ( self , *SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE) -> List[str]: return self.tokenizer.decode(*SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def UpperCamelCase_ ( self) -> Any: _lowerCamelCase : List[Any] = self.tokenizer.model_input_names _lowerCamelCase : List[str] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names)) def UpperCamelCase_ ( self , SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE) -> str: if os.path.isfile(SCREAMING_SNAKE_CASE): raise ValueError(F'Provided path ({save_directory}) should be a directory, not a file') os.makedirs(SCREAMING_SNAKE_CASE , exist_ok=SCREAMING_SNAKE_CASE) _lowerCamelCase : List[str] = os.path.join(SCREAMING_SNAKE_CASE , """qformer_tokenizer""") self.qformer_tokenizer.save_pretrained(SCREAMING_SNAKE_CASE) return super().save_pretrained(SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE) @classmethod def UpperCamelCase_ ( cls , SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE) -> str: _lowerCamelCase : Optional[Any] = AutoTokenizer.from_pretrained(SCREAMING_SNAKE_CASE , subfolder="""qformer_tokenizer""") _lowerCamelCase : Optional[Any] = cls._get_arguments_from_pretrained(SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE) args.append(SCREAMING_SNAKE_CASE) return cls(*SCREAMING_SNAKE_CASE)
88
from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
70
0
import os import tempfile import unittest from transformers import NezhaConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, ) from transformers.models.nezha.modeling_nezha import NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST class _lowerCamelCase: def __init__( self, lowerCamelCase, lowerCamelCase=13, lowerCamelCase=7, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=99, lowerCamelCase=32, lowerCamelCase=5, lowerCamelCase=4, lowerCamelCase=37, lowerCamelCase="gelu", lowerCamelCase=0.1, lowerCamelCase=0.1, lowerCamelCase=1_28, lowerCamelCase=32, lowerCamelCase=16, lowerCamelCase=2, lowerCamelCase=0.0_2, lowerCamelCase=3, lowerCamelCase=4, lowerCamelCase=None, ) -> Union[str, Any]: """simple docstring""" _lowercase : List[str] = parent _lowercase : Tuple = batch_size _lowercase : List[str] = seq_length _lowercase : Optional[int] = is_training _lowercase : Dict = use_input_mask _lowercase : Optional[int] = use_token_type_ids _lowercase : List[str] = use_labels _lowercase : List[str] = vocab_size _lowercase : Optional[Any] = hidden_size _lowercase : Optional[Any] = num_hidden_layers _lowercase : Dict = num_attention_heads _lowercase : List[str] = intermediate_size _lowercase : Optional[Any] = hidden_act _lowercase : Optional[Any] = hidden_dropout_prob _lowercase : Tuple = attention_probs_dropout_prob _lowercase : Optional[int] = max_position_embeddings _lowercase : List[str] = type_vocab_size _lowercase : Optional[int] = type_sequence_label_size _lowercase : Any = initializer_range _lowercase : Optional[int] = num_labels _lowercase : List[str] = num_choices _lowercase : Tuple = scope def UpperCamelCase ( self) -> Any: """simple docstring""" _lowercase : Optional[Any] = ids_tensor([self.batch_size, self.seq_length], self.vocab_size) _lowercase : Dict = None if self.use_input_mask: _lowercase : Dict = random_attention_mask([self.batch_size, self.seq_length]) _lowercase : str = None if self.use_token_type_ids: _lowercase : List[str] = ids_tensor([self.batch_size, self.seq_length], self.type_vocab_size) _lowercase : Union[str, Any] = None _lowercase : List[Any] = None _lowercase : str = None if self.use_labels: _lowercase : int = ids_tensor([self.batch_size], self.type_sequence_label_size) _lowercase : Any = ids_tensor([self.batch_size, self.seq_length], self.num_labels) _lowercase : List[str] = ids_tensor([self.batch_size], self.num_choices) _lowercase : List[Any] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase ( self) -> List[Any]: """simple docstring""" return NezhaConfig( vocab_size=self.vocab_size, hidden_size=self.hidden_size, num_hidden_layers=self.num_hidden_layers, num_attention_heads=self.num_attention_heads, intermediate_size=self.intermediate_size, hidden_act=self.hidden_act, hidden_dropout_prob=self.hidden_dropout_prob, attention_probs_dropout_prob=self.attention_probs_dropout_prob, max_position_embeddings=self.max_position_embeddings, type_vocab_size=self.type_vocab_size, is_decoder=lowerCamelCase, initializer_range=self.initializer_range, ) def UpperCamelCase ( self) -> Dict: """simple docstring""" ( ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ) : Any = self.prepare_config_and_inputs() _lowercase : Dict = True _lowercase : List[str] = floats_tensor([self.batch_size, self.seq_length, self.hidden_size]) _lowercase : List[str] = ids_tensor([self.batch_size, self.seq_length], vocab_size=2) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase) -> List[str]: """simple docstring""" _lowercase : Any = NezhaModel(config=lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : List[Any] = model(lowerCamelCase, attention_mask=lowerCamelCase, token_type_ids=lowerCamelCase) _lowercase : Union[str, Any] = model(lowerCamelCase, token_type_ids=lowerCamelCase) _lowercase : Tuple = model(lowerCamelCase) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size)) self.parent.assertEqual(result.pooler_output.shape, (self.batch_size, self.hidden_size)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> Any: """simple docstring""" _lowercase : Dict = True _lowercase : Union[str, Any] = NezhaModel(lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : Any = model( lowerCamelCase, attention_mask=lowerCamelCase, token_type_ids=lowerCamelCase, encoder_hidden_states=lowerCamelCase, encoder_attention_mask=lowerCamelCase, ) _lowercase : Dict = model( lowerCamelCase, attention_mask=lowerCamelCase, token_type_ids=lowerCamelCase, encoder_hidden_states=lowerCamelCase, ) _lowercase : Dict = model(lowerCamelCase, attention_mask=lowerCamelCase, token_type_ids=lowerCamelCase) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size)) self.parent.assertEqual(result.pooler_output.shape, (self.batch_size, self.hidden_size)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase) -> Any: """simple docstring""" _lowercase : Optional[int] = NezhaForMaskedLM(config=lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : str = model(lowerCamelCase, attention_mask=lowerCamelCase, token_type_ids=lowerCamelCase, labels=lowerCamelCase) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.vocab_size)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase) -> Tuple: """simple docstring""" _lowercase : Dict = NezhaForNextSentencePrediction(config=lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : str = model( lowerCamelCase, attention_mask=lowerCamelCase, token_type_ids=lowerCamelCase, labels=lowerCamelCase, ) self.parent.assertEqual(result.logits.shape, (self.batch_size, 2)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase) -> Optional[int]: """simple docstring""" _lowercase : Any = NezhaForPreTraining(config=lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : Optional[Any] = model( lowerCamelCase, attention_mask=lowerCamelCase, token_type_ids=lowerCamelCase, labels=lowerCamelCase, next_sentence_label=lowerCamelCase, ) self.parent.assertEqual(result.prediction_logits.shape, (self.batch_size, self.seq_length, self.vocab_size)) self.parent.assertEqual(result.seq_relationship_logits.shape, (self.batch_size, 2)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase) -> Union[str, Any]: """simple docstring""" _lowercase : Optional[int] = NezhaForQuestionAnswering(config=lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : Union[str, Any] = model( lowerCamelCase, attention_mask=lowerCamelCase, token_type_ids=lowerCamelCase, start_positions=lowerCamelCase, end_positions=lowerCamelCase, ) self.parent.assertEqual(result.start_logits.shape, (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape, (self.batch_size, self.seq_length)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase) -> Optional[int]: """simple docstring""" _lowercase : int = self.num_labels _lowercase : Tuple = NezhaForSequenceClassification(lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : int = model(lowerCamelCase, attention_mask=lowerCamelCase, token_type_ids=lowerCamelCase, labels=lowerCamelCase) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_labels)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase) -> Dict: """simple docstring""" _lowercase : Optional[Any] = self.num_labels _lowercase : Optional[Any] = NezhaForTokenClassification(config=lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : Dict = model(lowerCamelCase, attention_mask=lowerCamelCase, token_type_ids=lowerCamelCase, labels=lowerCamelCase) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.num_labels)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase) -> Dict: """simple docstring""" _lowercase : Optional[Any] = self.num_choices _lowercase : Any = NezhaForMultipleChoice(config=lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : str = input_ids.unsqueeze(1).expand(-1, self.num_choices, -1).contiguous() _lowercase : Tuple = token_type_ids.unsqueeze(1).expand(-1, self.num_choices, -1).contiguous() _lowercase : str = input_mask.unsqueeze(1).expand(-1, self.num_choices, -1).contiguous() _lowercase : Tuple = model( lowerCamelCase, attention_mask=lowerCamelCase, token_type_ids=lowerCamelCase, labels=lowerCamelCase, ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_choices)) def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : Optional[int] = self.prepare_config_and_inputs() ( ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ) : Optional[Any] = config_and_inputs _lowercase : List[str] = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class _lowerCamelCase( _a, _a, _a, unittest.TestCase ): lowercase_ : Union[str, Any] = ( ( NezhaModel, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, ) if is_torch_available() else () ) lowercase_ : str = ( { """feature-extraction""": NezhaModel, """fill-mask""": NezhaForMaskedLM, """question-answering""": NezhaForQuestionAnswering, """text-classification""": NezhaForSequenceClassification, """token-classification""": NezhaForTokenClassification, """zero-shot""": NezhaForSequenceClassification, } if is_torch_available() else {} ) lowercase_ : Dict = True def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase=False) -> Dict: """simple docstring""" _lowercase : Dict = super()._prepare_for_class(lowerCamelCase, lowerCamelCase, return_labels=lowerCamelCase) if return_labels: if model_class in get_values(lowerCamelCase): _lowercase : str = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length), dtype=torch.long, device=lowerCamelCase) _lowercase : Optional[int] = torch.zeros( self.model_tester.batch_size, dtype=torch.long, device=lowerCamelCase) return inputs_dict def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" _lowercase : Any = NezhaModelTester(self) _lowercase : Any = ConfigTester(self, config_class=lowerCamelCase, hidden_size=37) def UpperCamelCase ( self) -> Any: """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase) def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*lowerCamelCase) def UpperCamelCase ( self) -> List[str]: """simple docstring""" ( ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ) : List[str] = self.model_tester.prepare_config_and_inputs_for_decoder() _lowercase : List[str] = None self.model_tester.create_and_check_model_as_decoder( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowerCamelCase) def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowerCamelCase) def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_next_sequence_prediction(*lowerCamelCase) def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowerCamelCase) def UpperCamelCase ( self) -> Any: """simple docstring""" _lowercase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowerCamelCase) def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowerCamelCase) def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" _lowercase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowerCamelCase) @slow def UpperCamelCase ( self) -> int: """simple docstring""" for model_name in NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowercase : int = NezhaModel.from_pretrained(lowerCamelCase) self.assertIsNotNone(lowerCamelCase) @slow @require_torch_gpu def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase , _lowercase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # NezhaForMultipleChoice behaves incorrectly in JIT environments. if model_class == NezhaForMultipleChoice: return _lowercase : Tuple = True _lowercase : List[Any] = model_class(config=lowerCamelCase) _lowercase : Dict = self._prepare_for_class(lowerCamelCase, lowerCamelCase) _lowercase : List[str] = torch.jit.trace( lowerCamelCase, (inputs_dict['input_ids'].to('cpu'), inputs_dict['attention_mask'].to('cpu'))) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(lowerCamelCase, os.path.join(lowerCamelCase, 'bert.pt')) _lowercase : int = torch.jit.load(os.path.join(lowerCamelCase, 'bert.pt'), map_location=lowerCamelCase) loaded(inputs_dict['input_ids'].to(lowerCamelCase), inputs_dict['attention_mask'].to(lowerCamelCase)) @require_torch class _lowerCamelCase( unittest.TestCase ): @slow def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : Union[str, Any] = NezhaModel.from_pretrained('sijunhe/nezha-cn-base') _lowercase : Tuple = torch.tensor([[0, 1, 2, 3, 4, 5]]) _lowercase : Optional[Any] = torch.tensor([[0, 1, 1, 1, 1, 1]]) with torch.no_grad(): _lowercase : int = model(lowerCamelCase, attention_mask=lowerCamelCase)[0] _lowercase : Any = torch.Size((1, 6, 7_68)) self.assertEqual(output.shape, lowerCamelCase) _lowercase : List[Any] = torch.tensor([[[0.0_6_8_5, 0.2_4_4_1, 0.1_1_0_2], [0.0_6_0_0, 0.1_9_0_6, 0.1_3_4_9], [0.0_2_2_1, 0.0_8_1_9, 0.0_5_8_6]]]) self.assertTrue(torch.allclose(output[:, 1:4, 1:4], lowerCamelCase, atol=1E-4)) @slow def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : Optional[Any] = NezhaForMaskedLM.from_pretrained('sijunhe/nezha-cn-base') _lowercase : int = torch.tensor([[0, 1, 2, 3, 4, 5]]) _lowercase : Optional[int] = torch.tensor([[1, 1, 1, 1, 1, 1]]) with torch.no_grad(): _lowercase : Dict = model(lowerCamelCase, attention_mask=lowerCamelCase)[0] _lowercase : Optional[Any] = torch.Size((1, 6, 2_11_28)) self.assertEqual(output.shape, lowerCamelCase) _lowercase : Optional[int] = torch.tensor( [[-2.7_9_3_9, -1.7_9_0_2, -2.2_1_8_9], [-2.8_5_8_5, -1.8_9_0_8, -2.3_7_2_3], [-2.6_4_9_9, -1.7_7_5_0, -2.2_5_5_8]]) self.assertTrue(torch.allclose(output[:, 1:4, 1:4], lowerCamelCase, atol=1E-4))
89
from collections import Counter from timeit import timeit def _SCREAMING_SNAKE_CASE ( lowercase : str = "" , ): '''simple docstring''' return sum(c % 2 for c in Counter(input_str.replace(' ' , '' ).lower() ).values() ) < 2 def _SCREAMING_SNAKE_CASE ( lowercase : str = "" ): '''simple docstring''' if len(lowercase ) == 0: return True lowerCamelCase_ = input_str.replace(' ' , '' ).lower() # character_freq_dict: Stores the frequency of every character in the input string lowerCamelCase_ = {} for character in lower_case_input_str: lowerCamelCase_ = character_freq_dict.get(lowercase , 0 ) + 1 lowerCamelCase_ = 0 for character_count in character_freq_dict.values(): if character_count % 2: odd_char += 1 if odd_char > 1: return False return True def _SCREAMING_SNAKE_CASE ( lowercase : str = "" ): '''simple docstring''' print('\nFor string = ' , lowercase , ':' ) print( '> can_string_be_rearranged_as_palindrome_counter()' , '\tans =' , can_string_be_rearranged_as_palindrome_counter(lowercase ) , '\ttime =' , timeit( 'z.can_string_be_rearranged_as_palindrome_counter(z.check_str)' , setup='import __main__ as z' , ) , 'seconds' , ) print( '> can_string_be_rearranged_as_palindrome()' , '\tans =' , can_string_be_rearranged_as_palindrome(lowercase ) , '\ttime =' , timeit( 'z.can_string_be_rearranged_as_palindrome(z.check_str)' , setup='import __main__ as z' , ) , 'seconds' , ) if __name__ == "__main__": lowerCamelCase : Optional[Any] = input( "Enter string to determine if it can be rearranged as a palindrome or not: " ).strip() benchmark(check_str) lowerCamelCase : int = can_string_be_rearranged_as_palindrome_counter(check_str) print(F"""{check_str} can {'' if status else 'not '}be rearranged as a palindrome""")
70
0
'''simple docstring''' import os import unittest from transformers.models.cpmant.tokenization_cpmant import VOCAB_FILES_NAMES, CpmAntTokenizer from transformers.testing_utils import require_jieba, tooslow from ...test_tokenization_common import TokenizerTesterMixin @require_jieba class a__ ( a__ , unittest.TestCase ): '''simple docstring''' lowercase__ : List[Any] = CpmAntTokenizer lowercase__ : Any = False def __SCREAMING_SNAKE_CASE ( self ) -> List[Any]: super().setUp() lowerCAmelCase__ = [ '''<d>''', '''</d>''', '''<s>''', '''</s>''', '''</_>''', '''<unk>''', '''<pad>''', '''</n>''', '''我''', '''是''', '''C''', '''P''', '''M''', '''A''', '''n''', '''t''', ] lowerCAmelCase__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) @tooslow def __SCREAMING_SNAKE_CASE ( self ) -> Dict: lowerCAmelCase__ = CpmAntTokenizer.from_pretrained('''openbmb/cpm-ant-10b''' ) lowerCAmelCase__ = '''今天天气真好!''' lowerCAmelCase__ = ['''今天''', '''天气''', '''真''', '''好''', '''!'''] lowerCAmelCase__ = tokenizer.tokenize(lowerCamelCase_ ) self.assertListEqual(lowerCamelCase_ , lowerCamelCase_ ) lowerCAmelCase__ = '''今天天气真好!''' lowerCAmelCase__ = [tokenizer.bos_token] + tokens lowerCAmelCase__ = [6, 98_02, 1_49_62, 20_82, 8_31, 2_44] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase_ ) , lowerCamelCase_ ) lowerCAmelCase__ = tokenizer.decode(lowerCamelCase_ ) self.assertEqual(lowerCamelCase_ , lowerCamelCase_ )
90
from __future__ import annotations from decimal import Decimal from math import * # noqa: F403 from sympy import diff def _SCREAMING_SNAKE_CASE ( lowercase : str , lowercase : float | Decimal , lowercase : float = 10**-10 ): '''simple docstring''' lowerCamelCase_ = a while True: lowerCamelCase_ = Decimal(lowercase ) - ( Decimal(eval(lowercase ) ) / Decimal(eval(str(diff(lowercase ) ) ) ) # noqa: S307 ) # This number dictates the accuracy of the answer if abs(eval(lowercase ) ) < precision: # noqa: S307 return float(lowercase ) # Let's Execute if __name__ == "__main__": # Find root of trigonometric function # Find value of pi print(F"""The root of sin(x) = 0 is {newton_raphson('sin(x)', 2)}""") # Find root of polynomial print(F"""The root of x**2 - 5*x + 2 = 0 is {newton_raphson('x**2 - 5*x + 2', 0.4)}""") # Find Square Root of 5 print(F"""The root of log(x) - 1 = 0 is {newton_raphson('log(x) - 1', 2)}""") # Exponential Roots print(F"""The root of exp(x) - 1 = 0 is {newton_raphson('exp(x) - 1', 0)}""")
70
0
"""simple docstring""" def _snake_case ( snake_case__ : int = 1000 ): A = 2**power A = 0 while n: A , A = r + n % 10, n // 10 return r if __name__ == "__main__": print(solution(int(str(input()).strip())))
91
from __future__ import annotations from typing import Any class A( UpperCamelCase ): '''simple docstring''' pass class A: '''simple docstring''' def __init__( self : List[str] , A_ : Any ) -> None: """simple docstring""" lowerCamelCase_ = data lowerCamelCase_ = None def __iter__( self : int ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self lowerCamelCase_ = [] while node: if node in visited: raise ContainsLoopError visited.append(A_ ) yield node.data lowerCamelCase_ = node.next_node @property def a__ ( self : List[str] ) -> bool: """simple docstring""" try: list(self ) return False except ContainsLoopError: return True if __name__ == "__main__": lowerCamelCase : int = Node(1) lowerCamelCase : Optional[int] = Node(2) lowerCamelCase : Union[str, Any] = Node(3) lowerCamelCase : List[Any] = Node(4) print(root_node.has_loop) # False lowerCamelCase : int = root_node.next_node print(root_node.has_loop) # True lowerCamelCase : Dict = Node(5) lowerCamelCase : Optional[int] = Node(6) lowerCamelCase : str = Node(5) lowerCamelCase : Union[str, Any] = Node(6) print(root_node.has_loop) # False lowerCamelCase : List[str] = Node(1) print(root_node.has_loop) # False
70
0
'''simple docstring''' import torch from diffusers import DPMSolverSDEScheduler from diffusers.utils import torch_device from diffusers.utils.testing_utils import require_torchsde from .test_schedulers import SchedulerCommonTest @require_torchsde class __SCREAMING_SNAKE_CASE ( lowercase__ ): lowerCamelCase_ = (DPMSolverSDEScheduler,) lowerCamelCase_ = 10 def lowerCamelCase_ ( self : List[Any] , **UpperCAmelCase__ : Tuple ): '''simple docstring''' lowercase : Union[str, Any] ={ '''num_train_timesteps''': 1100, '''beta_start''': 0.00_01, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', '''noise_sampler_seed''': 0, } config.update(**UpperCAmelCase__ ) return config def lowerCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' for timesteps in [10, 50, 100, 1000]: self.check_over_configs(num_train_timesteps=UpperCAmelCase__ ) def lowerCamelCase_ ( self : str ): '''simple docstring''' for beta_start, beta_end in zip([0.0_00_01, 0.00_01, 0.0_01] , [0.00_02, 0.0_02, 0.02] ): self.check_over_configs(beta_start=UpperCAmelCase__ , beta_end=UpperCAmelCase__ ) def lowerCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=UpperCAmelCase__ ) def lowerCamelCase_ ( self : str ): '''simple docstring''' for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=UpperCAmelCase__ ) def lowerCamelCase_ ( self : Optional[int] ): '''simple docstring''' lowercase : List[Any] =self.scheduler_classes[0] lowercase : Dict =self.get_scheduler_config() lowercase : Tuple =scheduler_class(**UpperCAmelCase__ ) scheduler.set_timesteps(self.num_inference_steps ) lowercase : int =self.dummy_model() lowercase : Dict =self.dummy_sample_deter * scheduler.init_noise_sigma lowercase : List[str] =sample.to(UpperCAmelCase__ ) for i, t in enumerate(scheduler.timesteps ): lowercase : List[Any] =scheduler.scale_model_input(UpperCAmelCase__ , UpperCAmelCase__ ) lowercase : List[str] =model(UpperCAmelCase__ , UpperCAmelCase__ ) lowercase : str =scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) lowercase : Tuple =output.prev_sample lowercase : int =torch.sum(torch.abs(UpperCAmelCase__ ) ) lowercase : Optional[Any] =torch.mean(torch.abs(UpperCAmelCase__ ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 1_67.47_82_10_44_92_18_75 ) < 1E-2 assert abs(result_mean.item() - 0.21_78_70_59_64_56_52_77 ) < 1E-3 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 1_71.59_35_21_11_81_64_06 ) < 1E-2 assert abs(result_mean.item() - 0.2_23_42_90_68_92_29_96_52 ) < 1E-3 else: assert abs(result_sum.item() - 1_62.52_38_34_22_85_15_62 ) < 1E-2 assert abs(result_mean.item() - 0.2_11_61_95_70_85_13_26 ) < 1E-3 def lowerCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' lowercase : List[str] =self.scheduler_classes[0] lowercase : str =self.get_scheduler_config(prediction_type='''v_prediction''' ) lowercase : Tuple =scheduler_class(**UpperCAmelCase__ ) scheduler.set_timesteps(self.num_inference_steps ) lowercase : Optional[int] =self.dummy_model() lowercase : str =self.dummy_sample_deter * scheduler.init_noise_sigma lowercase : str =sample.to(UpperCAmelCase__ ) for i, t in enumerate(scheduler.timesteps ): lowercase : Optional[int] =scheduler.scale_model_input(UpperCAmelCase__ , UpperCAmelCase__ ) lowercase : Any =model(UpperCAmelCase__ , UpperCAmelCase__ ) lowercase : List[str] =scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) lowercase : int =output.prev_sample lowercase : Any =torch.sum(torch.abs(UpperCAmelCase__ ) ) lowercase : List[str] =torch.mean(torch.abs(UpperCAmelCase__ ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 1_24.77_14_92_00_43_94_53 ) < 1E-2 assert abs(result_mean.item() - 0.1_62_26_28_90_14_81_62_84 ) < 1E-3 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 1_28.1_66_33_60_59_57_03 ) < 1E-2 assert abs(result_mean.item() - 0.1_66_88_32_60_01_16_72_97 ) < 1E-3 else: assert abs(result_sum.item() - 1_19.8_48_75_48_82_81_25 ) < 1E-2 assert abs(result_mean.item() - 0.15_60_53_06_62_53_66_21 ) < 1E-3 def lowerCamelCase_ ( self : str ): '''simple docstring''' lowercase : Optional[int] =self.scheduler_classes[0] lowercase : List[str] =self.get_scheduler_config() lowercase : List[Any] =scheduler_class(**UpperCAmelCase__ ) scheduler.set_timesteps(self.num_inference_steps , device=UpperCAmelCase__ ) lowercase : List[Any] =self.dummy_model() lowercase : Union[str, Any] =self.dummy_sample_deter.to(UpperCAmelCase__ ) * scheduler.init_noise_sigma for t in scheduler.timesteps: lowercase : List[Any] =scheduler.scale_model_input(UpperCAmelCase__ , UpperCAmelCase__ ) lowercase : Optional[Any] =model(UpperCAmelCase__ , UpperCAmelCase__ ) lowercase : Tuple =scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) lowercase : List[str] =output.prev_sample lowercase : Any =torch.sum(torch.abs(UpperCAmelCase__ ) ) lowercase : Tuple =torch.mean(torch.abs(UpperCAmelCase__ ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 1_67.46_95_73_97_46_09_38 ) < 1E-2 assert abs(result_mean.item() - 0.2_18_05_93_46_07_98_26_35 ) < 1E-3 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 1_71.59_35_36_37_69_53_12 ) < 1E-2 assert abs(result_mean.item() - 0.2_23_42_90_83_82_41_57_71 ) < 1E-3 else: assert abs(result_sum.item() - 1_62.52_38_34_22_85_15_62 ) < 1E-2 assert abs(result_mean.item() - 0.2_11_61_95_70_85_13_26 ) < 1E-3 def lowerCamelCase_ ( self : int ): '''simple docstring''' lowercase : Tuple =self.scheduler_classes[0] lowercase : Optional[Any] =self.get_scheduler_config() lowercase : Union[str, Any] =scheduler_class(**UpperCAmelCase__ , use_karras_sigmas=UpperCAmelCase__ ) scheduler.set_timesteps(self.num_inference_steps , device=UpperCAmelCase__ ) lowercase : Union[str, Any] =self.dummy_model() lowercase : str =self.dummy_sample_deter.to(UpperCAmelCase__ ) * scheduler.init_noise_sigma lowercase : List[str] =sample.to(UpperCAmelCase__ ) for t in scheduler.timesteps: lowercase : int =scheduler.scale_model_input(UpperCAmelCase__ , UpperCAmelCase__ ) lowercase : int =model(UpperCAmelCase__ , UpperCAmelCase__ ) lowercase : Dict =scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) lowercase : Optional[int] =output.prev_sample lowercase : Any =torch.sum(torch.abs(UpperCAmelCase__ ) ) lowercase : List[Any] =torch.mean(torch.abs(UpperCAmelCase__ ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 1_76.66_97_41_35_74_21_88 ) < 1E-2 assert abs(result_mean.item() - 0.2_30_03_87_27_30_98_18_11 ) < 1E-2 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 1_77.63_65_35_64_45_31_25 ) < 1E-2 assert abs(result_mean.item() - 0.2_30_03_87_27_30_98_18_11 ) < 1E-2 else: assert abs(result_sum.item() - 1_70.3_13_52_23_38_86_72 ) < 1E-2 assert abs(result_mean.item() - 0.2_30_03_87_27_30_98_18_11 ) < 1E-2
92
import unittest import torch from diffusers import DDIMScheduler, DDPMScheduler, UNetaDModel from diffusers.training_utils import set_seed from diffusers.utils.testing_utils import slow lowerCamelCase : int = False class A( unittest.TestCase ): '''simple docstring''' def a__ ( self : int , A_ : Dict=32 ) -> Any: """simple docstring""" set_seed(0 ) lowerCamelCase_ = UNetaDModel(sample_size=A_ , in_channels=3 , out_channels=3 ) lowerCamelCase_ = torch.optim.SGD(model.parameters() , lr=0.0001 ) return model, optimizer @slow def a__ ( self : int ) -> str: """simple docstring""" lowerCamelCase_ = 'cpu' # ensure full determinism without setting the CUBLAS_WORKSPACE_CONFIG env variable lowerCamelCase_ = DDPMScheduler( num_train_timesteps=1000 , beta_start=0.0001 , beta_end=0.02 , beta_schedule='linear' , clip_sample=A_ , ) lowerCamelCase_ = DDIMScheduler( num_train_timesteps=1000 , beta_start=0.0001 , beta_end=0.02 , beta_schedule='linear' , clip_sample=A_ , ) assert ddpm_scheduler.config.num_train_timesteps == ddim_scheduler.config.num_train_timesteps # shared batches for DDPM and DDIM set_seed(0 ) lowerCamelCase_ = [torch.randn((4, 3, 32, 32) ).clip(-1 , 1 ).to(A_ ) for _ in range(4 )] lowerCamelCase_ = [torch.randn((4, 3, 32, 32) ).to(A_ ) for _ in range(4 )] lowerCamelCase_ = [torch.randint(0 , 1000 , (4,) ).long().to(A_ ) for _ in range(4 )] # train with a DDPM scheduler lowerCamelCase_ , lowerCamelCase_ = self.get_model_optimizer(resolution=32 ) model.train().to(A_ ) for i in range(4 ): optimizer.zero_grad() lowerCamelCase_ = ddpm_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) lowerCamelCase_ = model(A_ , timesteps[i] ).sample lowerCamelCase_ = torch.nn.functional.mse_loss(A_ , noise[i] ) loss.backward() optimizer.step() del model, optimizer # recreate the model and optimizer, and retry with DDIM lowerCamelCase_ , lowerCamelCase_ = self.get_model_optimizer(resolution=32 ) model.train().to(A_ ) for i in range(4 ): optimizer.zero_grad() lowerCamelCase_ = ddim_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) lowerCamelCase_ = model(A_ , timesteps[i] ).sample lowerCamelCase_ = torch.nn.functional.mse_loss(A_ , noise[i] ) loss.backward() optimizer.step() del model, optimizer self.assertTrue(torch.allclose(A_ , A_ , atol=1E-5 ) ) self.assertTrue(torch.allclose(A_ , A_ , atol=1E-5 ) )
70
0
"""simple docstring""" # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __A = { """configuration_mgp_str""": ["""MGP_STR_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MgpstrConfig"""], """processing_mgp_str""": ["""MgpstrProcessor"""], """tokenization_mgp_str""": ["""MgpstrTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ """MGP_STR_PRETRAINED_MODEL_ARCHIVE_LIST""", """MgpstrModel""", """MgpstrPreTrainedModel""", """MgpstrForSceneTextRecognition""", ] if TYPE_CHECKING: from .configuration_mgp_str import MGP_STR_PRETRAINED_CONFIG_ARCHIVE_MAP, MgpstrConfig from .processing_mgp_str import MgpstrProcessor from .tokenization_mgp_str import MgpstrTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mgp_str import ( MGP_STR_PRETRAINED_MODEL_ARCHIVE_LIST, MgpstrForSceneTextRecognition, MgpstrModel, MgpstrPreTrainedModel, ) else: import sys __A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
93
def _SCREAMING_SNAKE_CASE ( lowercase : str , lowercase : str ): '''simple docstring''' if len(lowercase ) != len(lowercase ): raise ValueError('String lengths must match!' ) lowerCamelCase_ = 0 for chara, chara in zip(lowercase , lowercase ): if chara != chara: count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
70
0
'''simple docstring''' from collections.abc import Sequence def lowercase_ ( __A : Sequence[int] | None = None ) -> int: """simple docstring""" if nums is None or not nums: raise ValueError('''Input sequence should not be empty''' ) lowercase : List[str] =nums[0] for i in range(1 , len(__A ) ): lowercase : str =nums[i] lowercase : Optional[Any] =max(__A , ans + num , __A ) return ans if __name__ == "__main__": import doctest doctest.testmod() # Try on a sample input from the user SCREAMING_SNAKE_CASE = int(input('Enter number of elements : ').strip()) SCREAMING_SNAKE_CASE = list(map(int, input('\nEnter the numbers : ').strip().split()))[:n] print(max_subsequence_sum(array))
94
def _SCREAMING_SNAKE_CASE ( lowercase : int = 10 ): '''simple docstring''' if not isinstance(lowercase , lowercase ) or n < 0: raise ValueError('Invalid input' ) lowerCamelCase_ = 10**n lowerCamelCase_ = 2_84_33 * (pow(2 , 7_83_04_57 , lowercase )) + 1 return str(number % modulus ) if __name__ == "__main__": from doctest import testmod testmod() print(F"""{solution(10) = }""")
70
0
"""simple docstring""" from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging lowerCamelCase_ = logging.get_logger(__name__) # pylint: disable=invalid-name class UpperCamelCase_ (__A ): def __init__( self : Optional[int] , lowerCAmelCase_ : CLIPSegForImageSegmentation , lowerCAmelCase_ : CLIPSegProcessor , lowerCAmelCase_ : AutoencoderKL , lowerCAmelCase_ : CLIPTextModel , lowerCAmelCase_ : CLIPTokenizer , lowerCAmelCase_ : UNetaDConditionModel , lowerCAmelCase_ : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , lowerCAmelCase_ : StableDiffusionSafetyChecker , lowerCAmelCase_ : CLIPImageProcessor , ) -> Optional[int]: super().__init__() if hasattr(scheduler.config , "steps_offset" ) and scheduler.config.steps_offset != 1: UpperCAmelCase_ : Optional[int] = ( f"""The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`""" f""" should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure """ "to update the config accordingly as leaving `steps_offset` might led to incorrect results" " in future versions. If you have downloaded this checkpoint from the Hugging Face Hub," " it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`" " file" ) deprecate("steps_offset!=1" , "1.0.0" , lowerCAmelCase_ , standard_warn=lowerCAmelCase_ ) UpperCAmelCase_ : str = dict(scheduler.config ) UpperCAmelCase_ : Tuple = 1 UpperCAmelCase_ : Optional[int] = FrozenDict(lowerCAmelCase_ ) if hasattr(scheduler.config , "skip_prk_steps" ) and scheduler.config.skip_prk_steps is False: UpperCAmelCase_ : Tuple = ( f"""The configuration file of this scheduler: {scheduler} has not set the configuration""" " `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make" " sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to" " incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face" " Hub, it would be very nice if you could open a Pull request for the" " `scheduler/scheduler_config.json` file" ) deprecate("skip_prk_steps not set" , "1.0.0" , lowerCAmelCase_ , standard_warn=lowerCAmelCase_ ) UpperCAmelCase_ : Dict = dict(scheduler.config ) UpperCAmelCase_ : int = True UpperCAmelCase_ : Optional[Any] = FrozenDict(lowerCAmelCase_ ) if safety_checker is None: logger.warning( f"""You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure""" " that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered" " results in services or applications open to the public. Both the diffusers team and Hugging Face" " strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling" " it only for use-cases that involve analyzing network behavior or auditing its results. For more" " information, please have a look at https://github.com/huggingface/diffusers/pull/254 ." ) self.register_modules( segmentation_model=lowerCAmelCase_ , segmentation_processor=lowerCAmelCase_ , vae=lowerCAmelCase_ , text_encoder=lowerCAmelCase_ , tokenizer=lowerCAmelCase_ , unet=lowerCAmelCase_ , scheduler=lowerCAmelCase_ , safety_checker=lowerCAmelCase_ , feature_extractor=lowerCAmelCase_ , ) def _SCREAMING_SNAKE_CASE ( self : Dict , lowerCAmelCase_ : Optional[Union[str, int]] = "auto" ) -> Tuple: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory UpperCAmelCase_ : Optional[int] = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowerCAmelCase_ ) def _SCREAMING_SNAKE_CASE ( self : Dict ) -> str: self.enable_attention_slicing(lowerCAmelCase_ ) def _SCREAMING_SNAKE_CASE ( self : int ) -> Dict: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) UpperCAmelCase_ : Optional[int] = torch.device("cuda" ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(lowerCAmelCase_ , lowerCAmelCase_ ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def _SCREAMING_SNAKE_CASE ( self : Tuple ) -> Optional[Any]: if self.device != torch.device("meta" ) or not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowerCAmelCase_ , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self : int , lowerCAmelCase_ : Union[str, List[str]] , lowerCAmelCase_ : Union[torch.FloatTensor, PIL.Image.Image] , lowerCAmelCase_ : str , lowerCAmelCase_ : int = 512 , lowerCAmelCase_ : int = 512 , lowerCAmelCase_ : int = 50 , lowerCAmelCase_ : float = 7.5 , lowerCAmelCase_ : Optional[Union[str, List[str]]] = None , lowerCAmelCase_ : Optional[int] = 1 , lowerCAmelCase_ : float = 0.0 , lowerCAmelCase_ : Optional[torch.Generator] = None , lowerCAmelCase_ : Optional[torch.FloatTensor] = None , lowerCAmelCase_ : Optional[str] = "pil" , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , lowerCAmelCase_ : int = 1 , **lowerCAmelCase_ : Tuple , ) -> Optional[int]: UpperCAmelCase_ : List[str] = self.segmentation_processor( text=[text] , images=[image] , padding="max_length" , return_tensors="pt" ).to(self.device ) UpperCAmelCase_ : Any = self.segmentation_model(**lowerCAmelCase_ ) UpperCAmelCase_ : Any = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() UpperCAmelCase_ : List[Any] = self.numpy_to_pil(lowerCAmelCase_ )[0].resize(image.size ) # Run inpainting pipeline with the generated mask UpperCAmelCase_ : str = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=lowerCAmelCase_ , image=lowerCAmelCase_ , mask_image=lowerCAmelCase_ , height=lowerCAmelCase_ , width=lowerCAmelCase_ , num_inference_steps=lowerCAmelCase_ , guidance_scale=lowerCAmelCase_ , negative_prompt=lowerCAmelCase_ , num_images_per_prompt=lowerCAmelCase_ , eta=lowerCAmelCase_ , generator=lowerCAmelCase_ , latents=lowerCAmelCase_ , output_type=lowerCAmelCase_ , return_dict=lowerCAmelCase_ , callback=lowerCAmelCase_ , callback_steps=lowerCAmelCase_ , )
95
from maths.prime_check import is_prime def _SCREAMING_SNAKE_CASE ( lowercase : int ): '''simple docstring''' if not isinstance(lowercase , lowercase ): lowerCamelCase_ = f"""Input value of [number={number}] must be an integer""" raise TypeError(lowercase ) if is_prime(lowercase ) and is_prime(number + 2 ): return number + 2 else: return -1 if __name__ == "__main__": import doctest doctest.testmod()
70
0
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool __lowerCamelCase = { 'Acehnese Arabic': 'ace_Arab', 'Acehnese Latin': 'ace_Latn', 'Mesopotamian Arabic': 'acm_Arab', 'Ta\'izzi-Adeni Arabic': 'acq_Arab', 'Tunisian Arabic': 'aeb_Arab', 'Afrikaans': 'afr_Latn', 'South Levantine Arabic': 'ajp_Arab', 'Akan': 'aka_Latn', 'Amharic': 'amh_Ethi', 'North Levantine Arabic': 'apc_Arab', 'Modern Standard Arabic': 'arb_Arab', 'Modern Standard Arabic Romanized': 'arb_Latn', 'Najdi Arabic': 'ars_Arab', 'Moroccan Arabic': 'ary_Arab', 'Egyptian Arabic': 'arz_Arab', 'Assamese': 'asm_Beng', 'Asturian': 'ast_Latn', 'Awadhi': 'awa_Deva', 'Central Aymara': 'ayr_Latn', 'South Azerbaijani': 'azb_Arab', 'North Azerbaijani': 'azj_Latn', 'Bashkir': 'bak_Cyrl', 'Bambara': 'bam_Latn', 'Balinese': 'ban_Latn', 'Belarusian': 'bel_Cyrl', 'Bemba': 'bem_Latn', 'Bengali': 'ben_Beng', 'Bhojpuri': 'bho_Deva', 'Banjar Arabic': 'bjn_Arab', 'Banjar Latin': 'bjn_Latn', 'Standard Tibetan': 'bod_Tibt', 'Bosnian': 'bos_Latn', 'Buginese': 'bug_Latn', 'Bulgarian': 'bul_Cyrl', 'Catalan': 'cat_Latn', 'Cebuano': 'ceb_Latn', 'Czech': 'ces_Latn', 'Chokwe': 'cjk_Latn', 'Central Kurdish': 'ckb_Arab', 'Crimean Tatar': 'crh_Latn', 'Welsh': 'cym_Latn', 'Danish': 'dan_Latn', 'German': 'deu_Latn', 'Southwestern Dinka': 'dik_Latn', 'Dyula': 'dyu_Latn', 'Dzongkha': 'dzo_Tibt', 'Greek': 'ell_Grek', 'English': 'eng_Latn', 'Esperanto': 'epo_Latn', 'Estonian': 'est_Latn', 'Basque': 'eus_Latn', 'Ewe': 'ewe_Latn', 'Faroese': 'fao_Latn', 'Fijian': 'fij_Latn', 'Finnish': 'fin_Latn', 'Fon': 'fon_Latn', 'French': 'fra_Latn', 'Friulian': 'fur_Latn', 'Nigerian Fulfulde': 'fuv_Latn', 'Scottish Gaelic': 'gla_Latn', 'Irish': 'gle_Latn', 'Galician': 'glg_Latn', 'Guarani': 'grn_Latn', 'Gujarati': 'guj_Gujr', 'Haitian Creole': 'hat_Latn', 'Hausa': 'hau_Latn', 'Hebrew': 'heb_Hebr', 'Hindi': 'hin_Deva', 'Chhattisgarhi': 'hne_Deva', 'Croatian': 'hrv_Latn', 'Hungarian': 'hun_Latn', 'Armenian': 'hye_Armn', 'Igbo': 'ibo_Latn', 'Ilocano': 'ilo_Latn', 'Indonesian': 'ind_Latn', 'Icelandic': 'isl_Latn', 'Italian': 'ita_Latn', 'Javanese': 'jav_Latn', 'Japanese': 'jpn_Jpan', 'Kabyle': 'kab_Latn', 'Jingpho': 'kac_Latn', 'Kamba': 'kam_Latn', 'Kannada': 'kan_Knda', 'Kashmiri Arabic': 'kas_Arab', 'Kashmiri Devanagari': 'kas_Deva', 'Georgian': 'kat_Geor', 'Central Kanuri Arabic': 'knc_Arab', 'Central Kanuri Latin': 'knc_Latn', 'Kazakh': 'kaz_Cyrl', 'Kabiyè': 'kbp_Latn', 'Kabuverdianu': 'kea_Latn', 'Khmer': 'khm_Khmr', 'Kikuyu': 'kik_Latn', 'Kinyarwanda': 'kin_Latn', 'Kyrgyz': 'kir_Cyrl', 'Kimbundu': 'kmb_Latn', 'Northern Kurdish': 'kmr_Latn', 'Kikongo': 'kon_Latn', 'Korean': 'kor_Hang', 'Lao': 'lao_Laoo', 'Ligurian': 'lij_Latn', 'Limburgish': 'lim_Latn', 'Lingala': 'lin_Latn', 'Lithuanian': 'lit_Latn', 'Lombard': 'lmo_Latn', 'Latgalian': 'ltg_Latn', 'Luxembourgish': 'ltz_Latn', 'Luba-Kasai': 'lua_Latn', 'Ganda': 'lug_Latn', 'Luo': 'luo_Latn', 'Mizo': 'lus_Latn', 'Standard Latvian': 'lvs_Latn', 'Magahi': 'mag_Deva', 'Maithili': 'mai_Deva', 'Malayalam': 'mal_Mlym', 'Marathi': 'mar_Deva', 'Minangkabau Arabic ': 'min_Arab', 'Minangkabau Latin': 'min_Latn', 'Macedonian': 'mkd_Cyrl', 'Plateau Malagasy': 'plt_Latn', 'Maltese': 'mlt_Latn', 'Meitei Bengali': 'mni_Beng', 'Halh Mongolian': 'khk_Cyrl', 'Mossi': 'mos_Latn', 'Maori': 'mri_Latn', 'Burmese': 'mya_Mymr', 'Dutch': 'nld_Latn', 'Norwegian Nynorsk': 'nno_Latn', 'Norwegian Bokmål': 'nob_Latn', 'Nepali': 'npi_Deva', 'Northern Sotho': 'nso_Latn', 'Nuer': 'nus_Latn', 'Nyanja': 'nya_Latn', 'Occitan': 'oci_Latn', 'West Central Oromo': 'gaz_Latn', 'Odia': 'ory_Orya', 'Pangasinan': 'pag_Latn', 'Eastern Panjabi': 'pan_Guru', 'Papiamento': 'pap_Latn', 'Western Persian': 'pes_Arab', 'Polish': 'pol_Latn', 'Portuguese': 'por_Latn', 'Dari': 'prs_Arab', 'Southern Pashto': 'pbt_Arab', 'Ayacucho Quechua': 'quy_Latn', 'Romanian': 'ron_Latn', 'Rundi': 'run_Latn', 'Russian': 'rus_Cyrl', 'Sango': 'sag_Latn', 'Sanskrit': 'san_Deva', 'Santali': 'sat_Olck', 'Sicilian': 'scn_Latn', 'Shan': 'shn_Mymr', 'Sinhala': 'sin_Sinh', 'Slovak': 'slk_Latn', 'Slovenian': 'slv_Latn', 'Samoan': 'smo_Latn', 'Shona': 'sna_Latn', 'Sindhi': 'snd_Arab', 'Somali': 'som_Latn', 'Southern Sotho': 'sot_Latn', 'Spanish': 'spa_Latn', 'Tosk Albanian': 'als_Latn', 'Sardinian': 'srd_Latn', 'Serbian': 'srp_Cyrl', 'Swati': 'ssw_Latn', 'Sundanese': 'sun_Latn', 'Swedish': 'swe_Latn', 'Swahili': 'swh_Latn', 'Silesian': 'szl_Latn', 'Tamil': 'tam_Taml', 'Tatar': 'tat_Cyrl', 'Telugu': 'tel_Telu', 'Tajik': 'tgk_Cyrl', 'Tagalog': 'tgl_Latn', 'Thai': 'tha_Thai', 'Tigrinya': 'tir_Ethi', 'Tamasheq Latin': 'taq_Latn', 'Tamasheq Tifinagh': 'taq_Tfng', 'Tok Pisin': 'tpi_Latn', 'Tswana': 'tsn_Latn', 'Tsonga': 'tso_Latn', 'Turkmen': 'tuk_Latn', 'Tumbuka': 'tum_Latn', 'Turkish': 'tur_Latn', 'Twi': 'twi_Latn', 'Central Atlas Tamazight': 'tzm_Tfng', 'Uyghur': 'uig_Arab', 'Ukrainian': 'ukr_Cyrl', 'Umbundu': 'umb_Latn', 'Urdu': 'urd_Arab', 'Northern Uzbek': 'uzn_Latn', 'Venetian': 'vec_Latn', 'Vietnamese': 'vie_Latn', 'Waray': 'war_Latn', 'Wolof': 'wol_Latn', 'Xhosa': 'xho_Latn', 'Eastern Yiddish': 'ydd_Hebr', 'Yoruba': 'yor_Latn', 'Yue Chinese': 'yue_Hant', 'Chinese Simplified': 'zho_Hans', 'Chinese Traditional': 'zho_Hant', 'Standard Malay': 'zsm_Latn', 'Zulu': 'zul_Latn', } class __A ( SCREAMING_SNAKE_CASE_ ): UpperCAmelCase__ = "facebook/nllb-200-distilled-600M" UpperCAmelCase__ = ( "This is a tool that translates text from a language to another. It takes three inputs: `text`, which should " "be the text to translate, `src_lang`, which should be the language of the text to translate and `tgt_lang`, " "which should be the language for the desired ouput language. Both `src_lang` and `tgt_lang` are written in " "plain English, such as 'Romanian', or 'Albanian'. It returns the text translated in `tgt_lang`." ) UpperCAmelCase__ = "translator" UpperCAmelCase__ = AutoTokenizer UpperCAmelCase__ = AutoModelForSeqaSeqLM UpperCAmelCase__ = LANGUAGE_CODES UpperCAmelCase__ = ["text", "text", "text"] UpperCAmelCase__ = ["text"] def lowerCamelCase__ ( self : List[Any] , __snake_case : int , __snake_case : Union[str, Any] , __snake_case : Dict ) -> Union[str, Any]: if src_lang not in self.lang_to_code: raise ValueError(F'{src_lang} is not a supported language.' ) if tgt_lang not in self.lang_to_code: raise ValueError(F'{tgt_lang} is not a supported language.' ) __magic_name__: Union[str, Any] = self.lang_to_code[src_lang] __magic_name__: int = self.lang_to_code[tgt_lang] return self.pre_processor._build_translation_inputs( __snake_case , return_tensors="""pt""" , src_lang=__snake_case , tgt_lang=__snake_case ) def lowerCamelCase__ ( self : Any , __snake_case : Tuple ) -> Any: return self.model.generate(**__snake_case ) def lowerCamelCase__ ( self : List[Any] , __snake_case : List[str] ) -> Dict: return self.post_processor.decode(outputs[0].tolist() , skip_special_tokens=__snake_case )
96
# Algorithm for the pigeonhole sorting def _SCREAMING_SNAKE_CASE ( lowercase : str ): '''simple docstring''' lowerCamelCase_ = min(lowercase ) # min() finds the minimum value lowerCamelCase_ = max(lowercase ) # max() finds the maximum value lowerCamelCase_ = max_val - min_val + 1 # size is difference of max and min values plus one # list of pigeonholes of size equal to the variable size lowerCamelCase_ = [0] * size # Populate the pigeonholes. for x in a: assert isinstance(lowercase , lowercase ), "integers only please" holes[x - min_val] += 1 # Putting the elements back into the array in an order. lowerCamelCase_ = 0 for count in range(lowercase ): while holes[count] > 0: holes[count] -= 1 lowerCamelCase_ = count + min_val i += 1 def _SCREAMING_SNAKE_CASE ( ): '''simple docstring''' lowerCamelCase_ = [8, 3, 2, 7, 4, 6, 8] pigeonhole_sort(lowercase ) print('Sorted order is:' , ' '.join(lowercase ) ) if __name__ == "__main__": main()
70
0
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __a = logging.get_logger(__name__) __a = {'vocab_file': 'spm_char.model'} __a = { 'vocab_file': { 'microsoft/speecht5_asr': 'https://huggingface.co/microsoft/speecht5_asr/resolve/main/spm_char.model', 'microsoft/speecht5_tts': 'https://huggingface.co/microsoft/speecht5_tts/resolve/main/spm_char.model', 'microsoft/speecht5_vc': 'https://huggingface.co/microsoft/speecht5_vc/resolve/main/spm_char.model', } } __a = { 'microsoft/speecht5_asr': 1_0_2_4, 'microsoft/speecht5_tts': 1_0_2_4, 'microsoft/speecht5_vc': 1_0_2_4, } class lowercase__( UpperCAmelCase ): """simple docstring""" a :Dict = VOCAB_FILES_NAMES a :Dict = PRETRAINED_VOCAB_FILES_MAP a :int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a :str = ['input_ids', 'attention_mask'] def __init__( self : Any , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : Any="<s>" , SCREAMING_SNAKE_CASE_ : List[Any]="</s>" , SCREAMING_SNAKE_CASE_ : str="<unk>" , SCREAMING_SNAKE_CASE_ : Tuple="<pad>" , SCREAMING_SNAKE_CASE_ : Optional[Dict[str, Any]] = None , **SCREAMING_SNAKE_CASE_ : Dict , ) -> None: lowercase_ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=SCREAMING_SNAKE_CASE_ , eos_token=SCREAMING_SNAKE_CASE_ , unk_token=SCREAMING_SNAKE_CASE_ , pad_token=SCREAMING_SNAKE_CASE_ , sp_model_kwargs=self.sp_model_kwargs , **SCREAMING_SNAKE_CASE_ , ) lowercase_ = vocab_file lowercase_ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(SCREAMING_SNAKE_CASE_ ) @property def _lowercase ( self : Optional[Any] ) -> Dict: return self.sp_model.get_piece_size() def _lowercase ( self : Optional[int] ) -> List[Any]: lowercase_ = {self.convert_ids_to_tokens(SCREAMING_SNAKE_CASE_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : int ) -> List[Any]: lowercase_ = self.__dict__.copy() lowercase_ = None return state def __setstate__( self : Union[str, Any] , SCREAMING_SNAKE_CASE_ : int ) -> Union[str, Any]: lowercase_ = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): lowercase_ = {} lowercase_ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _lowercase ( self : Dict , SCREAMING_SNAKE_CASE_ : str ) -> List[str]: return self.sp_model.encode(SCREAMING_SNAKE_CASE_ , out_type=SCREAMING_SNAKE_CASE_ ) def _lowercase ( self : Optional[Any] , SCREAMING_SNAKE_CASE_ : Optional[Any] ) -> Optional[int]: return self.sp_model.piece_to_id(SCREAMING_SNAKE_CASE_ ) def _lowercase ( self : Union[str, Any] , SCREAMING_SNAKE_CASE_ : List[Any] ) -> Optional[int]: lowercase_ = self.sp_model.IdToPiece(SCREAMING_SNAKE_CASE_ ) return token def _lowercase ( self : Tuple , SCREAMING_SNAKE_CASE_ : List[str] ) -> str: lowercase_ = [] lowercase_ = '''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(SCREAMING_SNAKE_CASE_ ) + token lowercase_ = [] else: current_sub_tokens.append(SCREAMING_SNAKE_CASE_ ) out_string += self.sp_model.decode(SCREAMING_SNAKE_CASE_ ) return out_string.strip() def _lowercase ( self : List[str] , SCREAMING_SNAKE_CASE_ : Union[str, Any] , SCREAMING_SNAKE_CASE_ : Dict=None ) -> List[int]: if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def _lowercase ( self : Union[str, Any] , SCREAMING_SNAKE_CASE_ : List[int] , SCREAMING_SNAKE_CASE_ : Optional[List[int]] = None , SCREAMING_SNAKE_CASE_ : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=SCREAMING_SNAKE_CASE_ , token_ids_a=SCREAMING_SNAKE_CASE_ , already_has_special_tokens=SCREAMING_SNAKE_CASE_ ) lowercase_ = [1] if token_ids_a is None: return ([0] * len(SCREAMING_SNAKE_CASE_ )) + suffix_ones return ([0] * len(SCREAMING_SNAKE_CASE_ )) + ([0] * len(SCREAMING_SNAKE_CASE_ )) + suffix_ones def _lowercase ( self : str , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(SCREAMING_SNAKE_CASE_ ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return lowercase_ = os.path.join( SCREAMING_SNAKE_CASE_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(SCREAMING_SNAKE_CASE_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , SCREAMING_SNAKE_CASE_ ) elif not os.path.isfile(self.vocab_file ): with open(SCREAMING_SNAKE_CASE_ , '''wb''' ) as fi: lowercase_ = self.sp_model.serialized_model_proto() fi.write(SCREAMING_SNAKE_CASE_ ) return (out_vocab_file,)
97
import os import unittest from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, BertTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class A( UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = BertTokenizer UpperCamelCase = BertTokenizerFast UpperCamelCase = True UpperCamelCase = True UpperCamelCase = filter_non_english def a__ ( self : List[Any] ) -> int: """simple docstring""" super().setUp() lowerCamelCase_ = [ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] lowerCamelCase_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def a__ ( self : Tuple , A_ : List[str] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = 'UNwant\u00E9d,running' lowerCamelCase_ = 'unwanted, running' return input_text, output_text def a__ ( self : Any ) -> Tuple: """simple docstring""" lowerCamelCase_ = self.tokenizer_class(self.vocab_file ) lowerCamelCase_ = tokenizer.tokenize('UNwant\u00E9d,running' ) self.assertListEqual(A_ , ['un', '##want', '##ed', ',', 'runn', '##ing'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A_ ) , [9, 6, 7, 12, 10, 11] ) def a__ ( self : Tuple ) -> List[str]: """simple docstring""" if not self.test_rust_tokenizer: return lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_rust_tokenizer() lowerCamelCase_ = 'UNwant\u00E9d,running' lowerCamelCase_ = tokenizer.tokenize(A_ ) lowerCamelCase_ = rust_tokenizer.tokenize(A_ ) self.assertListEqual(A_ , A_ ) lowerCamelCase_ = tokenizer.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = rust_tokenizer.encode(A_ , add_special_tokens=A_ ) self.assertListEqual(A_ , A_ ) lowerCamelCase_ = self.get_rust_tokenizer() lowerCamelCase_ = tokenizer.encode(A_ ) lowerCamelCase_ = rust_tokenizer.encode(A_ ) self.assertListEqual(A_ , A_ ) # With lower casing lowerCamelCase_ = self.get_tokenizer(do_lower_case=A_ ) lowerCamelCase_ = self.get_rust_tokenizer(do_lower_case=A_ ) lowerCamelCase_ = 'UNwant\u00E9d,running' lowerCamelCase_ = tokenizer.tokenize(A_ ) lowerCamelCase_ = rust_tokenizer.tokenize(A_ ) self.assertListEqual(A_ , A_ ) lowerCamelCase_ = tokenizer.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = rust_tokenizer.encode(A_ , add_special_tokens=A_ ) self.assertListEqual(A_ , A_ ) lowerCamelCase_ = self.get_rust_tokenizer() lowerCamelCase_ = tokenizer.encode(A_ ) lowerCamelCase_ = rust_tokenizer.encode(A_ ) self.assertListEqual(A_ , A_ ) def a__ ( self : Any ) -> Dict: """simple docstring""" lowerCamelCase_ = BasicTokenizer() self.assertListEqual(tokenizer.tokenize('ah\u535A\u63A8zz' ) , ['ah', '\u535A', '\u63A8', 'zz'] ) def a__ ( self : Dict ) -> int: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? ' ) , ['hello', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['hello'] ) def a__ ( self : str ) -> int: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ , strip_accents=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['hällo', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['h\u00E9llo'] ) def a__ ( self : Optional[int] ) -> Any: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ , strip_accents=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['hallo', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['hello'] ) def a__ ( self : Tuple ) -> str: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['hallo', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['hello'] ) def a__ ( self : int ) -> List[Any]: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? ' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?'] ) def a__ ( self : str ) -> List[str]: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ , strip_accents=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['HäLLo', '!', 'how', 'Are', 'yoU', '?'] ) def a__ ( self : Dict ) -> Any: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ , strip_accents=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['HaLLo', '!', 'how', 'Are', 'yoU', '?'] ) def a__ ( self : int ) -> Any: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ , never_split=['[UNK]'] ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? [UNK]' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?', '[UNK]'] ) def a__ ( self : List[Any] ) -> int: """simple docstring""" lowerCamelCase_ = BasicTokenizer() lowerCamelCase_ = 'a\n\'ll !!to?\'d of, can\'t.' lowerCamelCase_ = ['a', '\'', 'll', '!', '!', 'to', '?', '\'', 'd', 'of', ',', 'can', '\'', 't', '.'] self.assertListEqual(tokenizer.tokenize(A_ ) , A_ ) def a__ ( self : Optional[int] ) -> Dict: """simple docstring""" lowerCamelCase_ = ['[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing'] lowerCamelCase_ = {} for i, token in enumerate(A_ ): lowerCamelCase_ = i lowerCamelCase_ = WordpieceTokenizer(vocab=A_ , unk_token='[UNK]' ) self.assertListEqual(tokenizer.tokenize('' ) , [] ) self.assertListEqual(tokenizer.tokenize('unwanted running' ) , ['un', '##want', '##ed', 'runn', '##ing'] ) self.assertListEqual(tokenizer.tokenize('unwantedX running' ) , ['[UNK]', 'runn', '##ing'] ) def a__ ( self : Optional[Any] ) -> str: """simple docstring""" self.assertTrue(_is_whitespace(' ' ) ) self.assertTrue(_is_whitespace('\t' ) ) self.assertTrue(_is_whitespace('\r' ) ) self.assertTrue(_is_whitespace('\n' ) ) self.assertTrue(_is_whitespace('\u00A0' ) ) self.assertFalse(_is_whitespace('A' ) ) self.assertFalse(_is_whitespace('-' ) ) def a__ ( self : List[Any] ) -> int: """simple docstring""" self.assertTrue(_is_control('\u0005' ) ) self.assertFalse(_is_control('A' ) ) self.assertFalse(_is_control(' ' ) ) self.assertFalse(_is_control('\t' ) ) self.assertFalse(_is_control('\r' ) ) def a__ ( self : List[str] ) -> Optional[Any]: """simple docstring""" self.assertTrue(_is_punctuation('-' ) ) self.assertTrue(_is_punctuation('$' ) ) self.assertTrue(_is_punctuation('`' ) ) self.assertTrue(_is_punctuation('.' ) ) self.assertFalse(_is_punctuation('A' ) ) self.assertFalse(_is_punctuation(' ' ) ) def a__ ( self : int ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_rust_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(A_ ) for t in ['Test', '\xad', 'test']] , [['[UNK]'], [], ['[UNK]']] ) self.assertListEqual( [rust_tokenizer.tokenize(A_ ) for t in ['Test', '\xad', 'test']] , [['[UNK]'], [], ['[UNK]']] ) @slow def a__ ( self : Any ) -> int: """simple docstring""" lowerCamelCase_ = self.tokenizer_class.from_pretrained('bert-base-uncased' ) lowerCamelCase_ = tokenizer.encode('sequence builders' , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer.encode('multi-sequence build' , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer.build_inputs_with_special_tokens(A_ ) lowerCamelCase_ = tokenizer.build_inputs_with_special_tokens(A_ , A_ ) assert encoded_sentence == [101] + text + [102] assert encoded_pair == [101] + text + [102] + text_a + [102] def a__ ( self : str ) -> str: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCamelCase_ = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) lowerCamelCase_ = f"""A, naïve {tokenizer_r.mask_token} AllenNLP sentence.""" lowerCamelCase_ = tokenizer_r.encode_plus( A_ , return_attention_mask=A_ , return_token_type_ids=A_ , return_offsets_mapping=A_ , add_special_tokens=A_ , ) lowerCamelCase_ = tokenizer_r.do_lower_case if hasattr(A_ , 'do_lower_case' ) else False lowerCamelCase_ = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), 'A'), ((1, 2), ','), ((3, 5), 'na'), ((5, 6), '##ï'), ((6, 8), '##ve'), ((9, 15), tokenizer_r.mask_token), ((16, 21), 'Allen'), ((21, 23), '##NL'), ((23, 24), '##P'), ((25, 33), 'sentence'), ((33, 34), '.'), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), 'a'), ((1, 2), ','), ((3, 8), 'naive'), ((9, 15), tokenizer_r.mask_token), ((16, 21), 'allen'), ((21, 23), '##nl'), ((23, 24), '##p'), ((25, 33), 'sentence'), ((33, 34), '.'), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens['input_ids'] ) ) self.assertEqual([e[0] for e in expected_results] , tokens['offset_mapping'] ) def a__ ( self : List[Any] ) -> Dict: """simple docstring""" lowerCamelCase_ = ['的', '人', '有'] lowerCamelCase_ = ''.join(A_ ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCamelCase_ = True lowerCamelCase_ = self.tokenizer_class.from_pretrained(A_ , **A_ ) lowerCamelCase_ = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) lowerCamelCase_ = tokenizer_p.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer_r.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer_r.convert_ids_to_tokens(A_ ) lowerCamelCase_ = tokenizer_p.convert_ids_to_tokens(A_ ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(A_ , A_ ) self.assertListEqual(A_ , A_ ) lowerCamelCase_ = False lowerCamelCase_ = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) lowerCamelCase_ = self.tokenizer_class.from_pretrained(A_ , **A_ ) lowerCamelCase_ = tokenizer_r.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer_p.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer_r.convert_ids_to_tokens(A_ ) lowerCamelCase_ = tokenizer_p.convert_ids_to_tokens(A_ ) # it is expected that only the first Chinese character is not preceded by "##". lowerCamelCase_ = [ f"""##{token}""" if idx != 0 else token for idx, token in enumerate(A_ ) ] self.assertListEqual(A_ , A_ ) self.assertListEqual(A_ , A_ )
70
0
'''simple docstring''' from __future__ import annotations from collections.abc import MutableSequence class __lowerCAmelCase : """simple docstring""" def __init__( self : Tuple , lowerCAmelCase__ : int , lowerCAmelCase__ : MutableSequence[float] ) -> None: '''simple docstring''' if len(lowerCAmelCase__ ) != degree + 1: raise ValueError( '''The number of coefficients should be equal to the degree + 1.''' ) _UpperCamelCase = list(lowerCAmelCase__ ) _UpperCamelCase = degree def __add__( self : int , lowerCAmelCase__ : Polynomial ) -> Polynomial: '''simple docstring''' if self.degree > polynomial_a.degree: _UpperCamelCase = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , lowerCAmelCase__ ) else: _UpperCamelCase = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , lowerCAmelCase__ ) def __sub__( self : Dict , lowerCAmelCase__ : Polynomial ) -> Polynomial: '''simple docstring''' return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self : Dict ) -> Polynomial: '''simple docstring''' return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self : Dict , lowerCAmelCase__ : Polynomial ) -> Polynomial: '''simple docstring''' _UpperCamelCase = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , lowerCAmelCase__ ) def snake_case__ ( self : int , lowerCAmelCase__ : int | float ) -> int | float: '''simple docstring''' _UpperCamelCase = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Tuple ) -> str: '''simple docstring''' _UpperCamelCase = '''''' for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(lowerCAmelCase__ ) return polynomial def __repr__( self : List[Any] ) -> str: '''simple docstring''' return self.__str__() def snake_case__ ( self : Any ) -> Polynomial: '''simple docstring''' _UpperCamelCase = [0] * self.degree for i in range(self.degree ): _UpperCamelCase = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , lowerCAmelCase__ ) def snake_case__ ( self : Union[str, Any] , lowerCAmelCase__ : int | float = 0 ) -> Polynomial: '''simple docstring''' _UpperCamelCase = [0] * (self.degree + 2) _UpperCamelCase = constant for i in range(self.degree + 1 ): _UpperCamelCase = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , lowerCAmelCase__ ) def __eq__( self : Optional[Any] , lowerCAmelCase__ : object ) -> bool: '''simple docstring''' if not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : Any , lowerCAmelCase__ : object ) -> bool: '''simple docstring''' return not self.__eq__(lowerCAmelCase__ )
98
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) lowerCamelCase : Dict = logging.get_logger(__name__) # pylint: disable=invalid-name lowerCamelCase : List[str] = "\n Examples:\n ```py\n >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline\n >>> import torch\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-prior\")\n >>> pipe_prior.to(\"cuda\")\n >>> prompt = \"red cat, 4k photo\"\n >>> out = pipe_prior(prompt)\n >>> image_emb = out.image_embeds\n >>> zero_image_emb = out.negative_image_embeds\n >>> pipe = KandinskyV22Pipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-decoder\")\n >>> pipe.to(\"cuda\")\n >>> image = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=50,\n ... ).images\n >>> image[0].save(\"cat.png\")\n ```\n" def _SCREAMING_SNAKE_CASE ( lowercase : Any , lowercase : str , lowercase : Any=8 ): '''simple docstring''' lowerCamelCase_ = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 lowerCamelCase_ = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class A( UpperCamelCase ): '''simple docstring''' def __init__( self : str , A_ : UNetaDConditionModel , A_ : DDPMScheduler , A_ : VQModel , ) -> List[str]: """simple docstring""" super().__init__() self.register_modules( unet=A_ , scheduler=A_ , movq=A_ , ) lowerCamelCase_ = 2 ** (len(self.movq.config.block_out_channels ) - 1) def a__ ( self : List[Any] , A_ : Tuple , A_ : Dict , A_ : List[Any] , A_ : int , A_ : Any , A_ : Tuple ) -> Any: """simple docstring""" if latents is None: lowerCamelCase_ = randn_tensor(A_ , generator=A_ , device=A_ , dtype=A_ ) else: if latents.shape != shape: raise ValueError(f"""Unexpected latents shape, got {latents.shape}, expected {shape}""" ) lowerCamelCase_ = latents.to(A_ ) lowerCamelCase_ = latents * scheduler.init_noise_sigma return latents def a__ ( self : int , A_ : str=0 ) -> Optional[int]: """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) lowerCamelCase_ = torch.device(f"""cuda:{gpu_id}""" ) lowerCamelCase_ = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(A_ , A_ ) def a__ ( self : Tuple , A_ : Union[str, Any]=0 ) -> Dict: """simple docstring""" if is_accelerate_available() and is_accelerate_version('>=' , '0.17.0.dev0' ): from accelerate import cpu_offload_with_hook else: raise ImportError('`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.' ) lowerCamelCase_ = torch.device(f"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to('cpu' , silence_dtype_warnings=A_ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) lowerCamelCase_ = None for cpu_offloaded_model in [self.unet, self.movq]: lowerCamelCase_ , lowerCamelCase_ = cpu_offload_with_hook(A_ , A_ , prev_module_hook=A_ ) # We'll offload the last model manually. lowerCamelCase_ = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def a__ ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" if not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(A_ , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(A_ ) def __call__( self : List[Any] , A_ : Union[torch.FloatTensor, List[torch.FloatTensor]] , A_ : Union[torch.FloatTensor, List[torch.FloatTensor]] , A_ : int = 512 , A_ : int = 512 , A_ : int = 100 , A_ : float = 4.0 , A_ : int = 1 , A_ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , A_ : Optional[torch.FloatTensor] = None , A_ : Optional[str] = "pil" , A_ : bool = True , ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self._execution_device lowerCamelCase_ = guidance_scale > 1.0 if isinstance(A_ , A_ ): lowerCamelCase_ = torch.cat(A_ , dim=0 ) lowerCamelCase_ = image_embeds.shape[0] * num_images_per_prompt if isinstance(A_ , A_ ): lowerCamelCase_ = torch.cat(A_ , dim=0 ) if do_classifier_free_guidance: lowerCamelCase_ = image_embeds.repeat_interleave(A_ , dim=0 ) lowerCamelCase_ = negative_image_embeds.repeat_interleave(A_ , dim=0 ) lowerCamelCase_ = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=A_ ) self.scheduler.set_timesteps(A_ , device=A_ ) lowerCamelCase_ = self.scheduler.timesteps lowerCamelCase_ = self.unet.config.in_channels lowerCamelCase_ , lowerCamelCase_ = downscale_height_and_width(A_ , A_ , self.movq_scale_factor ) # create initial latent lowerCamelCase_ = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , A_ , A_ , A_ , self.scheduler , ) for i, t in enumerate(self.progress_bar(A_ ) ): # expand the latents if we are doing classifier free guidance lowerCamelCase_ = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowerCamelCase_ = {'image_embeds': image_embeds} lowerCamelCase_ = self.unet( sample=A_ , timestep=A_ , encoder_hidden_states=A_ , added_cond_kwargs=A_ , return_dict=A_ , )[0] if do_classifier_free_guidance: lowerCamelCase_ , lowerCamelCase_ = noise_pred.split(latents.shape[1] , dim=1 ) lowerCamelCase_ , lowerCamelCase_ = noise_pred.chunk(2 ) lowerCamelCase_ , lowerCamelCase_ = variance_pred.chunk(2 ) lowerCamelCase_ = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) lowerCamelCase_ = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , 'variance_type' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): lowerCamelCase_ , lowerCamelCase_ = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 lowerCamelCase_ = self.scheduler.step( A_ , A_ , A_ , generator=A_ , )[0] # post-processing lowerCamelCase_ = self.movq.decode(A_ , force_not_quantize=A_ )['sample'] if output_type not in ["pt", "np", "pil"]: raise ValueError(f"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: lowerCamelCase_ = image * 0.5 + 0.5 lowerCamelCase_ = image.clamp(0 , 1 ) lowerCamelCase_ = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowerCamelCase_ = self.numpy_to_pil(A_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=A_ )
70
0
# Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def a (lowerCAmelCase__ ): return 1 / (1 + np.exp(-z )) def a (lowerCAmelCase__ , lowerCAmelCase__ ): return (-y * np.log(lowerCAmelCase__ ) - (1 - y) * np.log(1 - h )).mean() def a (lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): __a = np.dot(lowerCAmelCase__ , lowerCAmelCase__ ) return np.sum(y * scores - np.log(1 + np.exp(lowerCAmelCase__ ) ) ) def a (lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=70_000 ): __a = np.zeros(x.shape[1] ) for iterations in range(lowerCAmelCase__ ): __a = np.dot(lowerCAmelCase__ , lowerCAmelCase__ ) __a = sigmoid_function(lowerCAmelCase__ ) __a = np.dot(x.T , h - y ) / y.size __a = theta - alpha * gradient # updating the weights __a = np.dot(lowerCAmelCase__ , lowerCAmelCase__ ) __a = sigmoid_function(lowerCAmelCase__ ) __a = cost_function(lowerCAmelCase__ , lowerCAmelCase__ ) if iterations % 100 == 0: print(f'''loss: {j} \t''' ) # printing the loss after every 100 iterations return theta # In[68]: if __name__ == "__main__": SCREAMING_SNAKE_CASE = datasets.load_iris() SCREAMING_SNAKE_CASE = iris.data[:, :2] SCREAMING_SNAKE_CASE = (iris.target != 0) * 1 SCREAMING_SNAKE_CASE = 0.1 SCREAMING_SNAKE_CASE = logistic_reg(alpha, x, y, max_iterations=7_0_0_0_0) print('theta: ', theta) # printing the theta i.e our weights vector def a (lowerCAmelCase__ ): return sigmoid_function( np.dot(lowerCAmelCase__ , lowerCAmelCase__ ) ) # predicting the value of probability from the logistic regression algorithm plt.figure(figsize=(1_0, 6)) plt.scatter(x[y == 0][:, 0], x[y == 0][:, 1], color='b', label='0') plt.scatter(x[y == 1][:, 0], x[y == 1][:, 1], color='r', label='1') ((SCREAMING_SNAKE_CASE) , (SCREAMING_SNAKE_CASE)) = (x[:, 0].min(), x[:, 0].max()) ((SCREAMING_SNAKE_CASE) , (SCREAMING_SNAKE_CASE)) = (x[:, 1].min(), x[:, 1].max()) ((SCREAMING_SNAKE_CASE) , (SCREAMING_SNAKE_CASE)) = np.meshgrid(np.linspace(xa_min, xa_max), np.linspace(xa_min, xa_max)) SCREAMING_SNAKE_CASE = np.c_[xxa.ravel(), xxa.ravel()] SCREAMING_SNAKE_CASE = predict_prob(grid).reshape(xxa.shape) plt.contour(xxa, xxa, probs, [0.5], linewidths=1, colors='black') plt.legend() plt.show()
99
from PIL import Image def _SCREAMING_SNAKE_CASE ( lowercase : Image ): '''simple docstring''' lowerCamelCase_ , lowerCamelCase_ = image.size lowerCamelCase_ = 0 lowerCamelCase_ = image.load() for i in range(lowercase ): for j in range(lowercase ): lowerCamelCase_ = pixels[j, i] mean += pixel mean //= width * height for j in range(lowercase ): for i in range(lowercase ): lowerCamelCase_ = 2_55 if pixels[i, j] > mean else 0 return image if __name__ == "__main__": lowerCamelCase : Optional[Any] = mean_threshold(Image.open("path_to_image").convert("L")) image.save("output_image_path")
70
0
import baseaa import io import json import os from copy import deepcopy from ..optimizer import AcceleratedOptimizer from ..scheduler import AcceleratedScheduler class __snake_case : '''simple docstring''' def __init__( self , A_ ): '''simple docstring''' if isinstance(A_ , A_ ): # Don't modify user's data should they want to reuse it (e.g. in tests), because once we # modified it, it will not be accepted here again, since `auto` values would have been overridden SCREAMING_SNAKE_CASE__ = deepcopy(A_ ) elif os.path.exists(A_ ): with io.open(A_ , '''r''' , encoding='''utf-8''' ) as f: SCREAMING_SNAKE_CASE__ = json.load(A_ ) else: try: SCREAMING_SNAKE_CASE__ = baseaa.urlsafe_baadecode(A_ ).decode('''utf-8''' ) SCREAMING_SNAKE_CASE__ = json.loads(A_ ) except (UnicodeDecodeError, AttributeError, ValueError): raise ValueError( f'''Expected a string path to an existing deepspeed config, or a dictionary, or a base64 encoded string. Received: {config_file_or_dict}''' ) SCREAMING_SNAKE_CASE__ = config self.set_stage_and_offload() def lowercase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = self.get_value('''zero_optimization.stage''' , -1 ) # offload SCREAMING_SNAKE_CASE__ = False if self.is_zeroa() or self.is_zeroa(): SCREAMING_SNAKE_CASE__ = set(['''cpu''', '''nvme'''] ) SCREAMING_SNAKE_CASE__ = set( [ self.get_value('''zero_optimization.offload_optimizer.device''' ), self.get_value('''zero_optimization.offload_param.device''' ), ] ) if len(offload_devices & offload_devices_valid ) > 0: SCREAMING_SNAKE_CASE__ = True def lowercase_ ( self , A_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = self.config # find the config node of interest if it exists SCREAMING_SNAKE_CASE__ = ds_key_long.split('''.''' ) SCREAMING_SNAKE_CASE__ = nodes.pop() for node in nodes: SCREAMING_SNAKE_CASE__ = config.get(A_ ) if config is None: return None, ds_key return config, ds_key def lowercase_ ( self , A_ , A_=None ): '''simple docstring''' SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.find_config_node(A_ ) if config is None: return default return config.get(A_ , A_ ) def lowercase_ ( self , A_ , A_=False ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = self.config # find the config node of interest if it exists SCREAMING_SNAKE_CASE__ = ds_key_long.split('''.''' ) for node in nodes: SCREAMING_SNAKE_CASE__ = config SCREAMING_SNAKE_CASE__ = config.get(A_ ) if config is None: if must_exist: raise ValueError(f'''Can\'t find {ds_key_long} entry in the config: {self.config}''' ) else: return # if found remove it if parent_config is not None: parent_config.pop(A_ ) def lowercase_ ( self , A_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = self.get_value(A_ ) return False if value is None else bool(A_ ) def lowercase_ ( self , A_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = self.get_value(A_ ) return False if value is None else not bool(A_ ) def lowercase_ ( self ): '''simple docstring''' return self._stage == 2 def lowercase_ ( self ): '''simple docstring''' return self._stage == 3 def lowercase_ ( self ): '''simple docstring''' return self._offload class __snake_case : '''simple docstring''' def __init__( self , A_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = engine def lowercase_ ( self , A_ , **A_ ): '''simple docstring''' self.engine.backward(A_ , **A_ ) # Deepspeed's `engine.step` performs the following operations: # - gradient accumulation check # - gradient clipping # - optimizer step # - zero grad # - checking overflow # - lr_scheduler step (only if engine.lr_scheduler is not None) self.engine.step() # and this plugin overrides the above calls with no-ops when Accelerate runs under # Deepspeed, but allows normal functionality for non-Deepspeed cases thus enabling a simple # training loop that works transparently under many training regimes. class __snake_case ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self , A_ ): '''simple docstring''' super().__init__(A_ , device_placement=A_ , scaler=A_ ) SCREAMING_SNAKE_CASE__ = hasattr(self.optimizer , '''overflow''' ) def lowercase_ ( self , A_=None ): '''simple docstring''' pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed def lowercase_ ( self ): '''simple docstring''' pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed @property def lowercase_ ( self ): '''simple docstring''' if self.__has_overflow__: return self.optimizer.overflow return False class __snake_case ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self , A_ , A_ ): '''simple docstring''' super().__init__(A_ , A_ ) def lowercase_ ( self ): '''simple docstring''' pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed class __snake_case : '''simple docstring''' def __init__( self , A_ , A_=0.001 , A_=0 , **A_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = params SCREAMING_SNAKE_CASE__ = lr SCREAMING_SNAKE_CASE__ = weight_decay SCREAMING_SNAKE_CASE__ = kwargs class __snake_case : '''simple docstring''' def __init__( self , A_ , A_=None , A_=0 , **A_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = optimizer SCREAMING_SNAKE_CASE__ = total_num_steps SCREAMING_SNAKE_CASE__ = warmup_num_steps SCREAMING_SNAKE_CASE__ = kwargs
100
# this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under git won't be considered # since the output of this script is fed into Makefile commands it doesn't print a newline after the results import re import subprocess import sys lowerCamelCase : List[Any] = subprocess.check_output("git merge-base main HEAD".split()).decode("utf-8") lowerCamelCase : Tuple = ( subprocess.check_output(F"""git diff --diff-filter=d --name-only {fork_point_sha}""".split()).decode("utf-8").split() ) lowerCamelCase : Tuple = "|".join(sys.argv[1:]) lowerCamelCase : Any = re.compile(rF"""^({joined_dirs}).*?\.py$""") lowerCamelCase : List[str] = [x for x in modified_files if regex.match(x)] print(" ".join(relevant_modified_files), end="")
70
0
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import ( AutoProcessor, BertTokenizerFast, BlipImageProcessor, GPTaTokenizer, InstructBlipProcessor, PreTrainedTokenizerFast, ) @require_vision class __lowercase (unittest.TestCase ): """simple docstring""" def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = tempfile.mkdtemp() SCREAMING_SNAKE_CASE_ : Dict = BlipImageProcessor() SCREAMING_SNAKE_CASE_ : Optional[int] = GPTaTokenizer.from_pretrained('hf-internal-testing/tiny-random-GPT2Model' ) SCREAMING_SNAKE_CASE_ : str = BertTokenizerFast.from_pretrained('hf-internal-testing/tiny-random-bert' ) SCREAMING_SNAKE_CASE_ : Optional[int] = InstructBlipProcessor(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) def UpperCamelCase__ ( self , **lowerCAmelCase__ ): """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCAmelCase__ ).tokenizer def UpperCamelCase__ ( self , **lowerCAmelCase__ ): """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCAmelCase__ ).image_processor def UpperCamelCase__ ( self , **lowerCAmelCase__ ): """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCAmelCase__ ).qformer_tokenizer def UpperCamelCase__ ( self ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = [np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta )] SCREAMING_SNAKE_CASE_ : Optional[int] = [Image.fromarray(np.moveaxis(lowerCAmelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str = InstructBlipProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() , qformer_tokenizer=self.get_qformer_tokenizer() , ) processor.save_pretrained(self.tmpdirname ) SCREAMING_SNAKE_CASE_ : str = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) SCREAMING_SNAKE_CASE_ : int = self.get_image_processor(do_normalize=lowerCAmelCase__ , padding_value=1.0 ) SCREAMING_SNAKE_CASE_ : Any = InstructBlipProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=lowerCAmelCase__ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , lowerCAmelCase__ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCAmelCase__ ) self.assertIsInstance(processor.qformer_tokenizer , lowerCAmelCase__ ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] = self.get_image_processor() SCREAMING_SNAKE_CASE_ : List[str] = self.get_tokenizer() SCREAMING_SNAKE_CASE_ : Dict = self.get_qformer_tokenizer() SCREAMING_SNAKE_CASE_ : str = InstructBlipProcessor( tokenizer=lowerCAmelCase__ , image_processor=lowerCAmelCase__ , qformer_tokenizer=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : List[Any] = self.prepare_image_inputs() SCREAMING_SNAKE_CASE_ : List[str] = image_processor(lowerCAmelCase__ , return_tensors='np' ) SCREAMING_SNAKE_CASE_ : Optional[int] = processor(images=lowerCAmelCase__ , return_tensors='np' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.get_image_processor() SCREAMING_SNAKE_CASE_ : Tuple = self.get_tokenizer() SCREAMING_SNAKE_CASE_ : Optional[int] = self.get_qformer_tokenizer() SCREAMING_SNAKE_CASE_ : Any = InstructBlipProcessor( tokenizer=lowerCAmelCase__ , image_processor=lowerCAmelCase__ , qformer_tokenizer=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Tuple = 'lower newer' SCREAMING_SNAKE_CASE_ : Union[str, Any] = processor(text=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Dict = tokenizer(lowerCAmelCase__ , return_token_type_ids=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Tuple = qformer_tokenizer(lowerCAmelCase__ , return_token_type_ids=lowerCAmelCase__ ) for key in encoded_tokens.keys(): self.assertListEqual(encoded_tokens[key] , encoded_processor[key] ) for key in encoded_tokens_qformer.keys(): self.assertListEqual(encoded_tokens_qformer[key] , encoded_processor['qformer_' + key] ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.get_image_processor() SCREAMING_SNAKE_CASE_ : str = self.get_tokenizer() SCREAMING_SNAKE_CASE_ : List[Any] = self.get_qformer_tokenizer() SCREAMING_SNAKE_CASE_ : Tuple = InstructBlipProcessor( tokenizer=lowerCAmelCase__ , image_processor=lowerCAmelCase__ , qformer_tokenizer=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Tuple = 'lower newer' SCREAMING_SNAKE_CASE_ : int = self.prepare_image_inputs() SCREAMING_SNAKE_CASE_ : List[str] = processor(text=lowerCAmelCase__ , images=lowerCAmelCase__ ) self.assertListEqual( list(inputs.keys() ) , ['input_ids', 'attention_mask', 'qformer_input_ids', 'qformer_attention_mask', 'pixel_values'] , ) # test if it raises when no input is passed with pytest.raises(lowerCAmelCase__ ): processor() def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = self.get_image_processor() SCREAMING_SNAKE_CASE_ : str = self.get_tokenizer() SCREAMING_SNAKE_CASE_ : List[str] = self.get_qformer_tokenizer() SCREAMING_SNAKE_CASE_ : List[str] = InstructBlipProcessor( tokenizer=lowerCAmelCase__ , image_processor=lowerCAmelCase__ , qformer_tokenizer=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : List[str] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] SCREAMING_SNAKE_CASE_ : Union[str, Any] = processor.batch_decode(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Any = tokenizer.batch_decode(lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = self.get_image_processor() SCREAMING_SNAKE_CASE_ : str = self.get_tokenizer() SCREAMING_SNAKE_CASE_ : List[Any] = self.get_qformer_tokenizer() SCREAMING_SNAKE_CASE_ : Tuple = InstructBlipProcessor( tokenizer=lowerCAmelCase__ , image_processor=lowerCAmelCase__ , qformer_tokenizer=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : List[Any] = 'lower newer' SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.prepare_image_inputs() SCREAMING_SNAKE_CASE_ : Any = processor(text=lowerCAmelCase__ , images=lowerCAmelCase__ ) self.assertListEqual( list(inputs.keys() ) , ['input_ids', 'attention_mask', 'qformer_input_ids', 'qformer_attention_mask', 'pixel_values'] , )
101
import argparse import json import subprocess def _SCREAMING_SNAKE_CASE ( lowercase : Dict , lowercase : List[str] ): '''simple docstring''' lowerCamelCase_ = [] lowerCamelCase_ = ( f"""curl -H \"Accept: application/vnd.github+json\" -H \"Authorization: Bearer {token}\"""" ' https://api.github.com/repos/huggingface/transformers/actions/runners' ) lowerCamelCase_ = subprocess.run(lowercase , shell=lowercase , stdout=subprocess.PIPE ) lowerCamelCase_ = output.stdout.decode('utf-8' ) lowerCamelCase_ = json.loads(lowercase ) lowerCamelCase_ = status['runners'] for runner in runners: if runner["name"] in target_runners: if runner["status"] == "offline": offline_runners.append(lowercase ) # save the result so we can report them on Slack with open('offline_runners.txt' , 'w' ) as fp: fp.write(json.dumps(lowercase ) ) if len(lowercase ) > 0: lowerCamelCase_ = '\n'.join([x['name'] for x in offline_runners] ) raise ValueError(f"""The following runners are offline:\n{failed}""" ) if __name__ == "__main__": def _SCREAMING_SNAKE_CASE ( lowercase : List[str] ): '''simple docstring''' return values.split(',' ) lowerCamelCase : str = argparse.ArgumentParser() # Required parameters parser.add_argument( "--target_runners", default=None, type=list_str, required=True, help="Comma-separated list of runners to check status.", ) parser.add_argument( "--token", default=None, type=str, required=True, help="A token that has actions:read permission." ) lowerCamelCase : Optional[int] = parser.parse_args() get_runner_status(args.target_runners, args.token)
70
0
"""simple docstring""" def UpperCamelCase (SCREAMING_SNAKE_CASE ): UpperCamelCase : List[Any] = [int(SCREAMING_SNAKE_CASE ) for i in ip_va_address.split(""".""" ) if i.isdigit()] return len(SCREAMING_SNAKE_CASE ) == 4 and all(0 <= int(SCREAMING_SNAKE_CASE ) <= 254 for octet in octets ) if __name__ == "__main__": __magic_name__ : str = input().strip() __magic_name__ : Tuple = """valid""" if is_ip_va_address_valid(ip) else """invalid""" print(f'''{ip} is a {valid_or_invalid} IP v4 address.''')
102
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from timm import create_model from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import BitConfig, BitForImageClassification, BitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase : Optional[Any] = logging.get_logger(__name__) def _SCREAMING_SNAKE_CASE ( lowercase : List[Any] ): '''simple docstring''' lowerCamelCase_ = 'huggingface/label-files' lowerCamelCase_ = 'imagenet-1k-id2label.json' lowerCamelCase_ = json.load(open(hf_hub_download(lowercase , lowercase , repo_type='dataset' ) , 'r' ) ) lowerCamelCase_ = {int(lowercase ): v for k, v in idalabel.items()} lowerCamelCase_ = {v: k for k, v in idalabel.items()} lowerCamelCase_ = 'std_conv' if 'bit' in model_name else False # note that when using BiT as backbone for ViT-hybrid checkpoints, # one needs to additionally set config.layer_type = "bottleneck", config.stem_type = "same", # config.conv_layer = "std_conv_same" lowerCamelCase_ = BitConfig( conv_layer=lowercase , num_labels=10_00 , idalabel=lowercase , labelaid=lowercase , ) return config def _SCREAMING_SNAKE_CASE ( lowercase : Any ): '''simple docstring''' if "stem.conv" in name: lowerCamelCase_ = name.replace('stem.conv' , 'bit.embedder.convolution' ) if "blocks" in name: lowerCamelCase_ = name.replace('blocks' , 'layers' ) if "head.fc" in name: lowerCamelCase_ = name.replace('head.fc' , 'classifier.1' ) if name.startswith('norm' ): lowerCamelCase_ = 'bit.' + name if "bit" not in name and "classifier" not in name: lowerCamelCase_ = 'bit.encoder.' + name return name def _SCREAMING_SNAKE_CASE ( ): '''simple docstring''' lowerCamelCase_ = 'http://images.cocodataset.org/val2017/000000039769.jpg' lowerCamelCase_ = Image.open(requests.get(lowercase , stream=lowercase ).raw ) return im @torch.no_grad() def _SCREAMING_SNAKE_CASE ( lowercase : str , lowercase : int , lowercase : Union[str, Any]=False ): '''simple docstring''' lowerCamelCase_ = get_config(lowercase ) # load original model from timm lowerCamelCase_ = create_model(lowercase , pretrained=lowercase ) timm_model.eval() # load state_dict of original model lowerCamelCase_ = timm_model.state_dict() for key in state_dict.copy().keys(): lowerCamelCase_ = state_dict.pop(lowercase ) lowerCamelCase_ = val.squeeze() if 'head' in key else val # load HuggingFace model lowerCamelCase_ = BitForImageClassification(lowercase ) model.eval() model.load_state_dict(lowercase ) # create image processor lowerCamelCase_ = create_transform(**resolve_data_config({} , model=lowercase ) ) lowerCamelCase_ = transform.transforms lowerCamelCase_ = { 'bilinear': PILImageResampling.BILINEAR, 'bicubic': PILImageResampling.BICUBIC, 'nearest': PILImageResampling.NEAREST, } lowerCamelCase_ = BitImageProcessor( do_resize=lowercase , size={'shortest_edge': timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=lowercase , crop_size={'height': timm_transforms[1].size[0], 'width': timm_transforms[1].size[1]} , do_normalize=lowercase , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) lowerCamelCase_ = prepare_img() lowerCamelCase_ = transform(lowercase ).unsqueeze(0 ) lowerCamelCase_ = processor(lowercase , return_tensors='pt' ).pixel_values # verify pixel values assert torch.allclose(lowercase , lowercase ) # verify logits with torch.no_grad(): lowerCamelCase_ = model(lowercase ) lowerCamelCase_ = outputs.logits print('Logits:' , logits[0, :3] ) print('Predicted class:' , model.config.idalabel[logits.argmax(-1 ).item()] ) lowerCamelCase_ = timm_model(lowercase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(lowercase , outputs.logits , atol=1e-3 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: Path(lowercase ).mkdir(exist_ok=lowercase ) print(f"""Saving model {model_name} and processor to {pytorch_dump_folder_path}""" ) model.save_pretrained(lowercase ) processor.save_pretrained(lowercase ) if push_to_hub: print(f"""Pushing model {model_name} and processor to the hub""" ) model.push_to_hub(f"""ybelkada/{model_name}""" ) processor.push_to_hub(f"""ybelkada/{model_name}""" ) if __name__ == "__main__": lowerCamelCase : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="resnetv2_50x1_bitm", type=str, help="Name of the BiT timm model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to push the model to the hub.", ) lowerCamelCase : Optional[int] = parser.parse_args() convert_bit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
70
0
"""simple docstring""" import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers import CLIPModel, CLIPTokenizerFast from utils import get_device, get_timestamp, show_pil class UpperCAmelCase : def __init__( self : Dict , __lowerCamelCase : str = "cpu" , __lowerCamelCase : str = "openai/clip-vit-large-patch14" ): """simple docstring""" _snake_case = device _snake_case = CLIPTokenizerFast.from_pretrained(__lowerCamelCase ) _snake_case = [0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3] _snake_case = [0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1] _snake_case = torchvision.transforms.Normalize(self.image_mean , self.image_std ) _snake_case = torchvision.transforms.Resize(2_2_4 ) _snake_case = torchvision.transforms.CenterCrop(2_2_4 ) def __UpperCAmelCase ( self : int , __lowerCamelCase : Optional[Any] ): """simple docstring""" _snake_case = self.resize(__lowerCamelCase ) _snake_case = self.center_crop(__lowerCamelCase ) _snake_case = self.normalize(__lowerCamelCase ) return images def __call__( self : Optional[int] , __lowerCamelCase : List[Any]=None , __lowerCamelCase : int=None , **__lowerCamelCase : List[Any] ): """simple docstring""" _snake_case = self.tokenizer(text=__lowerCamelCase , **__lowerCamelCase ) _snake_case = self.preprocess_img(__lowerCamelCase ) _snake_case = {key: value.to(self.device ) for (key, value) in encoding.items()} return encoding class UpperCAmelCase ( nn.Module ): def __init__( self : Union[str, Any] , __lowerCamelCase : Union[str, Any]=1_0 , __lowerCamelCase : Optional[Any]=0.0_1 , __lowerCamelCase : Dict=None , __lowerCamelCase : int=None , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : int=None , __lowerCamelCase : Tuple=None , __lowerCamelCase : List[Any]=False , __lowerCamelCase : int=True , __lowerCamelCase : int="image" , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Optional[Any]=False , __lowerCamelCase : Tuple=False , __lowerCamelCase : Optional[int]=False , ): """simple docstring""" super().__init__() _snake_case = None _snake_case = device if device else get_device() if vqgan: _snake_case = vqgan else: _snake_case = load_vqgan(self.device , conf_path=__lowerCamelCase , ckpt_path=__lowerCamelCase ) self.vqgan.eval() if clip: _snake_case = clip else: _snake_case = CLIPModel.from_pretrained('''openai/clip-vit-base-patch32''' ) self.clip.to(self.device ) _snake_case = ProcessorGradientFlow(device=self.device ) _snake_case = iterations _snake_case = lr _snake_case = log _snake_case = make_grid _snake_case = return_val _snake_case = quantize _snake_case = self.vqgan.decoder.z_shape def __UpperCAmelCase ( self : int , __lowerCamelCase : Optional[int]=None , __lowerCamelCase : Any=None , __lowerCamelCase : Optional[Any]=5 , __lowerCamelCase : Tuple=True ): """simple docstring""" _snake_case = [] if output_path is None: _snake_case = '''./animation.gif''' if input_path is None: _snake_case = self.save_path _snake_case = sorted(glob(input_path + '''/*''' ) ) if not len(__lowerCamelCase ): raise ValueError( '''No images found in save path, aborting (did you pass save_intermediate=True to the generate''' ''' function?)''' ) if len(__lowerCamelCase ) == 1: print('''Only one image found in save path, (did you pass save_intermediate=True to the generate function?)''' ) _snake_case = total_duration / len(__lowerCamelCase ) _snake_case = [frame_duration] * len(__lowerCamelCase ) if extend_frames: _snake_case = 1.5 _snake_case = 3 for file_name in paths: if file_name.endswith('''.png''' ): images.append(imageio.imread(__lowerCamelCase ) ) imageio.mimsave(__lowerCamelCase , __lowerCamelCase , duration=__lowerCamelCase ) print(f"""gif saved to {output_path}""" ) def __UpperCAmelCase ( self : Optional[int] , __lowerCamelCase : List[str]=None , __lowerCamelCase : Any=None ): """simple docstring""" if not (path or img): raise ValueError('''Input either path or tensor''' ) if img is not None: raise NotImplementedError _snake_case = preprocess(Image.open(__lowerCamelCase ) , target_image_size=2_5_6 ).to(self.device ) _snake_case = preprocess_vqgan(__lowerCamelCase ) _snake_case , *_snake_case = self.vqgan.encode(__lowerCamelCase ) return z def __UpperCAmelCase ( self : List[Any] , __lowerCamelCase : Dict ): """simple docstring""" _snake_case = self.latent.detach().requires_grad_() _snake_case = base_latent + transform_vector if self.quantize: _snake_case , *_snake_case = self.vqgan.quantize(__lowerCamelCase ) else: _snake_case = trans_latent return self.vqgan.decode(__lowerCamelCase ) def __UpperCAmelCase ( self : Optional[Any] , __lowerCamelCase : Tuple , __lowerCamelCase : List[Any] , __lowerCamelCase : Any=None ): """simple docstring""" _snake_case = self.clip_preprocessor(text=__lowerCamelCase , images=__lowerCamelCase , return_tensors='''pt''' , padding=__lowerCamelCase ) _snake_case = self.clip(**__lowerCamelCase ) _snake_case = clip_outputs.logits_per_image if weights is not None: _snake_case = similarity_logits * weights return similarity_logits.sum() def __UpperCAmelCase ( self : Optional[int] , __lowerCamelCase : str , __lowerCamelCase : List[Any] , __lowerCamelCase : Tuple ): """simple docstring""" _snake_case = self._get_clip_similarity(pos_prompts['''prompts'''] , __lowerCamelCase , weights=(1 / pos_prompts['''weights''']) ) if neg_prompts: _snake_case = self._get_clip_similarity(neg_prompts['''prompts'''] , __lowerCamelCase , weights=neg_prompts['''weights'''] ) else: _snake_case = torch.tensor([1] , device=self.device ) _snake_case = -torch.log(__lowerCamelCase ) + torch.log(__lowerCamelCase ) return loss def __UpperCAmelCase ( self : Optional[int] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Dict ): """simple docstring""" _snake_case = torch.randn_like(self.latent , requires_grad=__lowerCamelCase , device=self.device ) _snake_case = torch.optim.Adam([vector] , lr=self.lr ) for i in range(self.iterations ): optim.zero_grad() _snake_case = self._add_vector(__lowerCamelCase ) _snake_case = loop_post_process(__lowerCamelCase ) _snake_case = self._get_CLIP_loss(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) print('''CLIP loss''' , __lowerCamelCase ) if self.log: wandb.log({'''CLIP Loss''': clip_loss} ) clip_loss.backward(retain_graph=__lowerCamelCase ) optim.step() if self.return_val == "image": yield custom_to_pil(transformed_img[0] ) else: yield vector def __UpperCAmelCase ( self : Optional[Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Tuple , __lowerCamelCase : List[str] ): """simple docstring""" wandb.init(reinit=__lowerCamelCase , project='''face-editor''' ) wandb.config.update({'''Positive Prompts''': positive_prompts} ) wandb.config.update({'''Negative Prompts''': negative_prompts} ) wandb.config.update({'''lr''': self.lr, '''iterations''': self.iterations} ) if image_path: _snake_case = Image.open(__lowerCamelCase ) _snake_case = image.resize((2_5_6, 2_5_6) ) wandb.log('''Original Image''' , wandb.Image(__lowerCamelCase ) ) def __UpperCAmelCase ( self : Dict , __lowerCamelCase : Any ): """simple docstring""" if not prompts: return [] _snake_case = [] _snake_case = [] if isinstance(__lowerCamelCase , __lowerCamelCase ): _snake_case = [prompt.strip() for prompt in prompts.split('''|''' )] for prompt in prompts: if isinstance(__lowerCamelCase , (tuple, list) ): _snake_case = prompt[0] _snake_case = float(prompt[1] ) elif ":" in prompt: _snake_case , _snake_case = prompt.split(''':''' ) _snake_case = float(__lowerCamelCase ) else: _snake_case = prompt _snake_case = 1.0 processed_prompts.append(__lowerCamelCase ) weights.append(__lowerCamelCase ) return { "prompts": processed_prompts, "weights": torch.tensor(__lowerCamelCase , device=self.device ), } def __UpperCAmelCase ( self : int , __lowerCamelCase : Any , __lowerCamelCase : Tuple=None , __lowerCamelCase : List[Any]=None , __lowerCamelCase : Dict=True , __lowerCamelCase : int=False , __lowerCamelCase : Optional[int]=True , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Optional[int]=None , ): """simple docstring""" if image_path: _snake_case = self._get_latent(__lowerCamelCase ) else: _snake_case = torch.randn(self.latent_dim , device=self.device ) if self.log: self._init_logging(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) assert pos_prompts, "You must provide at least one positive prompt." _snake_case = self.process_prompts(__lowerCamelCase ) _snake_case = self.process_prompts(__lowerCamelCase ) if save_final and save_path is None: _snake_case = os.path.join('''./outputs/''' , '''_'''.join(pos_prompts['''prompts'''] ) ) if not os.path.exists(__lowerCamelCase ): os.makedirs(__lowerCamelCase ) else: _snake_case = save_path + '''_''' + get_timestamp() os.makedirs(__lowerCamelCase ) _snake_case = save_path _snake_case = self.vqgan.decode(self.latent )[0] if show_intermediate: print('''Original Image''' ) show_pil(custom_to_pil(__lowerCamelCase ) ) _snake_case = loop_post_process(__lowerCamelCase ) for iter, transformed_img in enumerate(self._optimize_CLIP(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) ): if show_intermediate: show_pil(__lowerCamelCase ) if save_intermediate: transformed_img.save(os.path.join(self.save_path , f"""iter_{iter:03d}.png""" ) ) if self.log: wandb.log({'''Image''': wandb.Image(__lowerCamelCase )} ) if show_final: show_pil(__lowerCamelCase ) if save_final: transformed_img.save(os.path.join(self.save_path , f"""iter_{iter:03d}_final.png""" ) )
103
from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class A: '''simple docstring''' def __init__( self : List[Any] , A_ : Dict , A_ : Union[str, Any]=13 , A_ : List[Any]=30 , A_ : Optional[Any]=2 , A_ : List[str]=3 , A_ : List[str]=True , A_ : Dict=True , A_ : List[Any]=32 , A_ : Any=2 , A_ : Any=4 , A_ : Optional[int]=37 , A_ : Dict="gelu" , A_ : List[Any]=0.1 , A_ : Optional[int]=0.1 , A_ : Union[str, Any]=10 , A_ : Optional[Any]=0.02 , A_ : List[Any]=3 , A_ : str=None , ) -> str: """simple docstring""" lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = image_size lowerCamelCase_ = patch_size lowerCamelCase_ = num_channels lowerCamelCase_ = is_training lowerCamelCase_ = use_labels lowerCamelCase_ = hidden_size lowerCamelCase_ = num_hidden_layers lowerCamelCase_ = num_attention_heads lowerCamelCase_ = intermediate_size lowerCamelCase_ = hidden_act lowerCamelCase_ = hidden_dropout_prob lowerCamelCase_ = attention_probs_dropout_prob lowerCamelCase_ = type_sequence_label_size lowerCamelCase_ = initializer_range lowerCamelCase_ = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) lowerCamelCase_ = (image_size // patch_size) ** 2 lowerCamelCase_ = num_patches + 1 def a__ ( self : List[str] ) -> Dict: """simple docstring""" lowerCamelCase_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase_ = None if self.use_labels: lowerCamelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase_ = self.get_config() return config, pixel_values, labels def a__ ( self : List[Any] ) -> Any: """simple docstring""" return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=A_ , initializer_range=self.initializer_range , ) def a__ ( self : Any , A_ : int , A_ : int , A_ : int ) -> List[Any]: """simple docstring""" lowerCamelCase_ = TFViTModel(config=A_ ) lowerCamelCase_ = model(A_ , training=A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # Test with an image with different size than the one specified in config. lowerCamelCase_ = self.image_size // 2 lowerCamelCase_ = pixel_values[:, :, :image_size, :image_size] lowerCamelCase_ = model(A_ , interpolate_pos_encoding=A_ , training=A_ ) lowerCamelCase_ = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, seq_length, self.hidden_size) ) def a__ ( self : List[Any] , A_ : List[Any] , A_ : Any , A_ : Any ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.type_sequence_label_size lowerCamelCase_ = TFViTForImageClassification(A_ ) lowerCamelCase_ = model(A_ , labels=A_ , training=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # Test with an image with different size than the one specified in config. lowerCamelCase_ = self.image_size // 2 lowerCamelCase_ = pixel_values[:, :, :image_size, :image_size] lowerCamelCase_ = model(A_ , interpolate_pos_encoding=A_ , training=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowerCamelCase_ = 1 lowerCamelCase_ = TFViTForImageClassification(A_ ) lowerCamelCase_ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCamelCase_ = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def a__ ( self : List[Any] ) -> Any: """simple docstring""" lowerCamelCase_ = self.prepare_config_and_inputs() lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = config_and_inputs lowerCamelCase_ = {'pixel_values': pixel_values} return config, inputs_dict @require_tf class A( UpperCamelCase , UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () UpperCamelCase = ( {'''feature-extraction''': TFViTModel, '''image-classification''': TFViTForImageClassification} if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def a__ ( self : int ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = TFViTModelTester(self ) lowerCamelCase_ = ConfigTester(self , config_class=A_ , has_text_modality=A_ , hidden_size=37 ) def a__ ( self : int ) -> List[str]: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='ViT does not use inputs_embeds' ) def a__ ( self : List[str] ) -> Tuple: """simple docstring""" pass @unittest.skip(reason='ViT does not use inputs_embeds' ) def a__ ( self : int ) -> Optional[Any]: """simple docstring""" pass def a__ ( self : str ) -> List[str]: """simple docstring""" lowerCamelCase_ , lowerCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase_ = model_class(A_ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) lowerCamelCase_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(A_ , tf.keras.layers.Layer ) ) def a__ ( self : List[Any] ) -> Dict: """simple docstring""" lowerCamelCase_ , lowerCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase_ = model_class(A_ ) lowerCamelCase_ = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase_ = [*signature.parameters.keys()] lowerCamelCase_ = ['pixel_values'] self.assertListEqual(arg_names[:1] , A_ ) def a__ ( self : Tuple ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def a__ ( self : Tuple ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A_ ) @slow def a__ ( self : Any ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = TFViTModel.from_pretrained('google/vit-base-patch16-224' ) self.assertIsNotNone(A_ ) def _SCREAMING_SNAKE_CASE ( ): '''simple docstring''' lowerCamelCase_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class A( unittest.TestCase ): '''simple docstring''' @cached_property def a__ ( self : int ) -> Tuple: """simple docstring""" return ViTImageProcessor.from_pretrained('google/vit-base-patch16-224' ) if is_vision_available() else None @slow def a__ ( self : List[str] ) -> str: """simple docstring""" lowerCamelCase_ = TFViTForImageClassification.from_pretrained('google/vit-base-patch16-224' ) lowerCamelCase_ = self.default_image_processor lowerCamelCase_ = prepare_img() lowerCamelCase_ = image_processor(images=A_ , return_tensors='tf' ) # forward pass lowerCamelCase_ = model(**A_ ) # verify the logits lowerCamelCase_ = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , A_ ) lowerCamelCase_ = tf.constant([-0.2744, 0.8215, -0.0836] ) tf.debugging.assert_near(outputs.logits[0, :3] , A_ , atol=1E-4 )
70
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/config.json""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/config.json""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/config.json""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/config.json""" ), """google/realm-orqa-nq-openqa""": """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/config.json""", """google/realm-orqa-nq-reader""": """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/config.json""", """google/realm-orqa-wq-openqa""": """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/config.json""", """google/realm-orqa-wq-reader""": """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/config.json""", # See all REALM models at https://huggingface.co/models?filter=realm } class UpperCamelCase__ ( _lowerCAmelCase ): """simple docstring""" A__ : List[Any] = "realm" def __init__( self , SCREAMING_SNAKE_CASE__=30522 , SCREAMING_SNAKE_CASE__=768 , SCREAMING_SNAKE_CASE__=128 , SCREAMING_SNAKE_CASE__=12 , SCREAMING_SNAKE_CASE__=12 , SCREAMING_SNAKE_CASE__=8 , SCREAMING_SNAKE_CASE__=3072 , SCREAMING_SNAKE_CASE__="gelu_new" , SCREAMING_SNAKE_CASE__=0.1 , SCREAMING_SNAKE_CASE__=0.1 , SCREAMING_SNAKE_CASE__=512 , SCREAMING_SNAKE_CASE__=2 , SCREAMING_SNAKE_CASE__=0.0_2 , SCREAMING_SNAKE_CASE__=1e-12 , SCREAMING_SNAKE_CASE__=256 , SCREAMING_SNAKE_CASE__=10 , SCREAMING_SNAKE_CASE__=1e-3 , SCREAMING_SNAKE_CASE__=5 , SCREAMING_SNAKE_CASE__=320 , SCREAMING_SNAKE_CASE__=13353718 , SCREAMING_SNAKE_CASE__=5000 , SCREAMING_SNAKE_CASE__=1 , SCREAMING_SNAKE_CASE__=0 , SCREAMING_SNAKE_CASE__=2 , **SCREAMING_SNAKE_CASE__ , ) -> Any: super().__init__(pad_token_id=SCREAMING_SNAKE_CASE__ , bos_token_id=SCREAMING_SNAKE_CASE__ , eos_token_id=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) # Common config A__ = vocab_size A__ = max_position_embeddings A__ = hidden_size A__ = retriever_proj_size A__ = num_hidden_layers A__ = num_attention_heads A__ = num_candidates A__ = intermediate_size A__ = hidden_act A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = initializer_range A__ = type_vocab_size A__ = layer_norm_eps # Reader config A__ = span_hidden_size A__ = max_span_width A__ = reader_layer_norm_eps A__ = reader_beam_size A__ = reader_seq_len # Retrieval config A__ = num_block_records A__ = searcher_beam_size
104
import itertools import random import unittest import numpy as np from transformers import is_speech_available from transformers.testing_utils import require_torch, require_torchaudio from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import SpeechaTextFeatureExtractor lowerCamelCase : Any = random.Random() def _SCREAMING_SNAKE_CASE ( lowercase : Union[str, Any] , lowercase : int=1.0 , lowercase : List[str]=None , lowercase : str=None ): '''simple docstring''' if rng is None: lowerCamelCase_ = global_rng lowerCamelCase_ = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class A( unittest.TestCase ): '''simple docstring''' def __init__( self : List[Any] , A_ : Dict , A_ : int=7 , A_ : str=400 , A_ : Dict=2000 , A_ : List[Any]=24 , A_ : List[Any]=24 , A_ : int=0.0 , A_ : Dict=16000 , A_ : List[Any]=True , A_ : str=True , ) -> Dict: """simple docstring""" lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = min_seq_length lowerCamelCase_ = max_seq_length lowerCamelCase_ = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) lowerCamelCase_ = feature_size lowerCamelCase_ = num_mel_bins lowerCamelCase_ = padding_value lowerCamelCase_ = sampling_rate lowerCamelCase_ = return_attention_mask lowerCamelCase_ = do_normalize def a__ ( self : List[str] ) -> Optional[int]: """simple docstring""" return { "feature_size": self.feature_size, "num_mel_bins": self.num_mel_bins, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def a__ ( self : List[Any] , A_ : str=False , A_ : Union[str, Any]=False ) -> str: """simple docstring""" def _flatten(A_ : List[Any] ): return list(itertools.chain(*A_ ) ) if equal_length: lowerCamelCase_ = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size lowerCamelCase_ = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: lowerCamelCase_ = [np.asarray(A_ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class A( UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = SpeechaTextFeatureExtractor if is_speech_available() else None def a__ ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = SpeechaTextFeatureExtractionTester(self ) def a__ ( self : str , A_ : Dict ) -> Dict: """simple docstring""" self.assertTrue(np.all(np.mean(A_ , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(A_ , axis=0 ) - 1 ) < 1E-3 ) ) def a__ ( self : int ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = [np.asarray(A_ ) for speech_input in speech_inputs] # Test feature size lowerCamelCase_ = feature_extractor(A_ , padding=A_ , return_tensors='np' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.feature_size ) # Test not batched input lowerCamelCase_ = feature_extractor(speech_inputs[0] , return_tensors='np' ).input_features lowerCamelCase_ = feature_extractor(np_speech_inputs[0] , return_tensors='np' ).input_features self.assertTrue(np.allclose(A_ , A_ , atol=1E-3 ) ) # Test batched lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. lowerCamelCase_ = [floats_list((1, x) )[0] for x in (800, 800, 800)] lowerCamelCase_ = np.asarray(A_ ) lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1E-3 ) ) def a__ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = ['longest', 'max_length', 'do_not_pad'] lowerCamelCase_ = [None, 16, None] for max_length, padding in zip(A_ , A_ ): lowerCamelCase_ = feature_extractor( A_ , padding=A_ , max_length=A_ , return_attention_mask=A_ ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = [np.sum(A_ ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def a__ ( self : List[Any] ) -> Any: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = ['longest', 'max_length', 'do_not_pad'] lowerCamelCase_ = [None, 16, None] for max_length, padding in zip(A_ , A_ ): lowerCamelCase_ = feature_extractor( A_ , max_length=A_ , padding=A_ , return_tensors='np' , return_attention_mask=A_ ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = [np.sum(A_ ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self.assertTrue(input_features[0][fbank_feat_lengths[0] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self.assertTrue(input_features[0][fbank_feat_lengths[1] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def a__ ( self : List[str] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = feature_extractor( A_ , padding='max_length' , max_length=4 , truncation=A_ , return_tensors='np' , return_attention_mask=A_ , ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1] ) self._check_zero_mean_unit_variance(input_features[2] ) def a__ ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = feature_extractor( A_ , padding='longest' , max_length=4 , truncation=A_ , return_tensors='np' , return_attention_mask=A_ , ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 4, 24) ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = feature_extractor( A_ , padding='longest' , max_length=16 , truncation=A_ , return_tensors='np' , return_attention_mask=A_ , ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 6, 24) ) def a__ ( self : List[Any] ) -> List[str]: """simple docstring""" import torch lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = np.random.rand(100 , 32 ).astype(np.floataa ) lowerCamelCase_ = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: lowerCamelCase_ = feature_extractor.pad([{'input_features': inputs}] , return_tensors='np' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) lowerCamelCase_ = feature_extractor.pad([{'input_features': inputs}] , return_tensors='pt' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def a__ ( self : List[str] , A_ : Union[str, Any] ) -> List[Any]: """simple docstring""" from datasets import load_dataset lowerCamelCase_ = load_dataset('hf-internal-testing/librispeech_asr_dummy' , 'clean' , split='validation' ) # automatic decoding with librispeech lowerCamelCase_ = ds.sort('id' ).select(range(A_ ) )[:num_samples]['audio'] return [x["array"] for x in speech_samples] def a__ ( self : str ) -> Tuple: """simple docstring""" lowerCamelCase_ = np.array([ -1.5745, -1.7713, -1.7020, -1.6069, -1.2250, -1.1105, -0.9072, -0.8241, -1.2310, -0.8098, -0.3320, -0.4101, -0.7985, -0.4996, -0.8213, -0.9128, -1.0420, -1.1286, -1.0440, -0.7999, -0.8405, -1.2275, -1.5443, -1.4625, ] ) # fmt: on lowerCamelCase_ = self._load_datasamples(1 ) lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = feature_extractor(A_ , return_tensors='pt' ).input_features self.assertEquals(input_features.shape , (1, 584, 24) ) self.assertTrue(np.allclose(input_features[0, 0, :30] , A_ , atol=1E-4 ) )
70
0
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BertTokenizer, BlipImageProcessor, BlipProcessor, PreTrainedTokenizerFast @require_vision class lowerCAmelCase_ ( unittest.TestCase ): def snake_case ( self ): SCREAMING_SNAKE_CASE_ : int = tempfile.mkdtemp() SCREAMING_SNAKE_CASE_ : Union[str, Any] = BlipImageProcessor() SCREAMING_SNAKE_CASE_ : List[Any] = BertTokenizer.from_pretrained('hf-internal-testing/tiny-random-BertModel' ) SCREAMING_SNAKE_CASE_ : Any = BlipProcessor(snake_case__ ,snake_case__ ) processor.save_pretrained(self.tmpdirname ) def snake_case ( self ,**snake_case__ ): return AutoProcessor.from_pretrained(self.tmpdirname ,**snake_case__ ).tokenizer def snake_case ( self ,**snake_case__ ): return AutoProcessor.from_pretrained(self.tmpdirname ,**snake_case__ ).image_processor def snake_case ( self ): shutil.rmtree(self.tmpdirname ) def snake_case ( self ): SCREAMING_SNAKE_CASE_ : Tuple = [np.random.randint(255 ,size=(3, 30, 400) ,dtype=np.uinta )] SCREAMING_SNAKE_CASE_ : Union[str, Any] = [Image.fromarray(np.moveaxis(snake_case__ ,0 ,-1 ) ) for x in image_inputs] return image_inputs def snake_case ( self ): SCREAMING_SNAKE_CASE_ : str = BlipProcessor(tokenizer=self.get_tokenizer() ,image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.get_tokenizer(bos_token='(BOS)' ,eos_token='(EOS)' ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.get_image_processor(do_normalize=snake_case__ ,padding_value=1.0 ) SCREAMING_SNAKE_CASE_ : Any = BlipProcessor.from_pretrained( self.tmpdirname ,bos_token='(BOS)' ,eos_token='(EOS)' ,do_normalize=snake_case__ ,padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() ,tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer ,snake_case__ ) self.assertEqual(processor.image_processor.to_json_string() ,image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor ,snake_case__ ) def snake_case ( self ): SCREAMING_SNAKE_CASE_ : int = self.get_image_processor() SCREAMING_SNAKE_CASE_ : int = self.get_tokenizer() SCREAMING_SNAKE_CASE_ : List[Any] = BlipProcessor(tokenizer=snake_case__ ,image_processor=snake_case__ ) SCREAMING_SNAKE_CASE_ : Tuple = self.prepare_image_inputs() SCREAMING_SNAKE_CASE_ : Optional[int] = image_processor(snake_case__ ,return_tensors='np' ) SCREAMING_SNAKE_CASE_ : Dict = processor(images=snake_case__ ,return_tensors='np' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1E-2 ) def snake_case ( self ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.get_image_processor() SCREAMING_SNAKE_CASE_ : List[Any] = self.get_tokenizer() SCREAMING_SNAKE_CASE_ : Any = BlipProcessor(tokenizer=snake_case__ ,image_processor=snake_case__ ) SCREAMING_SNAKE_CASE_ : Optional[Any] = 'lower newer' SCREAMING_SNAKE_CASE_ : str = processor(text=snake_case__ ) SCREAMING_SNAKE_CASE_ : List[Any] = tokenizer(snake_case__ ,return_token_type_ids=snake_case__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] ,encoded_processor[key] ) def snake_case ( self ): SCREAMING_SNAKE_CASE_ : Optional[int] = self.get_image_processor() SCREAMING_SNAKE_CASE_ : Optional[int] = self.get_tokenizer() SCREAMING_SNAKE_CASE_ : Optional[int] = BlipProcessor(tokenizer=snake_case__ ,image_processor=snake_case__ ) SCREAMING_SNAKE_CASE_ : Tuple = 'lower newer' SCREAMING_SNAKE_CASE_ : Tuple = self.prepare_image_inputs() SCREAMING_SNAKE_CASE_ : List[Any] = processor(text=snake_case__ ,images=snake_case__ ) self.assertListEqual(list(inputs.keys() ) ,['pixel_values', 'input_ids', 'attention_mask'] ) # test if it raises when no input is passed with pytest.raises(snake_case__ ): processor() def snake_case ( self ): SCREAMING_SNAKE_CASE_ : Any = self.get_image_processor() SCREAMING_SNAKE_CASE_ : str = self.get_tokenizer() SCREAMING_SNAKE_CASE_ : Optional[Any] = BlipProcessor(tokenizer=snake_case__ ,image_processor=snake_case__ ) SCREAMING_SNAKE_CASE_ : Optional[int] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] SCREAMING_SNAKE_CASE_ : str = processor.batch_decode(snake_case__ ) SCREAMING_SNAKE_CASE_ : Any = tokenizer.batch_decode(snake_case__ ) self.assertListEqual(snake_case__ ,snake_case__ ) def snake_case ( self ): SCREAMING_SNAKE_CASE_ : List[str] = self.get_image_processor() SCREAMING_SNAKE_CASE_ : List[Any] = self.get_tokenizer() SCREAMING_SNAKE_CASE_ : Dict = BlipProcessor(tokenizer=snake_case__ ,image_processor=snake_case__ ) SCREAMING_SNAKE_CASE_ : Optional[int] = 'lower newer' SCREAMING_SNAKE_CASE_ : List[Any] = self.prepare_image_inputs() SCREAMING_SNAKE_CASE_ : Optional[int] = processor(text=snake_case__ ,images=snake_case__ ) # For now the processor supports only ['pixel_values', 'input_ids', 'attention_mask'] self.assertListEqual(list(inputs.keys() ) ,['pixel_values', 'input_ids', 'attention_mask'] )
105
import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class A( UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = TransfoXLTokenizer UpperCamelCase = False UpperCamelCase = False def a__ ( self : Optional[Any] ) -> int: """simple docstring""" super().setUp() lowerCamelCase_ = [ '<unk>', '[CLS]', '[SEP]', 'want', 'unwanted', 'wa', 'un', 'running', ',', 'low', 'l', ] lowerCamelCase_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def a__ ( self : Optional[Any] , **A_ : Tuple ) -> Any: """simple docstring""" lowerCamelCase_ = True return TransfoXLTokenizer.from_pretrained(self.tmpdirname , **A_ ) def a__ ( self : List[str] , A_ : Dict ) -> Any: """simple docstring""" lowerCamelCase_ = '<unk> UNwanted , running' lowerCamelCase_ = '<unk> unwanted, running' return input_text, output_text def a__ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = TransfoXLTokenizer(vocab_file=self.vocab_file , lower_case=A_ ) lowerCamelCase_ = tokenizer.tokenize('<unk> UNwanted , running' ) self.assertListEqual(A_ , ['<unk>', 'unwanted', ',', 'running'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A_ ) , [0, 4, 8, 7] ) def a__ ( self : Any ) -> str: """simple docstring""" lowerCamelCase_ = TransfoXLTokenizer(lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['hello', '!', 'how', 'are', 'you', '?'] ) def a__ ( self : int ) -> Dict: """simple docstring""" lowerCamelCase_ = TransfoXLTokenizer(lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?'] ) def a__ ( self : List[Any] ) -> Tuple: """simple docstring""" lowerCamelCase_ = TransfoXLTokenizer(lower_case=A_ ) lowerCamelCase_ = 'Hello (bracket) and side-scrolled [and] Henry\'s $5,000 with 3.34 m. What\'s up!?' lowerCamelCase_ = [ 'Hello', '(', 'bracket', ')', 'and', 'side', '@-@', 'scrolled', '[', 'and', ']', 'Henry', '\'s', '$', '5', '@,@', '000', 'with', '3', '@.@', '34', 'm', '.', 'What', '\'s', 'up', '!', '?', ] self.assertListEqual(tokenizer.tokenize(A_ ) , A_ ) self.assertEqual(tokenizer.convert_tokens_to_string(A_ ) , A_ ) def a__ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = len(A_ ) tokenizer.add_tokens(['new1', 'new2'] ) tokenizer.move_added_token('new1' , 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(A_ ) , original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode('new1' ) , [1] ) self.assertEqual(tokenizer.decode([1] ) , 'new1' )
70
0
import math def lowerCamelCase_ ( lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : List[str] ) -> List[Any]: '''simple docstring''' if 0 not in (x, y): # We use the relation x^y = y*log10(x), where 10 is the base. return y * math.logaa(lowerCAmelCase__ ) else: if x == 0: # 0 raised to any number is 0 return 0 elif y == 0: return 1 # any number raised to 0 is 1 raise AssertionError('This should never happen' ) if __name__ == "__main__": # Main function # Read two numbers from input and typecast them to int using map function. # Here x is the base and y is the power. __snake_case :Optional[int] ='Enter the base and the power separated by a comma: ' __snake_case , __snake_case :str =map(int, input(prompt).split(',')) __snake_case , __snake_case :Optional[int] =map(int, input(prompt).split(',')) # We find the log of each number, using the function res(), which takes two # arguments. __snake_case :int =res(xa, ya) __snake_case :Any =res(xa, ya) # We check for the largest number if resa > resa: print('Largest number is', xa, '^', ya) elif resa > resa: print('Largest number is', xa, '^', ya) else: print('Both are equal')
106
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetrImageProcessor class A( unittest.TestCase ): '''simple docstring''' def __init__( self : Optional[Any] , A_ : Dict , A_ : int=7 , A_ : Any=3 , A_ : List[str]=30 , A_ : Union[str, Any]=400 , A_ : List[str]=True , A_ : int=None , A_ : Any=True , A_ : str=1 / 255 , A_ : int=True , A_ : List[Any]=[0.5, 0.5, 0.5] , A_ : Union[str, Any]=[0.5, 0.5, 0.5] , A_ : Union[str, Any]=True , ) -> List[str]: """simple docstring""" lowerCamelCase_ = size if size is not None else {'shortest_edge': 18, 'longest_edge': 1333} lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = num_channels lowerCamelCase_ = min_resolution lowerCamelCase_ = max_resolution lowerCamelCase_ = do_resize lowerCamelCase_ = size lowerCamelCase_ = do_rescale lowerCamelCase_ = rescale_factor lowerCamelCase_ = do_normalize lowerCamelCase_ = image_mean lowerCamelCase_ = image_std lowerCamelCase_ = do_pad def a__ ( self : Tuple ) -> Dict: """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_pad": self.do_pad, } def a__ ( self : Union[str, Any] , A_ : Dict , A_ : Any=False ) -> Union[str, Any]: """simple docstring""" if not batched: lowerCamelCase_ = image_inputs[0] if isinstance(A_ , Image.Image ): lowerCamelCase_ , lowerCamelCase_ = image.size else: lowerCamelCase_ , lowerCamelCase_ = image.shape[1], image.shape[2] if w < h: lowerCamelCase_ = int(self.size['shortest_edge'] * h / w ) lowerCamelCase_ = self.size['shortest_edge'] elif w > h: lowerCamelCase_ = self.size['shortest_edge'] lowerCamelCase_ = int(self.size['shortest_edge'] * w / h ) else: lowerCamelCase_ = self.size['shortest_edge'] lowerCamelCase_ = self.size['shortest_edge'] else: lowerCamelCase_ = [] for image in image_inputs: lowerCamelCase_ , lowerCamelCase_ = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) lowerCamelCase_ = max(A_ , key=lambda A_ : item[0] )[0] lowerCamelCase_ = max(A_ , key=lambda A_ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class A( UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = DetrImageProcessor if is_vision_available() else None def a__ ( self : List[Any] ) -> str: """simple docstring""" lowerCamelCase_ = DetrImageProcessingTester(self ) @property def a__ ( self : List[str] ) -> List[Any]: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def a__ ( self : Dict ) -> int: """simple docstring""" lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , 'image_mean' ) ) self.assertTrue(hasattr(A_ , 'image_std' ) ) self.assertTrue(hasattr(A_ , 'do_normalize' ) ) self.assertTrue(hasattr(A_ , 'do_rescale' ) ) self.assertTrue(hasattr(A_ , 'rescale_factor' ) ) self.assertTrue(hasattr(A_ , 'do_resize' ) ) self.assertTrue(hasattr(A_ , 'size' ) ) self.assertTrue(hasattr(A_ , 'do_pad' ) ) def a__ ( self : List[str] ) -> str: """simple docstring""" lowerCamelCase_ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 18, 'longest_edge': 1333} ) self.assertEqual(image_processor.do_pad , A_ ) lowerCamelCase_ = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A_ ) self.assertEqual(image_processor.size , {'shortest_edge': 42, 'longest_edge': 84} ) self.assertEqual(image_processor.do_pad , A_ ) def a__ ( self : Dict ) -> Any: """simple docstring""" pass def a__ ( self : Tuple ) -> Tuple: """simple docstring""" lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCamelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ ) for image in image_inputs: self.assertIsInstance(A_ , Image.Image ) # Test not batched input lowerCamelCase_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) lowerCamelCase_ = image_processing(A_ , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def a__ ( self : List[str] ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCamelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , np.ndarray ) # Test not batched input lowerCamelCase_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase_ = image_processing(A_ , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def a__ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCamelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test not batched input lowerCamelCase_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase_ = image_processing(A_ , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def a__ ( self : Tuple ) -> List[Any]: """simple docstring""" lowerCamelCase_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_annotations.txt' , 'r' ) as f: lowerCamelCase_ = json.loads(f.read() ) lowerCamelCase_ = {'image_id': 39769, 'annotations': target} # encode them lowerCamelCase_ = DetrImageProcessor.from_pretrained('facebook/detr-resnet-50' ) lowerCamelCase_ = image_processing(images=A_ , annotations=A_ , return_tensors='pt' ) # verify pixel values lowerCamelCase_ = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , A_ ) lowerCamelCase_ = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area lowerCamelCase_ = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , A_ ) ) # verify boxes lowerCamelCase_ = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , A_ ) lowerCamelCase_ = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , A_ , atol=1E-3 ) ) # verify image_id lowerCamelCase_ = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , A_ ) ) # verify is_crowd lowerCamelCase_ = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , A_ ) ) # verify class_labels lowerCamelCase_ = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , A_ ) ) # verify orig_size lowerCamelCase_ = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , A_ ) ) # verify size lowerCamelCase_ = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , A_ ) ) @slow def a__ ( self : str ) -> Any: """simple docstring""" lowerCamelCase_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt' , 'r' ) as f: lowerCamelCase_ = json.loads(f.read() ) lowerCamelCase_ = {'file_name': '000000039769.png', 'image_id': 39769, 'segments_info': target} lowerCamelCase_ = pathlib.Path('./tests/fixtures/tests_samples/COCO/coco_panoptic' ) # encode them lowerCamelCase_ = DetrImageProcessor.from_pretrained('facebook/detr-resnet-50-panoptic' ) lowerCamelCase_ = image_processing(images=A_ , annotations=A_ , masks_path=A_ , return_tensors='pt' ) # verify pixel values lowerCamelCase_ = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , A_ ) lowerCamelCase_ = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area lowerCamelCase_ = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , A_ ) ) # verify boxes lowerCamelCase_ = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , A_ ) lowerCamelCase_ = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , A_ , atol=1E-3 ) ) # verify image_id lowerCamelCase_ = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , A_ ) ) # verify is_crowd lowerCamelCase_ = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , A_ ) ) # verify class_labels lowerCamelCase_ = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , A_ ) ) # verify masks lowerCamelCase_ = 822873 self.assertEqual(encoding['labels'][0]['masks'].sum().item() , A_ ) # verify orig_size lowerCamelCase_ = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , A_ ) ) # verify size lowerCamelCase_ = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , A_ ) )
70
0
'''simple docstring''' import gc import unittest from diffusers import FlaxControlNetModel, FlaxStableDiffusionControlNetPipeline from diffusers.utils import is_flax_available, load_image, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class lowercase_ ( unittest.TestCase ): """simple docstring""" def __UpperCAmelCase ( self : str ) -> Dict: # clean up the VRAM after each test super().tearDown() gc.collect() def __UpperCAmelCase ( self : List[str] ) -> Union[str, Any]: _A , _A = FlaxControlNetModel.from_pretrained( 'lllyasviel/sd-controlnet-canny', from_pt=UpperCamelCase__, dtype=jnp.bfloataa ) _A , _A = FlaxStableDiffusionControlNetPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5', controlnet=UpperCamelCase__, from_pt=UpperCamelCase__, dtype=jnp.bfloataa ) _A = controlnet_params _A = 'bird' _A = jax.device_count() _A = pipe.prepare_text_inputs([prompts] * num_samples ) _A = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png' ) _A = pipe.prepare_image_inputs([canny_image] * num_samples ) _A = jax.random.PRNGKey(0 ) _A = jax.random.split(UpperCamelCase__, jax.device_count() ) _A = replicate(UpperCamelCase__ ) _A = shard(UpperCamelCase__ ) _A = shard(UpperCamelCase__ ) _A = pipe( prompt_ids=UpperCamelCase__, image=UpperCamelCase__, params=UpperCamelCase__, prng_seed=UpperCamelCase__, num_inference_steps=50, jit=UpperCamelCase__, ).images assert images.shape == (jax.device_count(), 1, 7_68, 5_12, 3) _A = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) _A = images[0, 2_53:2_56, 2_53:2_56, -1] _A = jnp.asarray(jax.device_get(image_slice.flatten() ) ) _A = jnp.array( [0.167_969, 0.116_699, 0.081_543, 0.154_297, 0.132_812, 0.108_887, 0.169_922, 0.169_922, 0.205_078] ) print(f'output_slice: {output_slice}' ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2 def __UpperCAmelCase ( self : Union[str, Any] ) -> List[str]: _A , _A = FlaxControlNetModel.from_pretrained( 'lllyasviel/sd-controlnet-openpose', from_pt=UpperCamelCase__, dtype=jnp.bfloataa ) _A , _A = FlaxStableDiffusionControlNetPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5', controlnet=UpperCamelCase__, from_pt=UpperCamelCase__, dtype=jnp.bfloataa ) _A = controlnet_params _A = 'Chef in the kitchen' _A = jax.device_count() _A = pipe.prepare_text_inputs([prompts] * num_samples ) _A = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/pose.png' ) _A = pipe.prepare_image_inputs([pose_image] * num_samples ) _A = jax.random.PRNGKey(0 ) _A = jax.random.split(UpperCamelCase__, jax.device_count() ) _A = replicate(UpperCamelCase__ ) _A = shard(UpperCamelCase__ ) _A = shard(UpperCamelCase__ ) _A = pipe( prompt_ids=UpperCamelCase__, image=UpperCamelCase__, params=UpperCamelCase__, prng_seed=UpperCamelCase__, num_inference_steps=50, jit=UpperCamelCase__, ).images assert images.shape == (jax.device_count(), 1, 7_68, 5_12, 3) _A = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) _A = images[0, 2_53:2_56, 2_53:2_56, -1] _A = jnp.asarray(jax.device_get(image_slice.flatten() ) ) _A = jnp.array( [[0.271_484, 0.261_719, 0.275_391, 0.277_344, 0.279_297, 0.291_016, 0.294_922, 0.302_734, 0.302_734]] ) print(f'output_slice: {output_slice}' ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2
107
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase : Dict = logging.get_logger(__name__) lowerCamelCase : int = { "microsoft/swinv2-tiny-patch4-window8-256": ( "https://huggingface.co/microsoft/swinv2-tiny-patch4-window8-256/resolve/main/config.json" ), } class A( UpperCamelCase ): '''simple docstring''' UpperCamelCase = '''swinv2''' UpperCamelCase = { '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers''', } def __init__( self : Dict , A_ : List[Any]=224 , A_ : Optional[Any]=4 , A_ : int=3 , A_ : Dict=96 , A_ : Any=[2, 2, 6, 2] , A_ : Optional[Any]=[3, 6, 12, 24] , A_ : Tuple=7 , A_ : Tuple=4.0 , A_ : str=True , A_ : str=0.0 , A_ : Union[str, Any]=0.0 , A_ : Optional[Any]=0.1 , A_ : str="gelu" , A_ : int=False , A_ : str=0.02 , A_ : List[Any]=1E-5 , A_ : Any=32 , **A_ : Tuple , ) -> Any: """simple docstring""" super().__init__(**A_ ) lowerCamelCase_ = image_size lowerCamelCase_ = patch_size lowerCamelCase_ = num_channels lowerCamelCase_ = embed_dim lowerCamelCase_ = depths lowerCamelCase_ = len(A_ ) lowerCamelCase_ = num_heads lowerCamelCase_ = window_size lowerCamelCase_ = mlp_ratio lowerCamelCase_ = qkv_bias lowerCamelCase_ = hidden_dropout_prob lowerCamelCase_ = attention_probs_dropout_prob lowerCamelCase_ = drop_path_rate lowerCamelCase_ = hidden_act lowerCamelCase_ = use_absolute_embeddings lowerCamelCase_ = layer_norm_eps lowerCamelCase_ = initializer_range lowerCamelCase_ = encoder_stride # we set the hidden_size attribute in order to make Swinv2 work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowerCamelCase_ = int(embed_dim * 2 ** (len(A_ ) - 1) ) lowerCamelCase_ = (0, 0, 0, 0)
70
0
import functools def _SCREAMING_SNAKE_CASE ( __snake_case , __snake_case ) -> int: # Validation if not isinstance(__snake_case , __snake_case ) or not all(isinstance(__snake_case , __snake_case ) for day in days ): raise ValueError("""The parameter days should be a list of integers""" ) if len(__snake_case ) != 3 or not all(isinstance(__snake_case , __snake_case ) for cost in costs ): raise ValueError("""The parameter costs should be a list of three integers""" ) if len(__snake_case ) == 0: return 0 if min(__snake_case ) <= 0: raise ValueError("""All days elements should be greater than 0""" ) if max(__snake_case ) >= 3_6_6: raise ValueError("""All days elements should be less than 366""" ) _UpperCAmelCase = set(__snake_case ) @functools.cache def dynamic_programming(__snake_case ) -> int: if index > 3_6_5: return 0 if index not in days_set: return dynamic_programming(index + 1 ) return min( costs[0] + dynamic_programming(index + 1 ) , costs[1] + dynamic_programming(index + 7 ) , costs[2] + dynamic_programming(index + 3_0 ) , ) return dynamic_programming(1 ) if __name__ == "__main__": import doctest doctest.testmod()
108
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import ( AutoProcessor, BertTokenizerFast, BlipImageProcessor, GPTaTokenizer, InstructBlipProcessor, PreTrainedTokenizerFast, ) @require_vision class A( unittest.TestCase ): '''simple docstring''' def a__ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = tempfile.mkdtemp() lowerCamelCase_ = BlipImageProcessor() lowerCamelCase_ = GPTaTokenizer.from_pretrained('hf-internal-testing/tiny-random-GPT2Model' ) lowerCamelCase_ = BertTokenizerFast.from_pretrained('hf-internal-testing/tiny-random-bert' ) lowerCamelCase_ = InstructBlipProcessor(A_ , A_ , A_ ) processor.save_pretrained(self.tmpdirname ) def a__ ( self : Optional[int] , **A_ : Optional[int] ) -> Dict: """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **A_ ).tokenizer def a__ ( self : List[str] , **A_ : str ) -> Optional[Any]: """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **A_ ).image_processor def a__ ( self : Tuple , **A_ : Any ) -> Optional[int]: """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **A_ ).qformer_tokenizer def a__ ( self : str ) -> str: """simple docstring""" shutil.rmtree(self.tmpdirname ) def a__ ( self : Dict ) -> Dict: """simple docstring""" lowerCamelCase_ = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] lowerCamelCase_ = [Image.fromarray(np.moveaxis(A_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def a__ ( self : Tuple ) -> Tuple: """simple docstring""" lowerCamelCase_ = InstructBlipProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() , qformer_tokenizer=self.get_qformer_tokenizer() , ) processor.save_pretrained(self.tmpdirname ) lowerCamelCase_ = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) lowerCamelCase_ = self.get_image_processor(do_normalize=A_ , padding_value=1.0 ) lowerCamelCase_ = InstructBlipProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=A_ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , A_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , A_ ) self.assertIsInstance(processor.qformer_tokenizer , A_ ) def a__ ( self : List[str] ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = self.prepare_image_inputs() lowerCamelCase_ = image_processor(A_ , return_tensors='np' ) lowerCamelCase_ = processor(images=A_ , return_tensors='np' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def a__ ( self : Union[str, Any] ) -> Tuple: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = 'lower newer' lowerCamelCase_ = processor(text=A_ ) lowerCamelCase_ = tokenizer(A_ , return_token_type_ids=A_ ) lowerCamelCase_ = qformer_tokenizer(A_ , return_token_type_ids=A_ ) for key in encoded_tokens.keys(): self.assertListEqual(encoded_tokens[key] , encoded_processor[key] ) for key in encoded_tokens_qformer.keys(): self.assertListEqual(encoded_tokens_qformer[key] , encoded_processor['qformer_' + key] ) def a__ ( self : Tuple ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = 'lower newer' lowerCamelCase_ = self.prepare_image_inputs() lowerCamelCase_ = processor(text=A_ , images=A_ ) self.assertListEqual( list(inputs.keys() ) , ['input_ids', 'attention_mask', 'qformer_input_ids', 'qformer_attention_mask', 'pixel_values'] , ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def a__ ( self : Optional[int] ) -> int: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowerCamelCase_ = processor.batch_decode(A_ ) lowerCamelCase_ = tokenizer.batch_decode(A_ ) self.assertListEqual(A_ , A_ ) def a__ ( self : Tuple ) -> List[str]: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = 'lower newer' lowerCamelCase_ = self.prepare_image_inputs() lowerCamelCase_ = processor(text=A_ , images=A_ ) self.assertListEqual( list(inputs.keys() ) , ['input_ids', 'attention_mask', 'qformer_input_ids', 'qformer_attention_mask', 'pixel_values'] , )
70
0
'''simple docstring''' from __future__ import annotations import requests a = set( "approved_at_utc approved_by author_flair_background_color\nauthor_flair_css_class author_flair_richtext author_flair_template_id author_fullname\nauthor_premium can_mod_post category clicked content_categories created_utc downs\nedited gilded gildings hidden hide_score is_created_from_ads_ui is_meta\nis_original_content is_reddit_media_domain is_video link_flair_css_class\nlink_flair_richtext link_flair_text link_flair_text_color media_embed mod_reason_title\nname permalink pwls quarantine saved score secure_media secure_media_embed selftext\nsubreddit subreddit_name_prefixed subreddit_type thumbnail title top_awarded_type\ntotal_awards_received ups upvote_ratio url user_reports".split() ) def __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase = 1 , __UpperCAmelCase = "new" , __UpperCAmelCase = None ) -> dict: '''simple docstring''' __SCREAMING_SNAKE_CASE = wanted_data or [] if invalid_search_terms := ", ".join(sorted(set(__UpperCAmelCase ) - valid_terms ) ): __SCREAMING_SNAKE_CASE = f"""Invalid search term: {invalid_search_terms}""" raise ValueError(__UpperCAmelCase ) __SCREAMING_SNAKE_CASE = requests.get( f"""https://reddit.com/r/{subreddit}/{age}.json?limit={limit}""" , headers={"""User-agent""": """A random string"""} , ) if response.status_code == 429: raise requests.HTTPError __SCREAMING_SNAKE_CASE = response.json() if not wanted_data: return {id_: data["data"]["children"][id_] for id_ in range(__UpperCAmelCase )} __SCREAMING_SNAKE_CASE = {} for id_ in range(__UpperCAmelCase ): __SCREAMING_SNAKE_CASE = { item: data["""data"""]["""children"""][id_]["""data"""][item] for item in wanted_data } return data_dict if __name__ == "__main__": # If you get Error 429, that means you are rate limited.Try after some time print(get_subreddit_data("learnpython", wanted_data=["title", "url", "selftext"]))
109
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_OBJECT_DETECTION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING lowerCamelCase : Tuple = logging.get_logger(__name__) lowerCamelCase : List[Any] = Dict[str, Any] lowerCamelCase : Dict = List[Prediction] @add_end_docstrings(UpperCamelCase ) class A( UpperCamelCase ): '''simple docstring''' def __init__( self : Tuple , *A_ : int , **A_ : int ) -> Optional[int]: """simple docstring""" super().__init__(*A_ , **A_ ) if self.framework == "tf": raise ValueError(f"""The {self.__class__} is only available in PyTorch.""" ) requires_backends(self , 'vision' ) self.check_model_type( dict(MODEL_FOR_OBJECT_DETECTION_MAPPING.items() + MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING.items() ) ) def a__ ( self : Union[str, Any] , **A_ : Union[str, Any] ) -> List[str]: """simple docstring""" lowerCamelCase_ = {} if "threshold" in kwargs: lowerCamelCase_ = kwargs['threshold'] return {}, {}, postprocess_kwargs def __call__( self : str , *A_ : Optional[int] , **A_ : Tuple ) -> Union[Predictions, List[Prediction]]: """simple docstring""" return super().__call__(*A_ , **A_ ) def a__ ( self : Union[str, Any] , A_ : Tuple ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = load_image(A_ ) lowerCamelCase_ = torch.IntTensor([[image.height, image.width]] ) lowerCamelCase_ = self.image_processor(images=[image] , return_tensors='pt' ) if self.tokenizer is not None: lowerCamelCase_ = self.tokenizer(text=inputs['words'] , boxes=inputs['boxes'] , return_tensors='pt' ) lowerCamelCase_ = target_size return inputs def a__ ( self : Union[str, Any] , A_ : List[str] ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = model_inputs.pop('target_size' ) lowerCamelCase_ = self.model(**A_ ) lowerCamelCase_ = outputs.__class__({'target_size': target_size, **outputs} ) if self.tokenizer is not None: lowerCamelCase_ = model_inputs['bbox'] return model_outputs def a__ ( self : str , A_ : Any , A_ : Tuple=0.9 ) -> str: """simple docstring""" lowerCamelCase_ = model_outputs['target_size'] if self.tokenizer is not None: # This is a LayoutLMForTokenClassification variant. # The OCR got the boxes and the model classified the words. lowerCamelCase_ , lowerCamelCase_ = target_size[0].tolist() def unnormalize(A_ : Dict ): return self._get_bounding_box( torch.Tensor( [ (width * bbox[0] / 1000), (height * bbox[1] / 1000), (width * bbox[2] / 1000), (height * bbox[3] / 1000), ] ) ) lowerCamelCase_ , lowerCamelCase_ = model_outputs['logits'].squeeze(0 ).softmax(dim=-1 ).max(dim=-1 ) lowerCamelCase_ = [self.model.config.idalabel[prediction] for prediction in classes.tolist()] lowerCamelCase_ = [unnormalize(A_ ) for bbox in model_outputs['bbox'].squeeze(0 )] lowerCamelCase_ = ['score', 'label', 'box'] lowerCamelCase_ = [dict(zip(A_ , A_ ) ) for vals in zip(scores.tolist() , A_ , A_ ) if vals[0] > threshold] else: # This is a regular ForObjectDetectionModel lowerCamelCase_ = self.image_processor.post_process_object_detection(A_ , A_ , A_ ) lowerCamelCase_ = raw_annotations[0] lowerCamelCase_ = raw_annotation['scores'] lowerCamelCase_ = raw_annotation['labels'] lowerCamelCase_ = raw_annotation['boxes'] lowerCamelCase_ = scores.tolist() lowerCamelCase_ = [self.model.config.idalabel[label.item()] for label in labels] lowerCamelCase_ = [self._get_bounding_box(A_ ) for box in boxes] # {"scores": [...], ...} --> [{"score":x, ...}, ...] lowerCamelCase_ = ['score', 'label', 'box'] lowerCamelCase_ = [ dict(zip(A_ , A_ ) ) for vals in zip(raw_annotation['scores'] , raw_annotation['labels'] , raw_annotation['boxes'] ) ] return annotation def a__ ( self : Union[str, Any] , A_ : "torch.Tensor" ) -> Dict[str, int]: """simple docstring""" if self.framework != "pt": raise ValueError('The ObjectDetectionPipeline is only available in PyTorch.' ) lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = box.int().tolist() lowerCamelCase_ = { 'xmin': xmin, 'ymin': ymin, 'xmax': xmax, 'ymax': ymax, } return bbox
70
0
"""simple docstring""" from dataclasses import asdict, dataclass from typing import Optional from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) # TODO Update this UpperCamelCase__ = { 'facebook/esm-1b': 'https://huggingface.co/facebook/esm-1b/resolve/main/config.json', # See all ESM models at https://huggingface.co/models?filter=esm } class a ( lowercase ): UpperCamelCase : str = """esm""" def __init__( self , UpperCamelCase_=None , UpperCamelCase_=None , UpperCamelCase_=None , UpperCamelCase_=768 , UpperCamelCase_=12 , UpperCamelCase_=12 , UpperCamelCase_=3_072 , UpperCamelCase_=0.1 , UpperCamelCase_=0.1 , UpperCamelCase_=1_026 , UpperCamelCase_=0.02 , UpperCamelCase_=1E-12 , UpperCamelCase_="absolute" , UpperCamelCase_=True , UpperCamelCase_=None , UpperCamelCase_=False , UpperCamelCase_=False , UpperCamelCase_=None , UpperCamelCase_=None , **UpperCamelCase_ , ): super().__init__(pad_token_id=UpperCamelCase_ , mask_token_id=UpperCamelCase_ , **UpperCamelCase_ ) UpperCAmelCase__ : Optional[int] = vocab_size UpperCAmelCase__ : Dict = hidden_size UpperCAmelCase__ : List[Any] = num_hidden_layers UpperCAmelCase__ : int = num_attention_heads UpperCAmelCase__ : Union[str, Any] = intermediate_size UpperCAmelCase__ : Optional[Any] = hidden_dropout_prob UpperCAmelCase__ : List[str] = attention_probs_dropout_prob UpperCAmelCase__ : Dict = max_position_embeddings UpperCAmelCase__ : Dict = initializer_range UpperCAmelCase__ : Any = layer_norm_eps UpperCAmelCase__ : Optional[int] = position_embedding_type UpperCAmelCase__ : Union[str, Any] = use_cache UpperCAmelCase__ : str = emb_layer_norm_before UpperCAmelCase__ : List[str] = token_dropout UpperCAmelCase__ : List[Any] = is_folding_model if is_folding_model: if esmfold_config is None: logger.info('No esmfold_config supplied for folding model, using default values.' ) UpperCAmelCase__ : str = EsmFoldConfig() elif isinstance(UpperCamelCase_ , UpperCamelCase_ ): UpperCAmelCase__ : Dict = EsmFoldConfig(**UpperCamelCase_ ) UpperCAmelCase__ : Any = esmfold_config if vocab_list is None: logger.warning('No vocab_list supplied for folding model, assuming the ESM-2 vocabulary!' ) UpperCAmelCase__ : Dict = get_default_vocab_list() else: UpperCAmelCase__ : Optional[int] = vocab_list else: UpperCAmelCase__ : str = None UpperCAmelCase__ : str = None if self.esmfold_config is not None and getattr(self.esmfold_config , 'use_esm_attn_map' , UpperCamelCase_ ): raise ValueError('The HuggingFace port of ESMFold does not support use_esm_attn_map at this time!' ) def __snake_case ( self ): UpperCAmelCase__ : Optional[int] = super().to_dict() if isinstance(self.esmfold_config , UpperCamelCase_ ): UpperCAmelCase__ : Optional[int] = self.esmfold_config.to_dict() return output @dataclass class a : UpperCamelCase : str = None UpperCamelCase : bool = True UpperCamelCase : bool = False UpperCamelCase : bool = False UpperCamelCase : bool = False UpperCamelCase : float = 0 UpperCamelCase : bool = True UpperCamelCase : bool = False UpperCamelCase : int = 1_2_8 UpperCamelCase : "TrunkConfig" = None def __snake_case ( self ): if self.trunk is None: UpperCAmelCase__ : Dict = TrunkConfig() elif isinstance(self.trunk , UpperCamelCase_ ): UpperCAmelCase__ : Any = TrunkConfig(**self.trunk ) def __snake_case ( self ): UpperCAmelCase__ : Dict = asdict(self ) UpperCAmelCase__ : Union[str, Any] = self.trunk.to_dict() return output @dataclass class a : UpperCamelCase : int = 4_8 UpperCamelCase : int = 1_0_2_4 UpperCamelCase : int = 1_2_8 UpperCamelCase : int = 3_2 UpperCamelCase : int = 3_2 UpperCamelCase : int = 3_2 UpperCamelCase : float = 0 UpperCamelCase : float = 0 UpperCamelCase : bool = False UpperCamelCase : int = 4 UpperCamelCase : Optional[int] = 1_2_8 UpperCamelCase : "StructureModuleConfig" = None def __snake_case ( self ): if self.structure_module is None: UpperCAmelCase__ : Optional[Any] = StructureModuleConfig() elif isinstance(self.structure_module , UpperCamelCase_ ): UpperCAmelCase__ : Optional[Any] = StructureModuleConfig(**self.structure_module ) if self.max_recycles <= 0: raise ValueError(F'''`max_recycles` should be positive, got {self.max_recycles}.''' ) if self.sequence_state_dim % self.sequence_state_dim != 0: raise ValueError( '`sequence_state_dim` should be a round multiple of `sequence_state_dim`, got' F''' {self.sequence_state_dim} and {self.sequence_state_dim}.''' ) if self.pairwise_state_dim % self.pairwise_state_dim != 0: raise ValueError( '`pairwise_state_dim` should be a round multiple of `pairwise_state_dim`, got' F''' {self.pairwise_state_dim} and {self.pairwise_state_dim}.''' ) UpperCAmelCase__ : Optional[int] = self.sequence_state_dim // self.sequence_head_width UpperCAmelCase__ : List[str] = self.pairwise_state_dim // self.pairwise_head_width if self.sequence_state_dim != sequence_num_heads * self.sequence_head_width: raise ValueError( '`sequence_state_dim` should be equal to `sequence_num_heads * sequence_head_width, got' F''' {self.sequence_state_dim} != {sequence_num_heads} * {self.sequence_head_width}.''' ) if self.pairwise_state_dim != pairwise_num_heads * self.pairwise_head_width: raise ValueError( '`pairwise_state_dim` should be equal to `pairwise_num_heads * pairwise_head_width, got' F''' {self.pairwise_state_dim} != {pairwise_num_heads} * {self.pairwise_head_width}.''' ) if self.pairwise_state_dim % 2 != 0: raise ValueError(F'''`pairwise_state_dim` should be even, got {self.pairwise_state_dim}.''' ) if self.dropout >= 0.4: raise ValueError(F'''`dropout` should not be greater than 0.4, got {self.dropout}.''' ) def __snake_case ( self ): UpperCAmelCase__ : str = asdict(self ) UpperCAmelCase__ : Union[str, Any] = self.structure_module.to_dict() return output @dataclass class a : UpperCamelCase : int = 3_8_4 UpperCamelCase : int = 1_2_8 UpperCamelCase : int = 1_6 UpperCamelCase : int = 1_2_8 UpperCamelCase : int = 1_2 UpperCamelCase : int = 4 UpperCamelCase : int = 8 UpperCamelCase : float = 0.1 UpperCamelCase : int = 8 UpperCamelCase : int = 1 UpperCamelCase : int = 2 UpperCamelCase : int = 7 UpperCamelCase : int = 1_0 UpperCamelCase : float = 1E-8 UpperCamelCase : float = 1E5 def __snake_case ( self ): return asdict(self ) def lowerCamelCase ( ): return ( "<cls>", "<pad>", "<eos>", "<unk>", "L", "A", "G", "V", "S", "E", "R", "T", "I", "D", "P", "K", "Q", "N", "F", "Y", "M", "H", "W", "C", "X", "B", "U", "Z", "O", ".", "-", "<null_1>", "<mask>", )
110
from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
70
0
import argparse import json import os from tensorflow.core.protobuf.saved_model_pba import SavedModel # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py A : Union[str, Any] = "." # Internal TensorFlow ops that can be safely ignored (mostly specific to a saved model) A : List[str] = [ "Assert", "AssignVariableOp", "EmptyTensorList", "MergeV2Checkpoints", "ReadVariableOp", "ResourceGather", "RestoreV2", "SaveV2", "ShardedFilename", "StatefulPartitionedCall", "StaticRegexFullMatch", "VarHandleOp", ] def __lowerCAmelCase ( a__ , a__ , a__ ) -> Optional[int]: __a = SavedModel() __a = [] with open(os.path.join(a__ , '''utils''' , '''tf_ops''' , '''onnx.json''' ) ) as f: __a = json.load(a__ )['''opsets'''] for i in range(1 , opset + 1 ): onnx_ops.extend(onnx_opsets[str(a__ )] ) with open(a__ , '''rb''' ) as f: saved_model.ParseFromString(f.read() ) __a = set() # Iterate over every metagraph in case there is more than one (a saved model can contain multiple graphs) for meta_graph in saved_model.meta_graphs: # Add operations in the graph definition model_op_names.update(node.op for node in meta_graph.graph_def.node ) # Go through the functions in the graph definition for func in meta_graph.graph_def.library.function: # Add operations in each function model_op_names.update(node.op for node in func.node_def ) # Convert to list, sorted if you want __a = sorted(a__ ) __a = [] for op in model_op_names: if op not in onnx_ops and op not in INTERNAL_OPS: incompatible_ops.append(a__ ) if strict and len(a__ ) > 0: raise Exception(F"""Found the following incompatible ops for the opset {opset}:\n""" + incompatible_ops ) elif len(a__ ) > 0: print(F"""Found the following incompatible ops for the opset {opset}:""" ) print(*a__ , sep='''\n''' ) else: print(F"""The saved model {saved_model_path} can properly be converted with ONNX.""" ) if __name__ == "__main__": A : Tuple = argparse.ArgumentParser() parser.add_argument('--saved_model_path', help='Path of the saved model to check (the .pb file).') parser.add_argument( '--opset', default=1_2, type=int, help='The ONNX opset against which the model has to be tested.' ) parser.add_argument( '--framework', choices=['onnx'], default='onnx', help='Frameworks against which to test the saved model.' ) parser.add_argument( '--strict', action='store_true', help='Whether make the checking strict (raise errors) or not (raise warnings)' ) A : Optional[Any] = parser.parse_args() if args.framework == "onnx": onnx_compliancy(args.saved_model_path, args.strict, args.opset)
219
from collections import Counter from timeit import timeit def _SCREAMING_SNAKE_CASE ( lowercase : str = "" , ): '''simple docstring''' return sum(c % 2 for c in Counter(input_str.replace(' ' , '' ).lower() ).values() ) < 2 def _SCREAMING_SNAKE_CASE ( lowercase : str = "" ): '''simple docstring''' if len(lowercase ) == 0: return True lowerCamelCase_ = input_str.replace(' ' , '' ).lower() # character_freq_dict: Stores the frequency of every character in the input string lowerCamelCase_ = {} for character in lower_case_input_str: lowerCamelCase_ = character_freq_dict.get(lowercase , 0 ) + 1 lowerCamelCase_ = 0 for character_count in character_freq_dict.values(): if character_count % 2: odd_char += 1 if odd_char > 1: return False return True def _SCREAMING_SNAKE_CASE ( lowercase : str = "" ): '''simple docstring''' print('\nFor string = ' , lowercase , ':' ) print( '> can_string_be_rearranged_as_palindrome_counter()' , '\tans =' , can_string_be_rearranged_as_palindrome_counter(lowercase ) , '\ttime =' , timeit( 'z.can_string_be_rearranged_as_palindrome_counter(z.check_str)' , setup='import __main__ as z' , ) , 'seconds' , ) print( '> can_string_be_rearranged_as_palindrome()' , '\tans =' , can_string_be_rearranged_as_palindrome(lowercase ) , '\ttime =' , timeit( 'z.can_string_be_rearranged_as_palindrome(z.check_str)' , setup='import __main__ as z' , ) , 'seconds' , ) if __name__ == "__main__": lowerCamelCase : Optional[Any] = input( "Enter string to determine if it can be rearranged as a palindrome or not: " ).strip() benchmark(check_str) lowerCamelCase : int = can_string_be_rearranged_as_palindrome_counter(check_str) print(F"""{check_str} can {'' if status else 'not '}be rearranged as a palindrome""")
70
0
'''simple docstring''' import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def UpperCamelCase__ ( _lowercase : Optional[Any] ) -> Dict: __UpperCAmelCase: Optional[Any] = args.pruning_method __UpperCAmelCase: Any = args.threshold __UpperCAmelCase: int = args.model_name_or_path.rstrip("""/""" ) __UpperCAmelCase: List[Any] = args.target_model_path print(F'''Load fine-pruned model from {model_name_or_path}''' ) __UpperCAmelCase: Dict = torch.load(os.path.join(_lowercase , """pytorch_model.bin""" ) ) __UpperCAmelCase: Optional[int] = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: __UpperCAmelCase: List[Any] = tensor print(F'''Copied layer {name}''' ) elif "classifier" in name or "qa_output" in name: __UpperCAmelCase: Optional[int] = tensor print(F'''Copied layer {name}''' ) elif "bias" in name: __UpperCAmelCase: Any = tensor print(F'''Copied layer {name}''' ) else: if pruning_method == "magnitude": __UpperCAmelCase: List[str] = MagnitudeBinarizer.apply(inputs=_lowercase , threshold=_lowercase ) __UpperCAmelCase: Any = tensor * mask print(F'''Pruned layer {name}''' ) elif pruning_method == "topK": if "mask_scores" in name: continue __UpperCAmelCase: int = name[:-6] __UpperCAmelCase: Optional[int] = model[F'''{prefix_}mask_scores'''] __UpperCAmelCase: Optional[int] = TopKBinarizer.apply(_lowercase , _lowercase ) __UpperCAmelCase: Dict = tensor * mask print(F'''Pruned layer {name}''' ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue __UpperCAmelCase: Union[str, Any] = name[:-6] __UpperCAmelCase: int = model[F'''{prefix_}mask_scores'''] __UpperCAmelCase: Optional[int] = ThresholdBinarizer.apply(_lowercase , _lowercase , _lowercase ) __UpperCAmelCase: Union[str, Any] = tensor * mask print(F'''Pruned layer {name}''' ) elif pruning_method == "l0": if "mask_scores" in name: continue __UpperCAmelCase: List[str] = name[:-6] __UpperCAmelCase: Tuple = model[F'''{prefix_}mask_scores'''] __UpperCAmelCase, __UpperCAmelCase: str = -0.1, 1.1 __UpperCAmelCase: List[str] = torch.sigmoid(_lowercase ) __UpperCAmelCase: Optional[Any] = s * (r - l) + l __UpperCAmelCase: Dict = s_bar.clamp(min=0.0 , max=1.0 ) __UpperCAmelCase: List[Any] = tensor * mask print(F'''Pruned layer {name}''' ) else: raise ValueError("""Unknown pruning method""" ) if target_model_path is None: __UpperCAmelCase: Any = os.path.join( os.path.dirname(_lowercase ) , F'''bertarized_{os.path.basename(_lowercase )}''' ) if not os.path.isdir(_lowercase ): shutil.copytree(_lowercase , _lowercase ) print(F'''\nCreated folder {target_model_path}''' ) torch.save(_lowercase , os.path.join(_lowercase , """pytorch_model.bin""" ) ) print("""\nPruned model saved! See you later!""" ) if __name__ == "__main__": SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser() parser.add_argument( '--pruning_method', choices=['l0', 'magnitude', 'topK', 'sigmoied_threshold'], type=str, required=True, help=( 'Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning,' ' sigmoied_threshold = Soft movement pruning)' ), ) parser.add_argument( '--threshold', type=float, required=False, help=( 'For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model.' 'For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared.' 'Not needed for `l0`' ), ) parser.add_argument( '--model_name_or_path', type=str, required=True, help='Folder containing the model that was previously fine-pruned', ) parser.add_argument( '--target_model_path', default=None, type=str, required=False, help='Folder containing the model that was previously fine-pruned', ) SCREAMING_SNAKE_CASE_ = parser.parse_args() main(args)
523
from __future__ import annotations from decimal import Decimal from math import * # noqa: F403 from sympy import diff def _SCREAMING_SNAKE_CASE ( lowercase : str , lowercase : float | Decimal , lowercase : float = 10**-10 ): '''simple docstring''' lowerCamelCase_ = a while True: lowerCamelCase_ = Decimal(lowercase ) - ( Decimal(eval(lowercase ) ) / Decimal(eval(str(diff(lowercase ) ) ) ) # noqa: S307 ) # This number dictates the accuracy of the answer if abs(eval(lowercase ) ) < precision: # noqa: S307 return float(lowercase ) # Let's Execute if __name__ == "__main__": # Find root of trigonometric function # Find value of pi print(F"""The root of sin(x) = 0 is {newton_raphson('sin(x)', 2)}""") # Find root of polynomial print(F"""The root of x**2 - 5*x + 2 = 0 is {newton_raphson('x**2 - 5*x + 2', 0.4)}""") # Find Square Root of 5 print(F"""The root of log(x) - 1 = 0 is {newton_raphson('log(x) - 1', 2)}""") # Exponential Roots print(F"""The root of exp(x) - 1 = 0 is {newton_raphson('exp(x) - 1', 0)}""")
70
0
import unittest import numpy as np import torch from torch import nn from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import KandinskyVaaPriorPipeline, PriorTransformer, UnCLIPScheduler from diffusers.utils import torch_device from diffusers.utils.testing_utils import enable_full_determinism, skip_mps from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class _SCREAMING_SNAKE_CASE ( snake_case, unittest.TestCase ): lowerCamelCase_ = KandinskyVaaPriorPipeline lowerCamelCase_ = ['prompt'] lowerCamelCase_ = ['prompt', 'negative_prompt'] lowerCamelCase_ = [ 'num_images_per_prompt', 'generator', 'num_inference_steps', 'latents', 'negative_prompt', 'guidance_scale', 'output_type', 'return_dict', ] lowerCamelCase_ = False @property def _UpperCAmelCase ( self : Union[str, Any] ): """simple docstring""" return 32 @property def _UpperCAmelCase ( self : str ): """simple docstring""" return 32 @property def _UpperCAmelCase ( self : Tuple ): """simple docstring""" return self.time_input_dim @property def _UpperCAmelCase ( self : str ): """simple docstring""" return self.time_input_dim * 4 @property def _UpperCAmelCase ( self : Optional[int] ): """simple docstring""" return 100 @property def _UpperCAmelCase ( self : Optional[Any] ): """simple docstring""" A : Tuple = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) return tokenizer @property def _UpperCAmelCase ( self : Any ): """simple docstring""" torch.manual_seed(0 ) A : int = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModelWithProjection(A_ ) @property def _UpperCAmelCase ( self : Optional[int] ): """simple docstring""" torch.manual_seed(0 ) A : List[str] = { '''num_attention_heads''': 2, '''attention_head_dim''': 12, '''embedding_dim''': self.text_embedder_hidden_size, '''num_layers''': 1, } A : Optional[int] = PriorTransformer(**A_ ) # clip_std and clip_mean is initialized to be 0 so PriorTransformer.post_process_latents will always return 0 - set clip_std to be 1 so it won't return 0 A : Union[str, Any] = nn.Parameter(torch.ones(model.clip_std.shape ) ) return model @property def _UpperCAmelCase ( self : int ): """simple docstring""" torch.manual_seed(0 ) A : Dict = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , image_size=224 , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_channels=3 , num_hidden_layers=5 , patch_size=14 , ) A : int = CLIPVisionModelWithProjection(A_ ) return model @property def _UpperCAmelCase ( self : Optional[int] ): """simple docstring""" A : str = CLIPImageProcessor( crop_size=224 , do_center_crop=A_ , do_normalize=A_ , do_resize=A_ , image_mean=[0.48_14_54_66, 0.4_57_82_75, 0.40_82_10_73] , image_std=[0.26_86_29_54, 0.26_13_02_58, 0.27_57_77_11] , resample=3 , size=224 , ) return image_processor def _UpperCAmelCase ( self : Union[str, Any] ): """simple docstring""" A : Optional[int] = self.dummy_prior A : Dict = self.dummy_image_encoder A : int = self.dummy_text_encoder A : List[str] = self.dummy_tokenizer A : Dict = self.dummy_image_processor A : str = UnCLIPScheduler( variance_type='''fixed_small_log''' , prediction_type='''sample''' , num_train_timesteps=1000 , clip_sample=A_ , clip_sample_range=10.0 , ) A : Tuple = { '''prior''': prior, '''image_encoder''': image_encoder, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''scheduler''': scheduler, '''image_processor''': image_processor, } return components def _UpperCAmelCase ( self : Dict , snake_case_ : Optional[int] , snake_case_ : Union[str, Any]=0 ): """simple docstring""" if str(A_ ).startswith('''mps''' ): A : Any = torch.manual_seed(A_ ) else: A : Optional[Any] = torch.Generator(device=A_ ).manual_seed(A_ ) A : Any = { '''prompt''': '''horse''', '''generator''': generator, '''guidance_scale''': 4.0, '''num_inference_steps''': 2, '''output_type''': '''np''', } return inputs def _UpperCAmelCase ( self : Dict ): """simple docstring""" A : Tuple = '''cpu''' A : List[str] = self.get_dummy_components() A : Optional[int] = self.pipeline_class(**A_ ) A : Optional[int] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) A : Union[str, Any] = pipe(**self.get_dummy_inputs(A_ ) ) A : List[Any] = output.image_embeds A : Union[str, Any] = pipe( **self.get_dummy_inputs(A_ ) , return_dict=A_ , )[0] A : Dict = image[0, -10:] A : Dict = image_from_tuple[0, -10:] assert image.shape == (1, 32) A : str = np.array( [-0.05_32, 1.71_20, 0.36_56, -1.08_52, -0.89_46, -1.17_56, 0.43_48, 0.24_82, 0.51_46, -0.11_56] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @skip_mps def _UpperCAmelCase ( self : Dict ): """simple docstring""" A : Optional[int] = torch_device == '''cpu''' A : Dict = True A : Union[str, Any] = False self._test_inference_batch_single_identical( test_max_difference=A_ , relax_max_difference=A_ , test_mean_pixel_difference=A_ , ) @skip_mps def _UpperCAmelCase ( self : List[str] ): """simple docstring""" A : Optional[int] = torch_device == '''cpu''' A : Tuple = False self._test_attention_slicing_forward_pass( test_max_difference=A_ , test_mean_pixel_difference=A_ , )
256
from __future__ import annotations from typing import Any class A( UpperCamelCase ): '''simple docstring''' pass class A: '''simple docstring''' def __init__( self : List[str] , A_ : Any ) -> None: """simple docstring""" lowerCamelCase_ = data lowerCamelCase_ = None def __iter__( self : int ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self lowerCamelCase_ = [] while node: if node in visited: raise ContainsLoopError visited.append(A_ ) yield node.data lowerCamelCase_ = node.next_node @property def a__ ( self : List[str] ) -> bool: """simple docstring""" try: list(self ) return False except ContainsLoopError: return True if __name__ == "__main__": lowerCamelCase : int = Node(1) lowerCamelCase : Optional[int] = Node(2) lowerCamelCase : Union[str, Any] = Node(3) lowerCamelCase : List[Any] = Node(4) print(root_node.has_loop) # False lowerCamelCase : int = root_node.next_node print(root_node.has_loop) # True lowerCamelCase : Dict = Node(5) lowerCamelCase : Optional[int] = Node(6) lowerCamelCase : str = Node(5) lowerCamelCase : Union[str, Any] = Node(6) print(root_node.has_loop) # False lowerCamelCase : List[str] = Node(1) print(root_node.has_loop) # False
70
0
"""simple docstring""" from collections import Counter from timeit import timeit def lowerCAmelCase_ ( lowercase_ : str = "" , ): '''simple docstring''' return sum(c % 2 for c in Counter(input_str.replace(''' ''' , '''''' ).lower() ).values() ) < 2 def lowerCAmelCase_ ( lowercase_ : str = "" ): '''simple docstring''' if len(lowercase_ ) == 0: return True __SCREAMING_SNAKE_CASE : List[str] = input_str.replace(''' ''' , '''''' ).lower() # character_freq_dict: Stores the frequency of every character in the input string __SCREAMING_SNAKE_CASE : Optional[int] = {} for character in lower_case_input_str: __SCREAMING_SNAKE_CASE : str = character_freq_dict.get(lowercase_ , 0 ) + 1 __SCREAMING_SNAKE_CASE : Optional[int] = 0 for character_count in character_freq_dict.values(): if character_count % 2: odd_char += 1 if odd_char > 1: return False return True def lowerCAmelCase_ ( lowercase_ : str = "" ): '''simple docstring''' print('''\nFor string = ''' , lowercase_ , ''':''' ) print( '''> can_string_be_rearranged_as_palindrome_counter()''' , '''\tans =''' , can_string_be_rearranged_as_palindrome_counter(lowercase_ ) , '''\ttime =''' , timeit( '''z.can_string_be_rearranged_as_palindrome_counter(z.check_str)''' , setup='''import __main__ as z''' , ) , '''seconds''' , ) print( '''> can_string_be_rearranged_as_palindrome()''' , '''\tans =''' , can_string_be_rearranged_as_palindrome(lowercase_ ) , '''\ttime =''' , timeit( '''z.can_string_be_rearranged_as_palindrome(z.check_str)''' , setup='''import __main__ as z''' , ) , '''seconds''' , ) if __name__ == "__main__": _lowerCamelCase = input( '''Enter string to determine if it can be rearranged as a palindrome or not: ''' ).strip() benchmark(check_str) _lowerCamelCase = can_string_be_rearranged_as_palindrome_counter(check_str) print(f'{check_str} can {"" if status else "not "}be rearranged as a palindrome')
674
import unittest import torch from diffusers import DDIMScheduler, DDPMScheduler, UNetaDModel from diffusers.training_utils import set_seed from diffusers.utils.testing_utils import slow lowerCamelCase : int = False class A( unittest.TestCase ): '''simple docstring''' def a__ ( self : int , A_ : Dict=32 ) -> Any: """simple docstring""" set_seed(0 ) lowerCamelCase_ = UNetaDModel(sample_size=A_ , in_channels=3 , out_channels=3 ) lowerCamelCase_ = torch.optim.SGD(model.parameters() , lr=0.0001 ) return model, optimizer @slow def a__ ( self : int ) -> str: """simple docstring""" lowerCamelCase_ = 'cpu' # ensure full determinism without setting the CUBLAS_WORKSPACE_CONFIG env variable lowerCamelCase_ = DDPMScheduler( num_train_timesteps=1000 , beta_start=0.0001 , beta_end=0.02 , beta_schedule='linear' , clip_sample=A_ , ) lowerCamelCase_ = DDIMScheduler( num_train_timesteps=1000 , beta_start=0.0001 , beta_end=0.02 , beta_schedule='linear' , clip_sample=A_ , ) assert ddpm_scheduler.config.num_train_timesteps == ddim_scheduler.config.num_train_timesteps # shared batches for DDPM and DDIM set_seed(0 ) lowerCamelCase_ = [torch.randn((4, 3, 32, 32) ).clip(-1 , 1 ).to(A_ ) for _ in range(4 )] lowerCamelCase_ = [torch.randn((4, 3, 32, 32) ).to(A_ ) for _ in range(4 )] lowerCamelCase_ = [torch.randint(0 , 1000 , (4,) ).long().to(A_ ) for _ in range(4 )] # train with a DDPM scheduler lowerCamelCase_ , lowerCamelCase_ = self.get_model_optimizer(resolution=32 ) model.train().to(A_ ) for i in range(4 ): optimizer.zero_grad() lowerCamelCase_ = ddpm_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) lowerCamelCase_ = model(A_ , timesteps[i] ).sample lowerCamelCase_ = torch.nn.functional.mse_loss(A_ , noise[i] ) loss.backward() optimizer.step() del model, optimizer # recreate the model and optimizer, and retry with DDIM lowerCamelCase_ , lowerCamelCase_ = self.get_model_optimizer(resolution=32 ) model.train().to(A_ ) for i in range(4 ): optimizer.zero_grad() lowerCamelCase_ = ddim_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) lowerCamelCase_ = model(A_ , timesteps[i] ).sample lowerCamelCase_ = torch.nn.functional.mse_loss(A_ , noise[i] ) loss.backward() optimizer.step() del model, optimizer self.assertTrue(torch.allclose(A_ , A_ , atol=1E-5 ) ) self.assertTrue(torch.allclose(A_ , A_ , atol=1E-5 ) )
70
0
'''simple docstring''' from manim import * class lowercase_ ( A ): """simple docstring""" def lowerCAmelCase_ ( self : Tuple ): """simple docstring""" _SCREAMING_SNAKE_CASE = Rectangle(height=0.5 , width=0.5 ) _SCREAMING_SNAKE_CASE = Rectangle(height=0.2_5 , width=0.2_5 ) _SCREAMING_SNAKE_CASE = Rectangle(height=0.4_6 , width=0.4_6 ).set_stroke(width=0 ) _SCREAMING_SNAKE_CASE = [mem.copy() for i in range(6 )] _SCREAMING_SNAKE_CASE = [mem.copy() for i in range(6 )] _SCREAMING_SNAKE_CASE = VGroup(*A_ ).arrange(A_ , buff=0 ) _SCREAMING_SNAKE_CASE = VGroup(*A_ ).arrange(A_ , buff=0 ) _SCREAMING_SNAKE_CASE = VGroup(A_ , A_ ).arrange(A_ , buff=0 ) _SCREAMING_SNAKE_CASE = Text("CPU" , font_size=2_4 ) _SCREAMING_SNAKE_CASE = Group(A_ , A_ ).arrange(A_ , buff=0.5 , aligned_edge=A_ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(A_ ) _SCREAMING_SNAKE_CASE = [mem.copy() for i in range(4 )] _SCREAMING_SNAKE_CASE = VGroup(*A_ ).arrange(A_ , buff=0 ) _SCREAMING_SNAKE_CASE = Text("GPU" , font_size=2_4 ) _SCREAMING_SNAKE_CASE = Group(A_ , A_ ).arrange(A_ , buff=0.5 , aligned_edge=A_ ) gpu.move_to([-1, -1, 0] ) self.add(A_ ) _SCREAMING_SNAKE_CASE = [mem.copy() for i in range(6 )] _SCREAMING_SNAKE_CASE = VGroup(*A_ ).arrange(A_ , buff=0 ) _SCREAMING_SNAKE_CASE = Text("Model" , font_size=2_4 ) _SCREAMING_SNAKE_CASE = Group(A_ , A_ ).arrange(A_ , buff=0.5 , aligned_edge=A_ ) model.move_to([3, -1.0, 0] ) self.add(A_ ) _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] for i, rect in enumerate(A_ ): rect.set_stroke(A_ ) _SCREAMING_SNAKE_CASE = Rectangle(height=0.4_6 / 4 , width=0.4_6 / 3 ).set_stroke(width=0.0 ).set_fill(A_ , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.0_2 , direction=A_ ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(model_cpu_arr[0] , direction=A_ , buff=0.0 ) else: cpu_target.next_to(model_cpu_arr[i - 1] , direction=A_ , buff=0.0 ) self.add(A_ ) model_cpu_arr.append(A_ ) self.add(*A_ , *A_ , *A_ ) _SCREAMING_SNAKE_CASE = [mem.copy() for i in range(6 )] _SCREAMING_SNAKE_CASE = VGroup(*A_ ).arrange(A_ , buff=0 ) _SCREAMING_SNAKE_CASE = Text("Loaded Checkpoint" , font_size=2_4 ) _SCREAMING_SNAKE_CASE = Group(A_ , A_ ).arrange(A_ , buff=0.5 , aligned_edge=A_ ) checkpoint.move_to([3, 0.5, 0] ) self.add(A_ ) _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] for i, rect in enumerate(A_ ): _SCREAMING_SNAKE_CASE = fill.copy().set_fill(A_ , opacity=0.7 ) target.move_to(A_ ) ckpt_arr.append(A_ ) _SCREAMING_SNAKE_CASE = target.copy() if i < 5: cpu_target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.move_to(cpu_right_col_base[i - 5] ) ckpt_cpu_arr.append(A_ ) self.add(*A_ , *A_ ) _SCREAMING_SNAKE_CASE = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _SCREAMING_SNAKE_CASE = MarkupText( F"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=1_8 , ) key_text.move_to([-5, 2.4, 0] ) self.add(A_ , A_ ) _SCREAMING_SNAKE_CASE = MarkupText( F"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=1_8 , ) blue_text.next_to(A_ , DOWN * 2.4 , aligned_edge=key_text.get_left() ) self.add(A_ ) _SCREAMING_SNAKE_CASE = MarkupText( F"""Based on the passed in configuration, weights are stored in\na variety of np.memmaps on disk or to a particular device.""" , font_size=2_4 , ) step_a.move_to([2, 2, 0] ) _SCREAMING_SNAKE_CASE = [meta_mem.copy() for i in range(6 )] _SCREAMING_SNAKE_CASE = [meta_mem.copy() for i in range(6 )] _SCREAMING_SNAKE_CASE = VGroup(*A_ ).arrange(A_ , buff=0 ) _SCREAMING_SNAKE_CASE = VGroup(*A_ ).arrange(A_ , buff=0 ) _SCREAMING_SNAKE_CASE = VGroup(A_ , A_ ).arrange(A_ , buff=0 ) _SCREAMING_SNAKE_CASE = Text("Disk" , font_size=2_4 ) _SCREAMING_SNAKE_CASE = Group(A_ , A_ ).arrange(A_ , buff=0.5 , aligned_edge=A_ ) disk.move_to([-4.0, -1.2_5, 0] ) self.play(Write(A_ , run_time=3 ) , Write(A_ , run_time=1 ) , Create(A_ , run_time=1 ) ) _SCREAMING_SNAKE_CASE = [] for i, rect in enumerate(A_ ): _SCREAMING_SNAKE_CASE = rect.copy() target.generate_target() target.target.move_to(disk_left_col_base[i] ).scale(0.5 ) animations.append(MoveToTarget(A_ , run_time=1.5 ) ) self.play(*A_ ) self.play(FadeOut(A_ ) ) _SCREAMING_SNAKE_CASE = MarkupText(F"""Then, the checkpoint is removed from memory\nthrough garbage collection.""" , font_size=2_4 ) step_a.move_to([2, 2, 0] ) self.play(Write(A_ , run_time=3 ) ) self.play( FadeOut(A_ , A_ , *A_ , *A_ ) , ) self.wait()
418
def _SCREAMING_SNAKE_CASE ( lowercase : str , lowercase : str ): '''simple docstring''' if len(lowercase ) != len(lowercase ): raise ValueError('String lengths must match!' ) lowerCamelCase_ = 0 for chara, chara in zip(lowercase , lowercase ): if chara != chara: count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
70
0
'''simple docstring''' import argparse import os import torch from transformers import FlavaConfig, FlavaForPreTraining from transformers.models.flava.convert_dalle_to_flava_codebook import convert_dalle_checkpoint def _A ( UpperCAmelCase ): '''simple docstring''' return sum(param.float().sum() if 'encoder.embeddings' not in key else 0 for key, param in state_dict.items() ) def _A ( UpperCAmelCase ,UpperCAmelCase ): '''simple docstring''' A__ = {} for key, value in state_dict.items(): if "text_encoder.embeddings" in key or "image_encoder.embeddings" in key: continue A__ = key.replace('heads.cmd.mim_head.cls.predictions' ,'mmm_image_head' ) A__ = key.replace('heads.cmd.mlm_head.cls.predictions' ,'mmm_text_head' ) A__ = key.replace('heads.cmd.itm_head.cls' ,'itm_head' ) A__ = key.replace('heads.cmd.itm_head.pooler' ,'itm_head.pooler' ) A__ = key.replace('heads.cmd.clip_head.logit_scale' ,'flava.logit_scale' ) A__ = key.replace('heads.fairseq_mlm.cls.predictions' ,'mlm_head' ) A__ = key.replace('heads.imagenet.mim_head.cls.predictions' ,'mim_head' ) A__ = key.replace('mm_text_projection' ,'flava.text_to_mm_projection' ) A__ = key.replace('mm_image_projection' ,'flava.image_to_mm_projection' ) A__ = key.replace('image_encoder.module' ,'flava.image_model' ) A__ = key.replace('text_encoder.module' ,'flava.text_model' ) A__ = key.replace('mm_encoder.module.encoder.cls_token' ,'flava.multimodal_model.cls_token' ) A__ = key.replace('mm_encoder.module' ,'flava.multimodal_model' ) A__ = key.replace('text_projection' ,'flava.text_projection' ) A__ = key.replace('image_projection' ,'flava.image_projection' ) A__ = value.float() for key, value in codebook_state_dict.items(): A__ = value return upgrade @torch.no_grad() def _A ( UpperCAmelCase ,UpperCAmelCase ,UpperCAmelCase ,UpperCAmelCase=None ): '''simple docstring''' if config_path is not None: A__ = FlavaConfig.from_pretrained(UpperCAmelCase ) else: A__ = FlavaConfig() A__ = FlavaForPreTraining(UpperCAmelCase ).eval() A__ = convert_dalle_checkpoint(UpperCAmelCase ,UpperCAmelCase ,save_checkpoint=UpperCAmelCase ) if os.path.exists(UpperCAmelCase ): A__ = torch.load(UpperCAmelCase ,map_location='cpu' ) else: A__ = torch.hub.load_state_dict_from_url(UpperCAmelCase ,map_location='cpu' ) A__ = upgrade_state_dict(UpperCAmelCase ,UpperCAmelCase ) hf_model.load_state_dict(UpperCAmelCase ) A__ = hf_model.state_dict() A__ = count_parameters(UpperCAmelCase ) A__ = count_parameters(UpperCAmelCase ) + count_parameters(UpperCAmelCase ) assert torch.allclose(UpperCAmelCase ,UpperCAmelCase ,atol=1e-3 ) hf_model.save_pretrained(UpperCAmelCase ) if __name__ == "__main__": lowerCAmelCase_ = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to flava checkpoint''') parser.add_argument('''--codebook_path''', default=None, type=str, help='''Path to flava codebook checkpoint''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') lowerCAmelCase_ = parser.parse_args() convert_flava_checkpoint(args.checkpoint_path, args.codebook_path, args.pytorch_dump_folder_path, args.config_path)
531
def _SCREAMING_SNAKE_CASE ( lowercase : int = 10 ): '''simple docstring''' if not isinstance(lowercase , lowercase ) or n < 0: raise ValueError('Invalid input' ) lowerCamelCase_ = 10**n lowerCamelCase_ = 2_84_33 * (pow(2 , 7_83_04_57 , lowercase )) + 1 return str(number % modulus ) if __name__ == "__main__": from doctest import testmod testmod() print(F"""{solution(10) = }""")
70
0
'''simple docstring''' import math def _lowerCAmelCase ( _UpperCamelCase : int ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE =0 _SCREAMING_SNAKE_CASE =0 while num > 0: _SCREAMING_SNAKE_CASE =num % 8 _SCREAMING_SNAKE_CASE =octal + (remainder * math.floor(math.pow(10 , _UpperCamelCase ) )) counter += 1 _SCREAMING_SNAKE_CASE =math.floor(num / 8 ) # basically /= 8 without remainder if any # This formatting removes trailing '.0' from `octal`. return f"0o{int(_UpperCamelCase )}" def _lowerCAmelCase ( ) -> int: """simple docstring""" print('\n2 in octal is:' ) print(decimal_to_octal(2 ) ) # = 2 print('\n8 in octal is:' ) print(decimal_to_octal(8 ) ) # = 10 print('\n65 in octal is:' ) print(decimal_to_octal(65 ) ) # = 101 print('\n216 in octal is:' ) print(decimal_to_octal(2_16 ) ) # = 330 print('\n512 in octal is:' ) print(decimal_to_octal(5_12 ) ) # = 1000 print('\n' ) if __name__ == "__main__": main()
405
from maths.prime_check import is_prime def _SCREAMING_SNAKE_CASE ( lowercase : int ): '''simple docstring''' if not isinstance(lowercase , lowercase ): lowerCamelCase_ = f"""Input value of [number={number}] must be an integer""" raise TypeError(lowercase ) if is_prime(lowercase ) and is_prime(number + 2 ): return number + 2 else: return -1 if __name__ == "__main__": import doctest doctest.testmod()
70
0
'''simple docstring''' import json import os import unittest from transformers.models.gptsan_japanese.tokenization_gptsan_japanese import ( VOCAB_FILES_NAMES, GPTSanJapaneseTokenizer, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _snake_case ( a_ , unittest.TestCase ): SCREAMING_SNAKE_CASE : Optional[Any] = GPTSanJapaneseTokenizer SCREAMING_SNAKE_CASE : Union[str, Any] = False SCREAMING_SNAKE_CASE : Union[str, Any] = {'''do_clean_text''': False, '''add_prefix_space''': False} def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' super().setUp() # fmt: off lowerCAmelCase = ['こん', 'こんに', 'にちは', 'ばんは', '世界,㔺界', '、', '。', '<BR>', '<SP>', '<TAB>', '<URL>', '<EMAIL>', '<TEL>', '<DATE>', '<PRICE>', '<BLOCK>', '<KIGOU>', '<U2000U2BFF>', '<|emoji1|>', '<unk>', '<|bagoftoken|>', '<|endoftext|>'] # fmt: on lowerCAmelCase = {'emoji': {'\ud83d\ude00': '<|emoji1|>'}, 'emoji_inv': {'<|emoji1|>': '\ud83d\ude00'}} # 😀 lowerCAmelCase = {'unk_token': '<unk>'} lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['emoji_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) with open(self.emoji_file , 'w' ) as emoji_writer: emoji_writer.write(json.dumps(A_ ) ) def _SCREAMING_SNAKE_CASE ( self , **_SCREAMING_SNAKE_CASE ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return GPTSanJapaneseTokenizer.from_pretrained(self.tmpdirname , **A_ ) def _SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE ): '''simple docstring''' lowerCAmelCase = 'こんにちは、世界。 \nこんばんは、㔺界。😀' lowerCAmelCase = 'こんにちは、世界。 \nこんばんは、世界。😀' return input_text, output_text def _SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE ): '''simple docstring''' lowerCAmelCase , lowerCAmelCase = self.get_input_output_texts(A_ ) lowerCAmelCase = tokenizer.encode(A_ , add_special_tokens=A_ ) lowerCAmelCase = tokenizer.decode(A_ , clean_up_tokenization_spaces=A_ ) return text, ids def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' pass # TODO add if relevant def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' pass # TODO add if relevant def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' pass # TODO add if relevant def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowerCAmelCase = self.get_tokenizer() # Testing tokenization lowerCAmelCase = 'こんにちは、世界。 こんばんは、㔺界。' lowerCAmelCase = ['こん', 'にちは', '、', '世界', '。', '<SP>', 'こん', 'ばんは', '、', '㔺界', '。'] lowerCAmelCase = tokenizer.tokenize(A_ ) self.assertListEqual(A_ , A_ ) # Testing conversion to ids without special tokens lowerCAmelCase = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6] lowerCAmelCase = tokenizer.convert_tokens_to_ids(A_ ) self.assertListEqual(A_ , A_ ) # Testing conversion to ids with special tokens lowerCAmelCase = tokens + [tokenizer.unk_token] lowerCAmelCase = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6, 19] lowerCAmelCase = tokenizer.convert_tokens_to_ids(A_ ) self.assertListEqual(A_ , A_ ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowerCAmelCase = self.get_tokenizer() # Testing tokenization lowerCAmelCase = 'こんにちは、<|bagoftoken|>世界。こんばんは、<|bagoftoken|>㔺界。' lowerCAmelCase = 'こんにちは、、、、世界。こんばんは、、、、世界。' lowerCAmelCase = tokenizer.encode(A_ ) lowerCAmelCase = tokenizer.decode(A_ ) self.assertEqual(A_ , A_ ) @slow def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowerCAmelCase = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) # Testing tokenization lowerCAmelCase = 'こんにちは、世界。' lowerCAmelCase = 'こんばんは、㔺界。😀' lowerCAmelCase = 'こんにちは、世界。こんばんは、世界。😀' lowerCAmelCase = tokenizer.encode(prefix_text + input_text ) lowerCAmelCase = tokenizer.encode('' , prefix_text=prefix_text + input_text ) lowerCAmelCase = tokenizer.encode(A_ , prefix_text=A_ ) lowerCAmelCase = tokenizer.decode(A_ ) lowerCAmelCase = tokenizer.decode(A_ ) lowerCAmelCase = tokenizer.decode(A_ ) self.assertEqual(A_ , A_ ) self.assertEqual(A_ , A_ ) self.assertEqual(A_ , A_ ) @slow def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowerCAmelCase = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) # Testing tokenization lowerCAmelCase = 'こんにちは、世界。' lowerCAmelCase = 'こんばんは、㔺界。😀' lowerCAmelCase = len(tokenizer.encode(A_ ) ) - 2 lowerCAmelCase = len(tokenizer.encode(A_ ) ) - 2 lowerCAmelCase = [1] + [0] * (len_prefix + len_text + 1) lowerCAmelCase = [1] * (len_prefix + len_text + 1) + [0] lowerCAmelCase = [1] + [1] * (len_prefix) + [0] * (len_text + 1) lowerCAmelCase = tokenizer(prefix_text + input_text ).token_type_ids lowerCAmelCase = tokenizer('' , prefix_text=prefix_text + input_text ).token_type_ids lowerCAmelCase = tokenizer(A_ , prefix_text=A_ ).token_type_ids self.assertListEqual(A_ , A_ ) self.assertListEqual(A_ , A_ ) self.assertListEqual(A_ , A_ ) @slow def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowerCAmelCase = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) lowerCAmelCase = tokenizer.encode('あンいワ' ) lowerCAmelCase = tokenizer.encode('' , prefix_text='あンいワ' ) lowerCAmelCase = tokenizer.encode('いワ' , prefix_text='あン' ) self.assertEqual(tokenizer.decode(A_ ) , tokenizer.decode(A_ ) ) self.assertEqual(tokenizer.decode(A_ ) , tokenizer.decode(A_ ) ) self.assertNotEqual(A_ , A_ ) self.assertNotEqual(A_ , A_ ) self.assertEqual(x_token_a[1] , x_token_a[-1] ) # SEG token self.assertEqual(x_token_a[1] , x_token_a[3] ) # SEG token @slow def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowerCAmelCase = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) lowerCAmelCase = [['武田信玄', 'は、'], ['織田信長', 'の配下の、']] lowerCAmelCase = tokenizer(A_ , padding=A_ ) lowerCAmelCase = tokenizer.batch_encode_plus(A_ , padding=A_ ) # fmt: off lowerCAmelCase = [[3_59_93, 86_40, 2_59_48, 3_59_98, 3_06_47, 3_56_75, 3_59_99, 3_59_99], [3_59_93, 1_03_82, 98_68, 3_59_98, 3_06_46, 94_59, 3_06_46, 3_56_75]] lowerCAmelCase = [[1, 1, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0]] lowerCAmelCase = [[1, 1, 1, 1, 1, 1, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1]] # fmt: on self.assertListEqual(x_token.input_ids , A_ ) self.assertListEqual(x_token.token_type_ids , A_ ) self.assertListEqual(x_token.attention_mask , A_ ) self.assertListEqual(x_token_a.input_ids , A_ ) self.assertListEqual(x_token_a.token_type_ids , A_ ) self.assertListEqual(x_token_a.attention_mask , A_ ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' pass def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' pass
284
# Algorithm for the pigeonhole sorting def _SCREAMING_SNAKE_CASE ( lowercase : str ): '''simple docstring''' lowerCamelCase_ = min(lowercase ) # min() finds the minimum value lowerCamelCase_ = max(lowercase ) # max() finds the maximum value lowerCamelCase_ = max_val - min_val + 1 # size is difference of max and min values plus one # list of pigeonholes of size equal to the variable size lowerCamelCase_ = [0] * size # Populate the pigeonholes. for x in a: assert isinstance(lowercase , lowercase ), "integers only please" holes[x - min_val] += 1 # Putting the elements back into the array in an order. lowerCamelCase_ = 0 for count in range(lowercase ): while holes[count] > 0: holes[count] -= 1 lowerCamelCase_ = count + min_val i += 1 def _SCREAMING_SNAKE_CASE ( ): '''simple docstring''' lowerCamelCase_ = [8, 3, 2, 7, 4, 6, 8] pigeonhole_sort(lowercase ) print('Sorted order is:' , ' '.join(lowercase ) ) if __name__ == "__main__": main()
70
0
'''simple docstring''' import warnings from ...utils import logging from .image_processing_perceiver import PerceiverImageProcessor __lowerCAmelCase = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE (__A ): """simple docstring""" def __init__( self , *UpperCamelCase__ , **UpperCamelCase__ ): """simple docstring""" warnings.warn( 'The class PerceiverFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use PerceiverImageProcessor instead.' , A_ , ) super().__init__(*A_ , **A_ )
536
import os import unittest from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, BertTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class A( UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = BertTokenizer UpperCamelCase = BertTokenizerFast UpperCamelCase = True UpperCamelCase = True UpperCamelCase = filter_non_english def a__ ( self : List[Any] ) -> int: """simple docstring""" super().setUp() lowerCamelCase_ = [ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] lowerCamelCase_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def a__ ( self : Tuple , A_ : List[str] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = 'UNwant\u00E9d,running' lowerCamelCase_ = 'unwanted, running' return input_text, output_text def a__ ( self : Any ) -> Tuple: """simple docstring""" lowerCamelCase_ = self.tokenizer_class(self.vocab_file ) lowerCamelCase_ = tokenizer.tokenize('UNwant\u00E9d,running' ) self.assertListEqual(A_ , ['un', '##want', '##ed', ',', 'runn', '##ing'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A_ ) , [9, 6, 7, 12, 10, 11] ) def a__ ( self : Tuple ) -> List[str]: """simple docstring""" if not self.test_rust_tokenizer: return lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_rust_tokenizer() lowerCamelCase_ = 'UNwant\u00E9d,running' lowerCamelCase_ = tokenizer.tokenize(A_ ) lowerCamelCase_ = rust_tokenizer.tokenize(A_ ) self.assertListEqual(A_ , A_ ) lowerCamelCase_ = tokenizer.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = rust_tokenizer.encode(A_ , add_special_tokens=A_ ) self.assertListEqual(A_ , A_ ) lowerCamelCase_ = self.get_rust_tokenizer() lowerCamelCase_ = tokenizer.encode(A_ ) lowerCamelCase_ = rust_tokenizer.encode(A_ ) self.assertListEqual(A_ , A_ ) # With lower casing lowerCamelCase_ = self.get_tokenizer(do_lower_case=A_ ) lowerCamelCase_ = self.get_rust_tokenizer(do_lower_case=A_ ) lowerCamelCase_ = 'UNwant\u00E9d,running' lowerCamelCase_ = tokenizer.tokenize(A_ ) lowerCamelCase_ = rust_tokenizer.tokenize(A_ ) self.assertListEqual(A_ , A_ ) lowerCamelCase_ = tokenizer.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = rust_tokenizer.encode(A_ , add_special_tokens=A_ ) self.assertListEqual(A_ , A_ ) lowerCamelCase_ = self.get_rust_tokenizer() lowerCamelCase_ = tokenizer.encode(A_ ) lowerCamelCase_ = rust_tokenizer.encode(A_ ) self.assertListEqual(A_ , A_ ) def a__ ( self : Any ) -> Dict: """simple docstring""" lowerCamelCase_ = BasicTokenizer() self.assertListEqual(tokenizer.tokenize('ah\u535A\u63A8zz' ) , ['ah', '\u535A', '\u63A8', 'zz'] ) def a__ ( self : Dict ) -> int: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? ' ) , ['hello', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['hello'] ) def a__ ( self : str ) -> int: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ , strip_accents=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['hällo', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['h\u00E9llo'] ) def a__ ( self : Optional[int] ) -> Any: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ , strip_accents=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['hallo', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['hello'] ) def a__ ( self : Tuple ) -> str: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['hallo', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['hello'] ) def a__ ( self : int ) -> List[Any]: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? ' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?'] ) def a__ ( self : str ) -> List[str]: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ , strip_accents=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['HäLLo', '!', 'how', 'Are', 'yoU', '?'] ) def a__ ( self : Dict ) -> Any: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ , strip_accents=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['HaLLo', '!', 'how', 'Are', 'yoU', '?'] ) def a__ ( self : int ) -> Any: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ , never_split=['[UNK]'] ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? [UNK]' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?', '[UNK]'] ) def a__ ( self : List[Any] ) -> int: """simple docstring""" lowerCamelCase_ = BasicTokenizer() lowerCamelCase_ = 'a\n\'ll !!to?\'d of, can\'t.' lowerCamelCase_ = ['a', '\'', 'll', '!', '!', 'to', '?', '\'', 'd', 'of', ',', 'can', '\'', 't', '.'] self.assertListEqual(tokenizer.tokenize(A_ ) , A_ ) def a__ ( self : Optional[int] ) -> Dict: """simple docstring""" lowerCamelCase_ = ['[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing'] lowerCamelCase_ = {} for i, token in enumerate(A_ ): lowerCamelCase_ = i lowerCamelCase_ = WordpieceTokenizer(vocab=A_ , unk_token='[UNK]' ) self.assertListEqual(tokenizer.tokenize('' ) , [] ) self.assertListEqual(tokenizer.tokenize('unwanted running' ) , ['un', '##want', '##ed', 'runn', '##ing'] ) self.assertListEqual(tokenizer.tokenize('unwantedX running' ) , ['[UNK]', 'runn', '##ing'] ) def a__ ( self : Optional[Any] ) -> str: """simple docstring""" self.assertTrue(_is_whitespace(' ' ) ) self.assertTrue(_is_whitespace('\t' ) ) self.assertTrue(_is_whitespace('\r' ) ) self.assertTrue(_is_whitespace('\n' ) ) self.assertTrue(_is_whitespace('\u00A0' ) ) self.assertFalse(_is_whitespace('A' ) ) self.assertFalse(_is_whitespace('-' ) ) def a__ ( self : List[Any] ) -> int: """simple docstring""" self.assertTrue(_is_control('\u0005' ) ) self.assertFalse(_is_control('A' ) ) self.assertFalse(_is_control(' ' ) ) self.assertFalse(_is_control('\t' ) ) self.assertFalse(_is_control('\r' ) ) def a__ ( self : List[str] ) -> Optional[Any]: """simple docstring""" self.assertTrue(_is_punctuation('-' ) ) self.assertTrue(_is_punctuation('$' ) ) self.assertTrue(_is_punctuation('`' ) ) self.assertTrue(_is_punctuation('.' ) ) self.assertFalse(_is_punctuation('A' ) ) self.assertFalse(_is_punctuation(' ' ) ) def a__ ( self : int ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_rust_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(A_ ) for t in ['Test', '\xad', 'test']] , [['[UNK]'], [], ['[UNK]']] ) self.assertListEqual( [rust_tokenizer.tokenize(A_ ) for t in ['Test', '\xad', 'test']] , [['[UNK]'], [], ['[UNK]']] ) @slow def a__ ( self : Any ) -> int: """simple docstring""" lowerCamelCase_ = self.tokenizer_class.from_pretrained('bert-base-uncased' ) lowerCamelCase_ = tokenizer.encode('sequence builders' , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer.encode('multi-sequence build' , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer.build_inputs_with_special_tokens(A_ ) lowerCamelCase_ = tokenizer.build_inputs_with_special_tokens(A_ , A_ ) assert encoded_sentence == [101] + text + [102] assert encoded_pair == [101] + text + [102] + text_a + [102] def a__ ( self : str ) -> str: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCamelCase_ = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) lowerCamelCase_ = f"""A, naïve {tokenizer_r.mask_token} AllenNLP sentence.""" lowerCamelCase_ = tokenizer_r.encode_plus( A_ , return_attention_mask=A_ , return_token_type_ids=A_ , return_offsets_mapping=A_ , add_special_tokens=A_ , ) lowerCamelCase_ = tokenizer_r.do_lower_case if hasattr(A_ , 'do_lower_case' ) else False lowerCamelCase_ = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), 'A'), ((1, 2), ','), ((3, 5), 'na'), ((5, 6), '##ï'), ((6, 8), '##ve'), ((9, 15), tokenizer_r.mask_token), ((16, 21), 'Allen'), ((21, 23), '##NL'), ((23, 24), '##P'), ((25, 33), 'sentence'), ((33, 34), '.'), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), 'a'), ((1, 2), ','), ((3, 8), 'naive'), ((9, 15), tokenizer_r.mask_token), ((16, 21), 'allen'), ((21, 23), '##nl'), ((23, 24), '##p'), ((25, 33), 'sentence'), ((33, 34), '.'), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens['input_ids'] ) ) self.assertEqual([e[0] for e in expected_results] , tokens['offset_mapping'] ) def a__ ( self : List[Any] ) -> Dict: """simple docstring""" lowerCamelCase_ = ['的', '人', '有'] lowerCamelCase_ = ''.join(A_ ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCamelCase_ = True lowerCamelCase_ = self.tokenizer_class.from_pretrained(A_ , **A_ ) lowerCamelCase_ = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) lowerCamelCase_ = tokenizer_p.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer_r.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer_r.convert_ids_to_tokens(A_ ) lowerCamelCase_ = tokenizer_p.convert_ids_to_tokens(A_ ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(A_ , A_ ) self.assertListEqual(A_ , A_ ) lowerCamelCase_ = False lowerCamelCase_ = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) lowerCamelCase_ = self.tokenizer_class.from_pretrained(A_ , **A_ ) lowerCamelCase_ = tokenizer_r.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer_p.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer_r.convert_ids_to_tokens(A_ ) lowerCamelCase_ = tokenizer_p.convert_ids_to_tokens(A_ ) # it is expected that only the first Chinese character is not preceded by "##". lowerCamelCase_ = [ f"""##{token}""" if idx != 0 else token for idx, token in enumerate(A_ ) ] self.assertListEqual(A_ , A_ ) self.assertListEqual(A_ , A_ )
70
0
import unittest from transformers import XLMConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMWithLMHeadModel, ) from transformers.models.xlm.modeling_xlm import XLM_PRETRAINED_MODEL_ARCHIVE_LIST class __lowerCAmelCase : """simple docstring""" def __init__( self : Dict , _snake_case : str , _snake_case : Any=13 , _snake_case : Any=7 , _snake_case : List[Any]=True , _snake_case : List[Any]=True , _snake_case : List[str]=True , _snake_case : Any=True , _snake_case : Union[str, Any]=True , _snake_case : List[str]=False , _snake_case : Any=False , _snake_case : List[Any]=False , _snake_case : Optional[int]=2 , _snake_case : Any=99 , _snake_case : List[Any]=0 , _snake_case : List[Any]=32 , _snake_case : List[Any]=5 , _snake_case : Optional[int]=4 , _snake_case : List[Any]=0.1 , _snake_case : Optional[int]=0.1 , _snake_case : Tuple=5_12 , _snake_case : Union[str, Any]=2 , _snake_case : List[Any]=0.02 , _snake_case : int=2 , _snake_case : str=4 , _snake_case : Optional[Any]="last" , _snake_case : Tuple=True , _snake_case : Any=None , _snake_case : List[Any]=0 , ): """simple docstring""" A__ = parent A__ = batch_size A__ = seq_length A__ = is_training A__ = use_input_lengths A__ = use_token_type_ids A__ = use_labels A__ = gelu_activation A__ = sinusoidal_embeddings A__ = causal A__ = asm A__ = n_langs A__ = vocab_size A__ = n_special A__ = hidden_size A__ = num_hidden_layers A__ = num_attention_heads A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = max_position_embeddings A__ = type_sequence_label_size A__ = initializer_range A__ = num_labels A__ = num_choices A__ = summary_type A__ = use_proj A__ = scope A__ = bos_token_id def _a ( self : int ): """simple docstring""" A__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ = random_attention_mask([self.batch_size, self.seq_length] ) A__ = None if self.use_input_lengths: A__ = ( ids_tensor([self.batch_size] , vocab_size=2 ) + self.seq_length - 2 ) # small variation of seq_length A__ = None if self.use_token_type_ids: A__ = ids_tensor([self.batch_size, self.seq_length] , self.n_langs ) A__ = None A__ = None A__ = None if self.use_labels: A__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A__ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A__ = ids_tensor([self.batch_size] , 2 ).float() A__ = ids_tensor([self.batch_size] , self.num_choices ) A__ = self.get_config() return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def _a ( self : Tuple ): """simple docstring""" return XLMConfig( vocab_size=self.vocab_size , n_special=self.n_special , emb_dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , gelu_activation=self.gelu_activation , sinusoidal_embeddings=self.sinusoidal_embeddings , asm=self.asm , causal=self.causal , n_langs=self.n_langs , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , summary_type=self.summary_type , use_proj=self.use_proj , num_labels=self.num_labels , bos_token_id=self.bos_token_id , ) def _a ( self : List[str] , _snake_case : List[str] , _snake_case : List[Any] , _snake_case : Union[str, Any] , _snake_case : Optional[int] , _snake_case : Any , _snake_case : str , _snake_case : str , _snake_case : List[str] , _snake_case : List[Any] , ): """simple docstring""" A__ = XLMModel(config=A_ ) model.to(A_ ) model.eval() A__ = model(A_ , lengths=A_ , langs=A_ ) A__ = model(A_ , langs=A_ ) A__ = model(A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a ( self : Optional[Any] , _snake_case : Any , _snake_case : Any , _snake_case : Optional[int] , _snake_case : int , _snake_case : List[Any] , _snake_case : Tuple , _snake_case : Any , _snake_case : str , _snake_case : List[Any] , ): """simple docstring""" A__ = XLMWithLMHeadModel(A_ ) model.to(A_ ) model.eval() A__ = model(A_ , token_type_ids=A_ , labels=A_ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a ( self : List[str] , _snake_case : Tuple , _snake_case : str , _snake_case : List[Any] , _snake_case : List[str] , _snake_case : Any , _snake_case : List[Any] , _snake_case : List[str] , _snake_case : Any , _snake_case : List[Any] , ): """simple docstring""" A__ = XLMForQuestionAnsweringSimple(A_ ) model.to(A_ ) model.eval() A__ = model(A_ ) A__ = model(A_ , start_positions=A_ , end_positions=A_ ) A__ = outputs self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _a ( self : List[str] , _snake_case : Union[str, Any] , _snake_case : Tuple , _snake_case : int , _snake_case : Any , _snake_case : int , _snake_case : str , _snake_case : Union[str, Any] , _snake_case : Union[str, Any] , _snake_case : Union[str, Any] , ): """simple docstring""" A__ = XLMForQuestionAnswering(A_ ) model.to(A_ ) model.eval() A__ = model(A_ ) A__ = model( A_ , start_positions=A_ , end_positions=A_ , cls_index=A_ , is_impossible=A_ , p_mask=A_ , ) A__ = model( A_ , start_positions=A_ , end_positions=A_ , cls_index=A_ , is_impossible=A_ , ) ((A__ ) , ) = result_with_labels.to_tuple() A__ = model(A_ , start_positions=A_ , end_positions=A_ ) ((A__ ) , ) = result_with_labels.to_tuple() self.parent.assertEqual(result_with_labels.loss.shape , () ) self.parent.assertEqual(result.start_top_log_probs.shape , (self.batch_size, model.config.start_n_top) ) self.parent.assertEqual(result.start_top_index.shape , (self.batch_size, model.config.start_n_top) ) self.parent.assertEqual( result.end_top_log_probs.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top) ) self.parent.assertEqual( result.end_top_index.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top) ) self.parent.assertEqual(result.cls_logits.shape , (self.batch_size,) ) def _a ( self : Dict , _snake_case : List[str] , _snake_case : str , _snake_case : List[Any] , _snake_case : Union[str, Any] , _snake_case : Optional[int] , _snake_case : Dict , _snake_case : Tuple , _snake_case : Union[str, Any] , _snake_case : int , ): """simple docstring""" A__ = XLMForSequenceClassification(A_ ) model.to(A_ ) model.eval() A__ = model(A_ ) A__ = model(A_ , labels=A_ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _a ( self : Optional[Any] , _snake_case : str , _snake_case : List[str] , _snake_case : str , _snake_case : str , _snake_case : Any , _snake_case : Union[str, Any] , _snake_case : Tuple , _snake_case : Optional[Any] , _snake_case : Optional[int] , ): """simple docstring""" A__ = self.num_labels A__ = XLMForTokenClassification(A_ ) model.to(A_ ) model.eval() A__ = model(A_ , attention_mask=A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _a ( self : Union[str, Any] , _snake_case : int , _snake_case : Union[str, Any] , _snake_case : int , _snake_case : List[Any] , _snake_case : List[str] , _snake_case : List[str] , _snake_case : List[Any] , _snake_case : str , _snake_case : str , ): """simple docstring""" A__ = self.num_choices A__ = XLMForMultipleChoice(config=A_ ) model.to(A_ ) model.eval() A__ = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A__ = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A__ = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A__ = model( A_ , attention_mask=A_ , token_type_ids=A_ , labels=A_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _a ( self : int ): """simple docstring""" A__ = self.prepare_config_and_inputs() ( ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ) = config_and_inputs A__ = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'lengths': input_lengths} return config, inputs_dict @require_torch class __lowerCAmelCase ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): """simple docstring""" A__ : Tuple = ( ( XLMModel, XLMWithLMHeadModel, XLMForQuestionAnswering, XLMForSequenceClassification, XLMForQuestionAnsweringSimple, XLMForTokenClassification, XLMForMultipleChoice, ) if is_torch_available() else () ) A__ : Union[str, Any] = ( (XLMWithLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Check other models whether language generation is also applicable A__ : List[Any] = ( { "feature-extraction": XLMModel, "fill-mask": XLMWithLMHeadModel, "question-answering": XLMForQuestionAnsweringSimple, "text-classification": XLMForSequenceClassification, "text-generation": XLMWithLMHeadModel, "token-classification": XLMForTokenClassification, "zero-shot": XLMForSequenceClassification, } if is_torch_available() else {} ) def _a ( self : Optional[Any] , _snake_case : Optional[Any] , _snake_case : Optional[Any] , _snake_case : List[Any] , _snake_case : Dict , _snake_case : List[str] ): """simple docstring""" if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith('Fast' ) ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def _a ( self : List[str] , _snake_case : List[Any] , _snake_case : List[str] , _snake_case : Tuple=False ): """simple docstring""" A__ = super()._prepare_for_class(A_ , A_ , return_labels=A_ ) if return_labels: if model_class.__name__ == "XLMForQuestionAnswering": A__ = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=A_ ) A__ = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=A_ ) return inputs_dict def _a ( self : Tuple ): """simple docstring""" A__ = XLMModelTester(self ) A__ = ConfigTester(self , config_class=A_ , emb_dim=37 ) def _a ( self : int ): """simple docstring""" self.config_tester.run_common_tests() def _a ( self : List[str] ): """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_model(*A_ ) def _a ( self : Optional[Any] ): """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_lm_head(*A_ ) def _a ( self : Tuple ): """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_simple_qa(*A_ ) def _a ( self : List[Any] ): """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_qa(*A_ ) def _a ( self : str ): """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_sequence_classif(*A_ ) def _a ( self : Optional[Any] ): """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_token_classif(*A_ ) def _a ( self : int ): """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_for_multiple_choice(*A_ ) def _a ( self : List[str] , _snake_case : Tuple , _snake_case : List[str] , _snake_case : Optional[int] , _snake_case : Optional[Any] , _snake_case : str , _snake_case : Tuple=False , _snake_case : Optional[int]=1 ): """simple docstring""" self.assertIsInstance(A_ , A_ ) self.assertListEqual( [isinstance(A_ , A_ ) for iter_attentions in attentions] , [True] * len(A_ ) ) self.assertEqual(len(A_ ) , (max_length - min_length) * num_beam_groups ) for idx, iter_attentions in enumerate(A_ ): # adds PAD dummy token A__ = min_length + idx + 1 A__ = min_length + idx + 1 A__ = ( batch_size * num_beam_groups, config.num_attention_heads, tgt_len, src_len, ) # check attn size self.assertListEqual( [layer_attention.shape for layer_attention in iter_attentions] , [expected_shape] * len(A_ ) ) def _a ( self : Union[str, Any] , _snake_case : Tuple , _snake_case : List[str] , _snake_case : str , _snake_case : Any , _snake_case : str , _snake_case : Any=False , _snake_case : List[str]=1 ): """simple docstring""" self.assertIsInstance(A_ , A_ ) self.assertListEqual( [isinstance(A_ , A_ ) for iter_hidden_states in hidden_states] , [True] * len(A_ ) , ) self.assertEqual(len(A_ ) , (max_length - min_length) * num_beam_groups ) for idx, iter_hidden_states in enumerate(A_ ): # adds PAD dummy token A__ = min_length + idx + 1 A__ = (batch_size * num_beam_groups, seq_len, config.hidden_size) # check hidden size self.assertListEqual( [layer_hidden_states.shape for layer_hidden_states in iter_hidden_states] , [expected_shape] * len(A_ ) , ) pass @slow def _a ( self : Tuple ): """simple docstring""" for model_name in XLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ = XLMModel.from_pretrained(A_ ) self.assertIsNotNone(A_ ) @require_torch class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def _a ( self : str ): """simple docstring""" A__ = XLMWithLMHeadModel.from_pretrained('xlm-mlm-en-2048' ) model.to(A_ ) A__ = torch.tensor([[14, 4_47]] , dtype=torch.long , device=A_ ) # the president A__ = [ 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, ] # the president the president the president the president the president the president the president the president the president the president # TODO(PVP): this and other input_ids I tried for generation give pretty bad results. Not sure why. Model might just not be made for auto-regressive inference A__ = model.generate(A_ , do_sample=A_ ) self.assertListEqual(output_ids[0].cpu().numpy().tolist() , A_ )
9
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) lowerCamelCase : Dict = logging.get_logger(__name__) # pylint: disable=invalid-name lowerCamelCase : List[str] = "\n Examples:\n ```py\n >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline\n >>> import torch\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-prior\")\n >>> pipe_prior.to(\"cuda\")\n >>> prompt = \"red cat, 4k photo\"\n >>> out = pipe_prior(prompt)\n >>> image_emb = out.image_embeds\n >>> zero_image_emb = out.negative_image_embeds\n >>> pipe = KandinskyV22Pipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-decoder\")\n >>> pipe.to(\"cuda\")\n >>> image = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=50,\n ... ).images\n >>> image[0].save(\"cat.png\")\n ```\n" def _SCREAMING_SNAKE_CASE ( lowercase : Any , lowercase : str , lowercase : Any=8 ): '''simple docstring''' lowerCamelCase_ = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 lowerCamelCase_ = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class A( UpperCamelCase ): '''simple docstring''' def __init__( self : str , A_ : UNetaDConditionModel , A_ : DDPMScheduler , A_ : VQModel , ) -> List[str]: """simple docstring""" super().__init__() self.register_modules( unet=A_ , scheduler=A_ , movq=A_ , ) lowerCamelCase_ = 2 ** (len(self.movq.config.block_out_channels ) - 1) def a__ ( self : List[Any] , A_ : Tuple , A_ : Dict , A_ : List[Any] , A_ : int , A_ : Any , A_ : Tuple ) -> Any: """simple docstring""" if latents is None: lowerCamelCase_ = randn_tensor(A_ , generator=A_ , device=A_ , dtype=A_ ) else: if latents.shape != shape: raise ValueError(f"""Unexpected latents shape, got {latents.shape}, expected {shape}""" ) lowerCamelCase_ = latents.to(A_ ) lowerCamelCase_ = latents * scheduler.init_noise_sigma return latents def a__ ( self : int , A_ : str=0 ) -> Optional[int]: """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) lowerCamelCase_ = torch.device(f"""cuda:{gpu_id}""" ) lowerCamelCase_ = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(A_ , A_ ) def a__ ( self : Tuple , A_ : Union[str, Any]=0 ) -> Dict: """simple docstring""" if is_accelerate_available() and is_accelerate_version('>=' , '0.17.0.dev0' ): from accelerate import cpu_offload_with_hook else: raise ImportError('`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.' ) lowerCamelCase_ = torch.device(f"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to('cpu' , silence_dtype_warnings=A_ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) lowerCamelCase_ = None for cpu_offloaded_model in [self.unet, self.movq]: lowerCamelCase_ , lowerCamelCase_ = cpu_offload_with_hook(A_ , A_ , prev_module_hook=A_ ) # We'll offload the last model manually. lowerCamelCase_ = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def a__ ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" if not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(A_ , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(A_ ) def __call__( self : List[Any] , A_ : Union[torch.FloatTensor, List[torch.FloatTensor]] , A_ : Union[torch.FloatTensor, List[torch.FloatTensor]] , A_ : int = 512 , A_ : int = 512 , A_ : int = 100 , A_ : float = 4.0 , A_ : int = 1 , A_ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , A_ : Optional[torch.FloatTensor] = None , A_ : Optional[str] = "pil" , A_ : bool = True , ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self._execution_device lowerCamelCase_ = guidance_scale > 1.0 if isinstance(A_ , A_ ): lowerCamelCase_ = torch.cat(A_ , dim=0 ) lowerCamelCase_ = image_embeds.shape[0] * num_images_per_prompt if isinstance(A_ , A_ ): lowerCamelCase_ = torch.cat(A_ , dim=0 ) if do_classifier_free_guidance: lowerCamelCase_ = image_embeds.repeat_interleave(A_ , dim=0 ) lowerCamelCase_ = negative_image_embeds.repeat_interleave(A_ , dim=0 ) lowerCamelCase_ = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=A_ ) self.scheduler.set_timesteps(A_ , device=A_ ) lowerCamelCase_ = self.scheduler.timesteps lowerCamelCase_ = self.unet.config.in_channels lowerCamelCase_ , lowerCamelCase_ = downscale_height_and_width(A_ , A_ , self.movq_scale_factor ) # create initial latent lowerCamelCase_ = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , A_ , A_ , A_ , self.scheduler , ) for i, t in enumerate(self.progress_bar(A_ ) ): # expand the latents if we are doing classifier free guidance lowerCamelCase_ = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowerCamelCase_ = {'image_embeds': image_embeds} lowerCamelCase_ = self.unet( sample=A_ , timestep=A_ , encoder_hidden_states=A_ , added_cond_kwargs=A_ , return_dict=A_ , )[0] if do_classifier_free_guidance: lowerCamelCase_ , lowerCamelCase_ = noise_pred.split(latents.shape[1] , dim=1 ) lowerCamelCase_ , lowerCamelCase_ = noise_pred.chunk(2 ) lowerCamelCase_ , lowerCamelCase_ = variance_pred.chunk(2 ) lowerCamelCase_ = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) lowerCamelCase_ = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , 'variance_type' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): lowerCamelCase_ , lowerCamelCase_ = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 lowerCamelCase_ = self.scheduler.step( A_ , A_ , A_ , generator=A_ , )[0] # post-processing lowerCamelCase_ = self.movq.decode(A_ , force_not_quantize=A_ )['sample'] if output_type not in ["pt", "np", "pil"]: raise ValueError(f"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: lowerCamelCase_ = image * 0.5 + 0.5 lowerCamelCase_ = image.clamp(0 , 1 ) lowerCamelCase_ = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowerCamelCase_ = self.numpy_to_pil(A_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=A_ )
70
0
'''simple docstring''' from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) __lowerCAmelCase =logging.get_logger(__name__) # pylint: disable=invalid-name __lowerCAmelCase ="\n Examples:\n ```py\n >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline\n >>> import torch\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-prior\")\n >>> pipe_prior.to(\"cuda\")\n >>> prompt = \"red cat, 4k photo\"\n >>> out = pipe_prior(prompt)\n >>> image_emb = out.image_embeds\n >>> zero_image_emb = out.negative_image_embeds\n >>> pipe = KandinskyV22Pipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-decoder\")\n >>> pipe.to(\"cuda\")\n >>> image = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=50,\n ... ).images\n >>> image[0].save(\"cat.png\")\n ```\n" def a ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=8 ) -> int: """simple docstring""" a_ = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 a_ = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class _snake_case ( snake_case ): """simple docstring""" def __init__( self , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , ) -> List[str]: super().__init__() self.register_modules( unet=A_ , scheduler=A_ , movq=A_ , ) a_ = 2 ** (len(self.movq.config.block_out_channels ) - 1) def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) -> Any: if latents is None: a_ = randn_tensor(A_ , generator=A_ , device=A_ , dtype=A_ ) else: if latents.shape != shape: raise ValueError(F'''Unexpected latents shape, got {latents.shape}, expected {shape}''' ) a_ = latents.to(A_ ) a_ = latents * scheduler.init_noise_sigma return latents def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase__=0 ) -> Optional[int]: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) a_ = torch.device(F'''cuda:{gpu_id}''' ) a_ = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(A_ , A_ ) def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase__=0 ) -> Dict: if is_accelerate_available() and is_accelerate_version('>=' , '0.17.0.dev0' ): from accelerate import cpu_offload_with_hook else: raise ImportError('`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.' ) a_ = torch.device(F'''cuda:{gpu_id}''' ) if self.device.type != "cpu": self.to('cpu' , silence_dtype_warnings=A_ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) a_ = None for cpu_offloaded_model in [self.unet, self.movq]: a_ , a_ = cpu_offload_with_hook(A_ , A_ , prev_module_hook=A_ ) # We'll offload the last model manually. a_ = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __SCREAMING_SNAKE_CASE ( self ) -> Union[str, Any]: if not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(A_ , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(A_ ) def __call__( self , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = 512 , UpperCAmelCase__ = 512 , UpperCAmelCase__ = 100 , UpperCAmelCase__ = 4.0 , UpperCAmelCase__ = 1 , UpperCAmelCase__ = None , UpperCAmelCase__ = None , UpperCAmelCase__ = "pil" , UpperCAmelCase__ = True , ) -> Optional[int]: a_ = self._execution_device a_ = guidance_scale > 1.0 if isinstance(A_ , A_ ): a_ = torch.cat(A_ , dim=0 ) a_ = image_embeds.shape[0] * num_images_per_prompt if isinstance(A_ , A_ ): a_ = torch.cat(A_ , dim=0 ) if do_classifier_free_guidance: a_ = image_embeds.repeat_interleave(A_ , dim=0 ) a_ = negative_image_embeds.repeat_interleave(A_ , dim=0 ) a_ = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=A_ ) self.scheduler.set_timesteps(A_ , device=A_ ) a_ = self.scheduler.timesteps a_ = self.unet.config.in_channels a_ , a_ = downscale_height_and_width(A_ , A_ , self.movq_scale_factor ) # create initial latent a_ = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , A_ , A_ , A_ , self.scheduler , ) for i, t in enumerate(self.progress_bar(A_ ) ): # expand the latents if we are doing classifier free guidance a_ = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents a_ = {'image_embeds': image_embeds} a_ = self.unet( sample=A_ , timestep=A_ , encoder_hidden_states=A_ , added_cond_kwargs=A_ , return_dict=A_ , )[0] if do_classifier_free_guidance: a_ , a_ = noise_pred.split(latents.shape[1] , dim=1 ) a_ , a_ = noise_pred.chunk(2 ) a_ , a_ = variance_pred.chunk(2 ) a_ = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) a_ = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , 'variance_type' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): a_ , a_ = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 a_ = self.scheduler.step( A_ , A_ , A_ , generator=A_ , )[0] # post-processing a_ = self.movq.decode(A_ , force_not_quantize=A_ )['sample'] if output_type not in ["pt", "np", "pil"]: raise ValueError(F'''Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}''' ) if output_type in ["np", "pil"]: a_ = image * 0.5 + 0.5 a_ = image.clamp(0 , 1 ) a_ = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": a_ = self.numpy_to_pil(A_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=A_ )
697
from PIL import Image def _SCREAMING_SNAKE_CASE ( lowercase : Image ): '''simple docstring''' lowerCamelCase_ , lowerCamelCase_ = image.size lowerCamelCase_ = 0 lowerCamelCase_ = image.load() for i in range(lowercase ): for j in range(lowercase ): lowerCamelCase_ = pixels[j, i] mean += pixel mean //= width * height for j in range(lowercase ): for i in range(lowercase ): lowerCamelCase_ = 2_55 if pixels[i, j] > mean else 0 return image if __name__ == "__main__": lowerCamelCase : Optional[Any] = mean_threshold(Image.open("path_to_image").convert("L")) image.save("output_image_path")
70
0
from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax.numpy as jnp from jax import random from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .scheduling_utils_flax import FlaxSchedulerMixin @flax.struct.dataclass class __A: snake_case_ = None snake_case_ = None snake_case_ = None # sigma(t_i) @classmethod def SCREAMING_SNAKE_CASE_ ( cls ) -> List[Any]: '''simple docstring''' return cls() @dataclass class __A( a ): snake_case_ = 4_2 snake_case_ = 4_2 snake_case_ = 4_2 class __A( a , a ): @property def SCREAMING_SNAKE_CASE_ ( self ) -> int: '''simple docstring''' return True @register_to_config def __init__( self , _snake_case = 0.02 , _snake_case = 100 , _snake_case = 1.007 , _snake_case = 80 , _snake_case = 0.05 , _snake_case = 50 , ) -> Union[str, Any]: '''simple docstring''' pass def SCREAMING_SNAKE_CASE_ ( self ) -> List[str]: '''simple docstring''' return KarrasVeSchedulerState.create() def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case , _snake_case = () ) -> KarrasVeSchedulerState: '''simple docstring''' __a = jnp.arange(0 , A_ )[::-1].copy() __a = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in timesteps ] return state.replace( num_inference_steps=A_ , schedule=jnp.array(A_ , dtype=jnp.floataa ) , timesteps=A_ , ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case , _snake_case , _snake_case , ) -> Tuple[jnp.ndarray, float]: '''simple docstring''' if self.config.s_min <= sigma <= self.config.s_max: __a = min(self.config.s_churn / state.num_inference_steps , 2**0.5 - 1 ) else: __a = 0 # sample eps ~ N(0, S_noise^2 * I) __a = random.split(A_ , num=1 ) __a = self.config.s_noise * random.normal(key=A_ , shape=sample.shape ) __a = sigma + gamma * sigma __a = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case = True , ) -> Union[FlaxKarrasVeOutput, Tuple]: '''simple docstring''' __a = sample_hat + sigma_hat * model_output __a = (sample_hat - pred_original_sample) / sigma_hat __a = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative, state) return FlaxKarrasVeOutput(prev_sample=A_ , derivative=A_ , state=A_ ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case = True , ) -> Union[FlaxKarrasVeOutput, Tuple]: '''simple docstring''' __a = sample_prev + sigma_prev * model_output __a = (sample_prev - pred_original_sample) / sigma_prev __a = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative, state) return FlaxKarrasVeOutput(prev_sample=A_ , derivative=A_ , state=A_ ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case , _snake_case , _snake_case ) -> List[str]: '''simple docstring''' raise NotImplementedError()
219
# this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under git won't be considered # since the output of this script is fed into Makefile commands it doesn't print a newline after the results import re import subprocess import sys lowerCamelCase : List[Any] = subprocess.check_output("git merge-base main HEAD".split()).decode("utf-8") lowerCamelCase : Tuple = ( subprocess.check_output(F"""git diff --diff-filter=d --name-only {fork_point_sha}""".split()).decode("utf-8").split() ) lowerCamelCase : Tuple = "|".join(sys.argv[1:]) lowerCamelCase : Any = re.compile(rF"""^({joined_dirs}).*?\.py$""") lowerCamelCase : List[str] = [x for x in modified_files if regex.match(x)] print(" ".join(relevant_modified_files), end="")
70
0
'''simple docstring''' import unittest import numpy as np import torch from diffusers import ScoreSdeVePipeline, ScoreSdeVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class a ( unittest.TestCase ): """simple docstring""" @property def lowercase_ ( self ): '''simple docstring''' torch.manual_seed(0 ) __UpperCAmelCase: str = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=("""DownBlock2D""", """AttnDownBlock2D""") , up_block_types=("""AttnUpBlock2D""", """UpBlock2D""") , ) return model def lowercase_ ( self ): '''simple docstring''' __UpperCAmelCase: List[str] = self.dummy_uncond_unet __UpperCAmelCase: str = ScoreSdeVeScheduler() __UpperCAmelCase: Dict = ScoreSdeVePipeline(unet=A_ , scheduler=A_ ) sde_ve.to(A_ ) sde_ve.set_progress_bar_config(disable=A_ ) __UpperCAmelCase: Tuple = torch.manual_seed(0 ) __UpperCAmelCase: int = sde_ve(num_inference_steps=2 , output_type="""numpy""" , generator=A_ ).images __UpperCAmelCase: int = torch.manual_seed(0 ) __UpperCAmelCase: int = sde_ve(num_inference_steps=2 , output_type="""numpy""" , generator=A_ , return_dict=A_ )[ 0 ] __UpperCAmelCase: int = image[0, -3:, -3:, -1] __UpperCAmelCase: Optional[int] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __UpperCAmelCase: Any = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class a ( unittest.TestCase ): """simple docstring""" def lowercase_ ( self ): '''simple docstring''' __UpperCAmelCase: Union[str, Any] = """google/ncsnpp-church-256""" __UpperCAmelCase: List[str] = UNetaDModel.from_pretrained(A_ ) __UpperCAmelCase: Dict = ScoreSdeVeScheduler.from_pretrained(A_ ) __UpperCAmelCase: Union[str, Any] = ScoreSdeVePipeline(unet=A_ , scheduler=A_ ) sde_ve.to(A_ ) sde_ve.set_progress_bar_config(disable=A_ ) __UpperCAmelCase: str = torch.manual_seed(0 ) __UpperCAmelCase: Any = sde_ve(num_inference_steps=10 , output_type="""numpy""" , generator=A_ ).images __UpperCAmelCase: List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) __UpperCAmelCase: List[Any] = np.array([0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
523
import argparse import json import subprocess def _SCREAMING_SNAKE_CASE ( lowercase : Dict , lowercase : List[str] ): '''simple docstring''' lowerCamelCase_ = [] lowerCamelCase_ = ( f"""curl -H \"Accept: application/vnd.github+json\" -H \"Authorization: Bearer {token}\"""" ' https://api.github.com/repos/huggingface/transformers/actions/runners' ) lowerCamelCase_ = subprocess.run(lowercase , shell=lowercase , stdout=subprocess.PIPE ) lowerCamelCase_ = output.stdout.decode('utf-8' ) lowerCamelCase_ = json.loads(lowercase ) lowerCamelCase_ = status['runners'] for runner in runners: if runner["name"] in target_runners: if runner["status"] == "offline": offline_runners.append(lowercase ) # save the result so we can report them on Slack with open('offline_runners.txt' , 'w' ) as fp: fp.write(json.dumps(lowercase ) ) if len(lowercase ) > 0: lowerCamelCase_ = '\n'.join([x['name'] for x in offline_runners] ) raise ValueError(f"""The following runners are offline:\n{failed}""" ) if __name__ == "__main__": def _SCREAMING_SNAKE_CASE ( lowercase : List[str] ): '''simple docstring''' return values.split(',' ) lowerCamelCase : str = argparse.ArgumentParser() # Required parameters parser.add_argument( "--target_runners", default=None, type=list_str, required=True, help="Comma-separated list of runners to check status.", ) parser.add_argument( "--token", default=None, type=str, required=True, help="A token that has actions:read permission." ) lowerCamelCase : Optional[int] = parser.parse_args() get_runner_status(args.target_runners, args.token)
70
0
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation UpperCamelCase_ = logging.get_logger(__name__) UpperCamelCase_ = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} UpperCamelCase_ = { "tokenizer_file": { "EleutherAI/gpt-neox-20b": "https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/tokenizer.json", }, } UpperCamelCase_ = { "gpt-neox-20b": 20_48, } class _SCREAMING_SNAKE_CASE ( snake_case ): lowerCamelCase_ = VOCAB_FILES_NAMES lowerCamelCase_ = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase_ = ['input_ids', 'attention_mask'] def __init__( self : List[str] , snake_case_ : List[Any]=None , snake_case_ : Tuple=None , snake_case_ : Union[str, Any]=None , snake_case_ : Any="<|endoftext|>" , snake_case_ : str="<|endoftext|>" , snake_case_ : int="<|endoftext|>" , snake_case_ : Optional[Any]=False , **snake_case_ : List[Any] , ): """simple docstring""" super().__init__( A_ , A_ , tokenizer_file=A_ , unk_token=A_ , bos_token=A_ , eos_token=A_ , add_prefix_space=A_ , **A_ , ) A : Dict = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('''add_prefix_space''' , A_ ) != add_prefix_space: A : Dict = getattr(A_ , pre_tok_state.pop('''type''' ) ) A : List[Any] = add_prefix_space A : Dict = pre_tok_class(**A_ ) A : Tuple = add_prefix_space def _UpperCAmelCase ( self : Optional[int] , snake_case_ : str , snake_case_ : Optional[str] = None ): """simple docstring""" A : Dict = self._tokenizer.model.save(A_ , name=A_ ) return tuple(A_ ) def _UpperCAmelCase ( self : List[str] , snake_case_ : "Conversation" ): """simple docstring""" A : Dict = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(A_ , add_special_tokens=A_ ) + [self.eos_token_id] ) if len(A_ ) > self.model_max_length: A : List[str] = input_ids[-self.model_max_length :] return input_ids
256
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from timm import create_model from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import BitConfig, BitForImageClassification, BitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase : Optional[Any] = logging.get_logger(__name__) def _SCREAMING_SNAKE_CASE ( lowercase : List[Any] ): '''simple docstring''' lowerCamelCase_ = 'huggingface/label-files' lowerCamelCase_ = 'imagenet-1k-id2label.json' lowerCamelCase_ = json.load(open(hf_hub_download(lowercase , lowercase , repo_type='dataset' ) , 'r' ) ) lowerCamelCase_ = {int(lowercase ): v for k, v in idalabel.items()} lowerCamelCase_ = {v: k for k, v in idalabel.items()} lowerCamelCase_ = 'std_conv' if 'bit' in model_name else False # note that when using BiT as backbone for ViT-hybrid checkpoints, # one needs to additionally set config.layer_type = "bottleneck", config.stem_type = "same", # config.conv_layer = "std_conv_same" lowerCamelCase_ = BitConfig( conv_layer=lowercase , num_labels=10_00 , idalabel=lowercase , labelaid=lowercase , ) return config def _SCREAMING_SNAKE_CASE ( lowercase : Any ): '''simple docstring''' if "stem.conv" in name: lowerCamelCase_ = name.replace('stem.conv' , 'bit.embedder.convolution' ) if "blocks" in name: lowerCamelCase_ = name.replace('blocks' , 'layers' ) if "head.fc" in name: lowerCamelCase_ = name.replace('head.fc' , 'classifier.1' ) if name.startswith('norm' ): lowerCamelCase_ = 'bit.' + name if "bit" not in name and "classifier" not in name: lowerCamelCase_ = 'bit.encoder.' + name return name def _SCREAMING_SNAKE_CASE ( ): '''simple docstring''' lowerCamelCase_ = 'http://images.cocodataset.org/val2017/000000039769.jpg' lowerCamelCase_ = Image.open(requests.get(lowercase , stream=lowercase ).raw ) return im @torch.no_grad() def _SCREAMING_SNAKE_CASE ( lowercase : str , lowercase : int , lowercase : Union[str, Any]=False ): '''simple docstring''' lowerCamelCase_ = get_config(lowercase ) # load original model from timm lowerCamelCase_ = create_model(lowercase , pretrained=lowercase ) timm_model.eval() # load state_dict of original model lowerCamelCase_ = timm_model.state_dict() for key in state_dict.copy().keys(): lowerCamelCase_ = state_dict.pop(lowercase ) lowerCamelCase_ = val.squeeze() if 'head' in key else val # load HuggingFace model lowerCamelCase_ = BitForImageClassification(lowercase ) model.eval() model.load_state_dict(lowercase ) # create image processor lowerCamelCase_ = create_transform(**resolve_data_config({} , model=lowercase ) ) lowerCamelCase_ = transform.transforms lowerCamelCase_ = { 'bilinear': PILImageResampling.BILINEAR, 'bicubic': PILImageResampling.BICUBIC, 'nearest': PILImageResampling.NEAREST, } lowerCamelCase_ = BitImageProcessor( do_resize=lowercase , size={'shortest_edge': timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=lowercase , crop_size={'height': timm_transforms[1].size[0], 'width': timm_transforms[1].size[1]} , do_normalize=lowercase , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) lowerCamelCase_ = prepare_img() lowerCamelCase_ = transform(lowercase ).unsqueeze(0 ) lowerCamelCase_ = processor(lowercase , return_tensors='pt' ).pixel_values # verify pixel values assert torch.allclose(lowercase , lowercase ) # verify logits with torch.no_grad(): lowerCamelCase_ = model(lowercase ) lowerCamelCase_ = outputs.logits print('Logits:' , logits[0, :3] ) print('Predicted class:' , model.config.idalabel[logits.argmax(-1 ).item()] ) lowerCamelCase_ = timm_model(lowercase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(lowercase , outputs.logits , atol=1e-3 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: Path(lowercase ).mkdir(exist_ok=lowercase ) print(f"""Saving model {model_name} and processor to {pytorch_dump_folder_path}""" ) model.save_pretrained(lowercase ) processor.save_pretrained(lowercase ) if push_to_hub: print(f"""Pushing model {model_name} and processor to the hub""" ) model.push_to_hub(f"""ybelkada/{model_name}""" ) processor.push_to_hub(f"""ybelkada/{model_name}""" ) if __name__ == "__main__": lowerCamelCase : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="resnetv2_50x1_bitm", type=str, help="Name of the BiT timm model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to push the model to the hub.", ) lowerCamelCase : Optional[int] = parser.parse_args() convert_bit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
70
0
"""simple docstring""" from __future__ import annotations def lowerCAmelCase_ ( lowercase_ : int ): '''simple docstring''' __SCREAMING_SNAKE_CASE : Optional[int] = 2 __SCREAMING_SNAKE_CASE : int = [] while i * i <= n: if n % i: i += 1 else: n //= i factors.append(lowercase_ ) if n > 1: factors.append(lowercase_ ) return factors if __name__ == "__main__": import doctest doctest.testmod()
674
from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class A: '''simple docstring''' def __init__( self : List[Any] , A_ : Dict , A_ : Union[str, Any]=13 , A_ : List[Any]=30 , A_ : Optional[Any]=2 , A_ : List[str]=3 , A_ : List[str]=True , A_ : Dict=True , A_ : List[Any]=32 , A_ : Any=2 , A_ : Any=4 , A_ : Optional[int]=37 , A_ : Dict="gelu" , A_ : List[Any]=0.1 , A_ : Optional[int]=0.1 , A_ : Union[str, Any]=10 , A_ : Optional[Any]=0.02 , A_ : List[Any]=3 , A_ : str=None , ) -> str: """simple docstring""" lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = image_size lowerCamelCase_ = patch_size lowerCamelCase_ = num_channels lowerCamelCase_ = is_training lowerCamelCase_ = use_labels lowerCamelCase_ = hidden_size lowerCamelCase_ = num_hidden_layers lowerCamelCase_ = num_attention_heads lowerCamelCase_ = intermediate_size lowerCamelCase_ = hidden_act lowerCamelCase_ = hidden_dropout_prob lowerCamelCase_ = attention_probs_dropout_prob lowerCamelCase_ = type_sequence_label_size lowerCamelCase_ = initializer_range lowerCamelCase_ = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) lowerCamelCase_ = (image_size // patch_size) ** 2 lowerCamelCase_ = num_patches + 1 def a__ ( self : List[str] ) -> Dict: """simple docstring""" lowerCamelCase_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase_ = None if self.use_labels: lowerCamelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase_ = self.get_config() return config, pixel_values, labels def a__ ( self : List[Any] ) -> Any: """simple docstring""" return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=A_ , initializer_range=self.initializer_range , ) def a__ ( self : Any , A_ : int , A_ : int , A_ : int ) -> List[Any]: """simple docstring""" lowerCamelCase_ = TFViTModel(config=A_ ) lowerCamelCase_ = model(A_ , training=A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # Test with an image with different size than the one specified in config. lowerCamelCase_ = self.image_size // 2 lowerCamelCase_ = pixel_values[:, :, :image_size, :image_size] lowerCamelCase_ = model(A_ , interpolate_pos_encoding=A_ , training=A_ ) lowerCamelCase_ = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, seq_length, self.hidden_size) ) def a__ ( self : List[Any] , A_ : List[Any] , A_ : Any , A_ : Any ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.type_sequence_label_size lowerCamelCase_ = TFViTForImageClassification(A_ ) lowerCamelCase_ = model(A_ , labels=A_ , training=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # Test with an image with different size than the one specified in config. lowerCamelCase_ = self.image_size // 2 lowerCamelCase_ = pixel_values[:, :, :image_size, :image_size] lowerCamelCase_ = model(A_ , interpolate_pos_encoding=A_ , training=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowerCamelCase_ = 1 lowerCamelCase_ = TFViTForImageClassification(A_ ) lowerCamelCase_ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCamelCase_ = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def a__ ( self : List[Any] ) -> Any: """simple docstring""" lowerCamelCase_ = self.prepare_config_and_inputs() lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = config_and_inputs lowerCamelCase_ = {'pixel_values': pixel_values} return config, inputs_dict @require_tf class A( UpperCamelCase , UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () UpperCamelCase = ( {'''feature-extraction''': TFViTModel, '''image-classification''': TFViTForImageClassification} if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def a__ ( self : int ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = TFViTModelTester(self ) lowerCamelCase_ = ConfigTester(self , config_class=A_ , has_text_modality=A_ , hidden_size=37 ) def a__ ( self : int ) -> List[str]: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='ViT does not use inputs_embeds' ) def a__ ( self : List[str] ) -> Tuple: """simple docstring""" pass @unittest.skip(reason='ViT does not use inputs_embeds' ) def a__ ( self : int ) -> Optional[Any]: """simple docstring""" pass def a__ ( self : str ) -> List[str]: """simple docstring""" lowerCamelCase_ , lowerCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase_ = model_class(A_ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) lowerCamelCase_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(A_ , tf.keras.layers.Layer ) ) def a__ ( self : List[Any] ) -> Dict: """simple docstring""" lowerCamelCase_ , lowerCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase_ = model_class(A_ ) lowerCamelCase_ = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase_ = [*signature.parameters.keys()] lowerCamelCase_ = ['pixel_values'] self.assertListEqual(arg_names[:1] , A_ ) def a__ ( self : Tuple ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def a__ ( self : Tuple ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A_ ) @slow def a__ ( self : Any ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = TFViTModel.from_pretrained('google/vit-base-patch16-224' ) self.assertIsNotNone(A_ ) def _SCREAMING_SNAKE_CASE ( ): '''simple docstring''' lowerCamelCase_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class A( unittest.TestCase ): '''simple docstring''' @cached_property def a__ ( self : int ) -> Tuple: """simple docstring""" return ViTImageProcessor.from_pretrained('google/vit-base-patch16-224' ) if is_vision_available() else None @slow def a__ ( self : List[str] ) -> str: """simple docstring""" lowerCamelCase_ = TFViTForImageClassification.from_pretrained('google/vit-base-patch16-224' ) lowerCamelCase_ = self.default_image_processor lowerCamelCase_ = prepare_img() lowerCamelCase_ = image_processor(images=A_ , return_tensors='tf' ) # forward pass lowerCamelCase_ = model(**A_ ) # verify the logits lowerCamelCase_ = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , A_ ) lowerCamelCase_ = tf.constant([-0.2744, 0.8215, -0.0836] ) tf.debugging.assert_near(outputs.logits[0, :3] , A_ , atol=1E-4 )
70
0
'''simple docstring''' import csv from collections import defaultdict from dataclasses import dataclass, field from typing import List, Optional import matplotlib.pyplot as plt import numpy as np from matplotlib.ticker import ScalarFormatter from transformers import HfArgumentParser def SCREAMING_SNAKE_CASE_ ( __A : Union[str, Any]=None , __A : Dict=None ) -> int: return field(default_factory=lambda: default , metadata=__A ) @dataclass class lowercase_ : """simple docstring""" lowerCamelCase_ = field( metadata={'''help''': '''The csv file to plot.'''} , ) lowerCamelCase_ = field( default=A , metadata={'''help''': '''Whether to plot along batch size or sequence length. Defaults to sequence length.'''} , ) lowerCamelCase_ = field( default=A , metadata={'''help''': '''Whether the csv file has time results or memory results. Defaults to memory results.'''} , ) lowerCamelCase_ = field( default=A , metadata={'''help''': '''Disable logarithmic scale when plotting'''} , ) lowerCamelCase_ = field( default=A , metadata={ '''help''': '''Whether the csv file has training results or inference results. Defaults to inference results.''' } , ) lowerCamelCase_ = field( default=A , metadata={'''help''': '''Filename under which the plot will be saved. If unused no plot is saved.'''} , ) lowerCamelCase_ = list_field( default=A , metadata={'''help''': '''List of model names that are used instead of the ones in the csv file.'''} ) def SCREAMING_SNAKE_CASE_ ( __A : List[Any] ) -> Union[str, Any]: try: int(__A ) return True except ValueError: return False def SCREAMING_SNAKE_CASE_ ( __A : List[Any] ) -> int: try: float(__A ) return True except ValueError: return False class lowercase_ : """simple docstring""" def __init__( self : Optional[int] , __lowerCamelCase : List[Any] ): """simple docstring""" _SCREAMING_SNAKE_CASE = args _SCREAMING_SNAKE_CASE = defaultdict(lambda: {"bsz": [], "seq_len": [], "result": {}} ) with open(self.args.csv_file , newline="" ) as csv_file: _SCREAMING_SNAKE_CASE = csv.DictReader(A_ ) for row in reader: _SCREAMING_SNAKE_CASE = row["model"] self.result_dict[model_name]["bsz"].append(int(row["batch_size"] ) ) self.result_dict[model_name]["seq_len"].append(int(row["sequence_length"] ) ) if can_convert_to_int(row["result"] ): # value is not None _SCREAMING_SNAKE_CASE = int(row["result"] ) elif can_convert_to_float(row["result"] ): # value is not None _SCREAMING_SNAKE_CASE = float(row["result"] ) def lowerCAmelCase_ ( self : Union[str, Any] ): """simple docstring""" _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE = plt.subplots() _SCREAMING_SNAKE_CASE = "Time usage" if self.args.is_time else "Memory usage" _SCREAMING_SNAKE_CASE = title_str + " for training" if self.args.is_train else title_str + " for inference" if not self.args.no_log_scale: # set logarithm scales ax.set_xscale("log" ) ax.set_yscale("log" ) for axis in [ax.xaxis, ax.yaxis]: axis.set_major_formatter(ScalarFormatter() ) for model_name_idx, model_name in enumerate(self.result_dict.keys() ): _SCREAMING_SNAKE_CASE = sorted(set(self.result_dict[model_name]["bsz"] ) ) _SCREAMING_SNAKE_CASE = sorted(set(self.result_dict[model_name]["seq_len"] ) ) _SCREAMING_SNAKE_CASE = self.result_dict[model_name]["result"] ((_SCREAMING_SNAKE_CASE), (_SCREAMING_SNAKE_CASE)) = ( (batch_sizes, sequence_lengths) if self.args.plot_along_batch else (sequence_lengths, batch_sizes) ) _SCREAMING_SNAKE_CASE = ( model_name if self.args.short_model_names is None else self.args.short_model_names[model_name_idx] ) for inner_loop_value in inner_loop_array: if self.args.plot_along_batch: _SCREAMING_SNAKE_CASE = np.asarray( [results[(x, inner_loop_value)] for x in x_axis_array if (x, inner_loop_value) in results] , dtype=A_ , ) else: _SCREAMING_SNAKE_CASE = np.asarray( [results[(inner_loop_value, x)] for x in x_axis_array if (inner_loop_value, x) in results] , dtype=np.floataa , ) ((_SCREAMING_SNAKE_CASE), (_SCREAMING_SNAKE_CASE)) = ( ("batch_size", "len") if self.args.plot_along_batch else ("in #tokens", "bsz") ) _SCREAMING_SNAKE_CASE = np.asarray(A_ , A_ )[: len(A_ )] plt.scatter( A_ , A_ , label=F"""{label_model_name} - {inner_loop_label}: {inner_loop_value}""" ) plt.plot(A_ , A_ , "--" ) title_str += F""" {label_model_name} vs.""" _SCREAMING_SNAKE_CASE = title_str[:-4] _SCREAMING_SNAKE_CASE = "Time in s" if self.args.is_time else "Memory in MB" # plot plt.title(A_ ) plt.xlabel(A_ ) plt.ylabel(A_ ) plt.legend() if self.args.figure_png_file is not None: plt.savefig(self.args.figure_png_file ) else: plt.show() def SCREAMING_SNAKE_CASE_ ( ) -> str: _SCREAMING_SNAKE_CASE = HfArgumentParser(__A ) _SCREAMING_SNAKE_CASE = parser.parse_args_into_dataclasses()[0] _SCREAMING_SNAKE_CASE = Plot(args=__A ) plot.plot() if __name__ == "__main__": main()
418
import itertools import random import unittest import numpy as np from transformers import is_speech_available from transformers.testing_utils import require_torch, require_torchaudio from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import SpeechaTextFeatureExtractor lowerCamelCase : Any = random.Random() def _SCREAMING_SNAKE_CASE ( lowercase : Union[str, Any] , lowercase : int=1.0 , lowercase : List[str]=None , lowercase : str=None ): '''simple docstring''' if rng is None: lowerCamelCase_ = global_rng lowerCamelCase_ = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class A( unittest.TestCase ): '''simple docstring''' def __init__( self : List[Any] , A_ : Dict , A_ : int=7 , A_ : str=400 , A_ : Dict=2000 , A_ : List[Any]=24 , A_ : List[Any]=24 , A_ : int=0.0 , A_ : Dict=16000 , A_ : List[Any]=True , A_ : str=True , ) -> Dict: """simple docstring""" lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = min_seq_length lowerCamelCase_ = max_seq_length lowerCamelCase_ = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) lowerCamelCase_ = feature_size lowerCamelCase_ = num_mel_bins lowerCamelCase_ = padding_value lowerCamelCase_ = sampling_rate lowerCamelCase_ = return_attention_mask lowerCamelCase_ = do_normalize def a__ ( self : List[str] ) -> Optional[int]: """simple docstring""" return { "feature_size": self.feature_size, "num_mel_bins": self.num_mel_bins, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def a__ ( self : List[Any] , A_ : str=False , A_ : Union[str, Any]=False ) -> str: """simple docstring""" def _flatten(A_ : List[Any] ): return list(itertools.chain(*A_ ) ) if equal_length: lowerCamelCase_ = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size lowerCamelCase_ = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: lowerCamelCase_ = [np.asarray(A_ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class A( UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = SpeechaTextFeatureExtractor if is_speech_available() else None def a__ ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = SpeechaTextFeatureExtractionTester(self ) def a__ ( self : str , A_ : Dict ) -> Dict: """simple docstring""" self.assertTrue(np.all(np.mean(A_ , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(A_ , axis=0 ) - 1 ) < 1E-3 ) ) def a__ ( self : int ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = [np.asarray(A_ ) for speech_input in speech_inputs] # Test feature size lowerCamelCase_ = feature_extractor(A_ , padding=A_ , return_tensors='np' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.feature_size ) # Test not batched input lowerCamelCase_ = feature_extractor(speech_inputs[0] , return_tensors='np' ).input_features lowerCamelCase_ = feature_extractor(np_speech_inputs[0] , return_tensors='np' ).input_features self.assertTrue(np.allclose(A_ , A_ , atol=1E-3 ) ) # Test batched lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. lowerCamelCase_ = [floats_list((1, x) )[0] for x in (800, 800, 800)] lowerCamelCase_ = np.asarray(A_ ) lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1E-3 ) ) def a__ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = ['longest', 'max_length', 'do_not_pad'] lowerCamelCase_ = [None, 16, None] for max_length, padding in zip(A_ , A_ ): lowerCamelCase_ = feature_extractor( A_ , padding=A_ , max_length=A_ , return_attention_mask=A_ ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = [np.sum(A_ ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def a__ ( self : List[Any] ) -> Any: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = ['longest', 'max_length', 'do_not_pad'] lowerCamelCase_ = [None, 16, None] for max_length, padding in zip(A_ , A_ ): lowerCamelCase_ = feature_extractor( A_ , max_length=A_ , padding=A_ , return_tensors='np' , return_attention_mask=A_ ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = [np.sum(A_ ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self.assertTrue(input_features[0][fbank_feat_lengths[0] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self.assertTrue(input_features[0][fbank_feat_lengths[1] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def a__ ( self : List[str] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = feature_extractor( A_ , padding='max_length' , max_length=4 , truncation=A_ , return_tensors='np' , return_attention_mask=A_ , ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1] ) self._check_zero_mean_unit_variance(input_features[2] ) def a__ ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = feature_extractor( A_ , padding='longest' , max_length=4 , truncation=A_ , return_tensors='np' , return_attention_mask=A_ , ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 4, 24) ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = feature_extractor( A_ , padding='longest' , max_length=16 , truncation=A_ , return_tensors='np' , return_attention_mask=A_ , ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 6, 24) ) def a__ ( self : List[Any] ) -> List[str]: """simple docstring""" import torch lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = np.random.rand(100 , 32 ).astype(np.floataa ) lowerCamelCase_ = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: lowerCamelCase_ = feature_extractor.pad([{'input_features': inputs}] , return_tensors='np' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) lowerCamelCase_ = feature_extractor.pad([{'input_features': inputs}] , return_tensors='pt' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def a__ ( self : List[str] , A_ : Union[str, Any] ) -> List[Any]: """simple docstring""" from datasets import load_dataset lowerCamelCase_ = load_dataset('hf-internal-testing/librispeech_asr_dummy' , 'clean' , split='validation' ) # automatic decoding with librispeech lowerCamelCase_ = ds.sort('id' ).select(range(A_ ) )[:num_samples]['audio'] return [x["array"] for x in speech_samples] def a__ ( self : str ) -> Tuple: """simple docstring""" lowerCamelCase_ = np.array([ -1.5745, -1.7713, -1.7020, -1.6069, -1.2250, -1.1105, -0.9072, -0.8241, -1.2310, -0.8098, -0.3320, -0.4101, -0.7985, -0.4996, -0.8213, -0.9128, -1.0420, -1.1286, -1.0440, -0.7999, -0.8405, -1.2275, -1.5443, -1.4625, ] ) # fmt: on lowerCamelCase_ = self._load_datasamples(1 ) lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = feature_extractor(A_ , return_tensors='pt' ).input_features self.assertEquals(input_features.shape , (1, 584, 24) ) self.assertTrue(np.allclose(input_features[0, 0, :30] , A_ , atol=1E-4 ) )
70
0
'''simple docstring''' import cmath import math def _A ( UpperCAmelCase ,UpperCAmelCase ,UpperCAmelCase ,UpperCAmelCase ): '''simple docstring''' A__ = math.radians(UpperCAmelCase ) A__ = math.radians(UpperCAmelCase ) # Convert voltage and current to rectangular form A__ = cmath.rect(UpperCAmelCase ,UpperCAmelCase ) A__ = cmath.rect(UpperCAmelCase ,UpperCAmelCase ) # Calculate apparent power return voltage_rect * current_rect if __name__ == "__main__": import doctest doctest.testmod()
531
import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class A( UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = TransfoXLTokenizer UpperCamelCase = False UpperCamelCase = False def a__ ( self : Optional[Any] ) -> int: """simple docstring""" super().setUp() lowerCamelCase_ = [ '<unk>', '[CLS]', '[SEP]', 'want', 'unwanted', 'wa', 'un', 'running', ',', 'low', 'l', ] lowerCamelCase_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def a__ ( self : Optional[Any] , **A_ : Tuple ) -> Any: """simple docstring""" lowerCamelCase_ = True return TransfoXLTokenizer.from_pretrained(self.tmpdirname , **A_ ) def a__ ( self : List[str] , A_ : Dict ) -> Any: """simple docstring""" lowerCamelCase_ = '<unk> UNwanted , running' lowerCamelCase_ = '<unk> unwanted, running' return input_text, output_text def a__ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = TransfoXLTokenizer(vocab_file=self.vocab_file , lower_case=A_ ) lowerCamelCase_ = tokenizer.tokenize('<unk> UNwanted , running' ) self.assertListEqual(A_ , ['<unk>', 'unwanted', ',', 'running'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A_ ) , [0, 4, 8, 7] ) def a__ ( self : Any ) -> str: """simple docstring""" lowerCamelCase_ = TransfoXLTokenizer(lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['hello', '!', 'how', 'are', 'you', '?'] ) def a__ ( self : int ) -> Dict: """simple docstring""" lowerCamelCase_ = TransfoXLTokenizer(lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?'] ) def a__ ( self : List[Any] ) -> Tuple: """simple docstring""" lowerCamelCase_ = TransfoXLTokenizer(lower_case=A_ ) lowerCamelCase_ = 'Hello (bracket) and side-scrolled [and] Henry\'s $5,000 with 3.34 m. What\'s up!?' lowerCamelCase_ = [ 'Hello', '(', 'bracket', ')', 'and', 'side', '@-@', 'scrolled', '[', 'and', ']', 'Henry', '\'s', '$', '5', '@,@', '000', 'with', '3', '@.@', '34', 'm', '.', 'What', '\'s', 'up', '!', '?', ] self.assertListEqual(tokenizer.tokenize(A_ ) , A_ ) self.assertEqual(tokenizer.convert_tokens_to_string(A_ ) , A_ ) def a__ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = len(A_ ) tokenizer.add_tokens(['new1', 'new2'] ) tokenizer.move_added_token('new1' , 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(A_ ) , original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode('new1' ) , [1] ) self.assertEqual(tokenizer.decode([1] ) , 'new1' )
70
0
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TextClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. lowerCamelCase : Dict = {"LayoutLMv2Config", "LayoutLMv3Config"} @is_pipeline_test class A__ ( unittest.TestCase ): A__ = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING A__ = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: A__ = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: A__ = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } @require_torch def A ( self : List[str] ) -> Union[str, Any]: '''simple docstring''' _SCREAMING_SNAKE_CASE =pipeline( task='text-classification' , model='hf-internal-testing/tiny-random-distilbert' , framework='pt' ) _SCREAMING_SNAKE_CASE =text_classifier('This is great !' ) self.assertEqual(nested_simplify(A_ ) , [{'label': 'LABEL_0', 'score': 0.5_04}] ) _SCREAMING_SNAKE_CASE =text_classifier('This is great !' , top_k=2 ) self.assertEqual( nested_simplify(A_ ) , [{'label': 'LABEL_0', 'score': 0.5_04}, {'label': 'LABEL_1', 'score': 0.4_96}] ) _SCREAMING_SNAKE_CASE =text_classifier(['This is great !', 'This is bad'] , top_k=2 ) self.assertEqual( nested_simplify(A_ ) , [ [{'label': 'LABEL_0', 'score': 0.5_04}, {'label': 'LABEL_1', 'score': 0.4_96}], [{'label': 'LABEL_0', 'score': 0.5_04}, {'label': 'LABEL_1', 'score': 0.4_96}], ] , ) _SCREAMING_SNAKE_CASE =text_classifier('This is great !' , top_k=1 ) self.assertEqual(nested_simplify(A_ ) , [{'label': 'LABEL_0', 'score': 0.5_04}] ) # Legacy behavior _SCREAMING_SNAKE_CASE =text_classifier('This is great !' , return_all_scores=A_ ) self.assertEqual(nested_simplify(A_ ) , [{'label': 'LABEL_0', 'score': 0.5_04}] ) _SCREAMING_SNAKE_CASE =text_classifier('This is great !' , return_all_scores=A_ ) self.assertEqual( nested_simplify(A_ ) , [[{'label': 'LABEL_0', 'score': 0.5_04}, {'label': 'LABEL_1', 'score': 0.4_96}]] ) _SCREAMING_SNAKE_CASE =text_classifier(['This is great !', 'Something else'] , return_all_scores=A_ ) self.assertEqual( nested_simplify(A_ ) , [ [{'label': 'LABEL_0', 'score': 0.5_04}, {'label': 'LABEL_1', 'score': 0.4_96}], [{'label': 'LABEL_0', 'score': 0.5_04}, {'label': 'LABEL_1', 'score': 0.4_96}], ] , ) _SCREAMING_SNAKE_CASE =text_classifier(['This is great !', 'Something else'] , return_all_scores=A_ ) self.assertEqual( nested_simplify(A_ ) , [ {'label': 'LABEL_0', 'score': 0.5_04}, {'label': 'LABEL_0', 'score': 0.5_04}, ] , ) @require_torch def A ( self : Tuple ) -> str: '''simple docstring''' import torch _SCREAMING_SNAKE_CASE =pipeline( task='text-classification' , model='hf-internal-testing/tiny-random-distilbert' , framework='pt' , device=torch.device('cpu' ) , ) _SCREAMING_SNAKE_CASE =text_classifier('This is great !' ) self.assertEqual(nested_simplify(A_ ) , [{'label': 'LABEL_0', 'score': 0.5_04}] ) @require_tf def A ( self : Optional[int] ) -> Optional[int]: '''simple docstring''' _SCREAMING_SNAKE_CASE =pipeline( task='text-classification' , model='hf-internal-testing/tiny-random-distilbert' , framework='tf' ) _SCREAMING_SNAKE_CASE =text_classifier('This is great !' ) self.assertEqual(nested_simplify(A_ ) , [{'label': 'LABEL_0', 'score': 0.5_04}] ) @slow @require_torch def A ( self : Any ) -> Dict: '''simple docstring''' _SCREAMING_SNAKE_CASE =pipeline('text-classification' ) _SCREAMING_SNAKE_CASE =text_classifier('This is great !' ) self.assertEqual(nested_simplify(A_ ) , [{'label': 'POSITIVE', 'score': 1.0}] ) _SCREAMING_SNAKE_CASE =text_classifier('This is bad !' ) self.assertEqual(nested_simplify(A_ ) , [{'label': 'NEGATIVE', 'score': 1.0}] ) _SCREAMING_SNAKE_CASE =text_classifier('Birds are a type of animal' ) self.assertEqual(nested_simplify(A_ ) , [{'label': 'POSITIVE', 'score': 0.9_88}] ) @slow @require_tf def A ( self : Optional[Any] ) -> str: '''simple docstring''' _SCREAMING_SNAKE_CASE =pipeline('text-classification' , framework='tf' ) _SCREAMING_SNAKE_CASE =text_classifier('This is great !' ) self.assertEqual(nested_simplify(A_ ) , [{'label': 'POSITIVE', 'score': 1.0}] ) _SCREAMING_SNAKE_CASE =text_classifier('This is bad !' ) self.assertEqual(nested_simplify(A_ ) , [{'label': 'NEGATIVE', 'score': 1.0}] ) _SCREAMING_SNAKE_CASE =text_classifier('Birds are a type of animal' ) self.assertEqual(nested_simplify(A_ ) , [{'label': 'POSITIVE', 'score': 0.9_88}] ) def A ( self : str , _a : List[str] , _a : Any , _a : Optional[Any] ) -> Optional[Any]: '''simple docstring''' _SCREAMING_SNAKE_CASE =TextClassificationPipeline(model=A_ , tokenizer=A_ ) return text_classifier, ["HuggingFace is in", "This is another test"] def A ( self : int , _a : Union[str, Any] , _a : str ) -> Dict: '''simple docstring''' _SCREAMING_SNAKE_CASE =text_classifier.model # Small inputs because BartTokenizer tiny has maximum position embeddings = 22 _SCREAMING_SNAKE_CASE ='HuggingFace is in' _SCREAMING_SNAKE_CASE =text_classifier(A_ ) self.assertEqual(nested_simplify(A_ ) , [{'label': ANY(A_ ), 'score': ANY(A_ )}] ) self.assertTrue(outputs[0]['label'] in model.config.idalabel.values() ) _SCREAMING_SNAKE_CASE =['HuggingFace is in ', 'Paris is in France'] _SCREAMING_SNAKE_CASE =text_classifier(A_ ) self.assertEqual( nested_simplify(A_ ) , [{'label': ANY(A_ ), 'score': ANY(A_ )}, {'label': ANY(A_ ), 'score': ANY(A_ )}] , ) self.assertTrue(outputs[0]['label'] in model.config.idalabel.values() ) self.assertTrue(outputs[1]['label'] in model.config.idalabel.values() ) # Forcing to get all results with `top_k=None` # This is NOT the legacy format _SCREAMING_SNAKE_CASE =text_classifier(A_ , top_k=A_ ) _SCREAMING_SNAKE_CASE =len(model.config.idalabel.values() ) self.assertEqual( nested_simplify(A_ ) , [[{'label': ANY(A_ ), 'score': ANY(A_ )}] * N, [{'label': ANY(A_ ), 'score': ANY(A_ )}] * N] , ) _SCREAMING_SNAKE_CASE ={'text': 'HuggingFace is in ', 'text_pair': 'Paris is in France'} _SCREAMING_SNAKE_CASE =text_classifier(A_ ) self.assertEqual( nested_simplify(A_ ) , {'label': ANY(A_ ), 'score': ANY(A_ )} , ) self.assertTrue(outputs['label'] in model.config.idalabel.values() ) # This might be used a text pair, but tokenizer + pipe interaction # makes it hard to understand that it's not using the pair properly # https://github.com/huggingface/transformers/issues/17305 # We disabled this usage instead as it was outputting wrong outputs. _SCREAMING_SNAKE_CASE =[['HuggingFace is in ', 'Paris is in France']] with self.assertRaises(A_ ): text_classifier(A_ ) # This used to be valid for doing text pairs # We're keeping it working because of backward compatibility _SCREAMING_SNAKE_CASE =text_classifier([[['HuggingFace is in ', 'Paris is in France']]] ) self.assertEqual( nested_simplify(A_ ) , [{'label': ANY(A_ ), 'score': ANY(A_ )}] , ) self.assertTrue(outputs[0]['label'] in model.config.idalabel.values() )
405
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetrImageProcessor class A( unittest.TestCase ): '''simple docstring''' def __init__( self : Optional[Any] , A_ : Dict , A_ : int=7 , A_ : Any=3 , A_ : List[str]=30 , A_ : Union[str, Any]=400 , A_ : List[str]=True , A_ : int=None , A_ : Any=True , A_ : str=1 / 255 , A_ : int=True , A_ : List[Any]=[0.5, 0.5, 0.5] , A_ : Union[str, Any]=[0.5, 0.5, 0.5] , A_ : Union[str, Any]=True , ) -> List[str]: """simple docstring""" lowerCamelCase_ = size if size is not None else {'shortest_edge': 18, 'longest_edge': 1333} lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = num_channels lowerCamelCase_ = min_resolution lowerCamelCase_ = max_resolution lowerCamelCase_ = do_resize lowerCamelCase_ = size lowerCamelCase_ = do_rescale lowerCamelCase_ = rescale_factor lowerCamelCase_ = do_normalize lowerCamelCase_ = image_mean lowerCamelCase_ = image_std lowerCamelCase_ = do_pad def a__ ( self : Tuple ) -> Dict: """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_pad": self.do_pad, } def a__ ( self : Union[str, Any] , A_ : Dict , A_ : Any=False ) -> Union[str, Any]: """simple docstring""" if not batched: lowerCamelCase_ = image_inputs[0] if isinstance(A_ , Image.Image ): lowerCamelCase_ , lowerCamelCase_ = image.size else: lowerCamelCase_ , lowerCamelCase_ = image.shape[1], image.shape[2] if w < h: lowerCamelCase_ = int(self.size['shortest_edge'] * h / w ) lowerCamelCase_ = self.size['shortest_edge'] elif w > h: lowerCamelCase_ = self.size['shortest_edge'] lowerCamelCase_ = int(self.size['shortest_edge'] * w / h ) else: lowerCamelCase_ = self.size['shortest_edge'] lowerCamelCase_ = self.size['shortest_edge'] else: lowerCamelCase_ = [] for image in image_inputs: lowerCamelCase_ , lowerCamelCase_ = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) lowerCamelCase_ = max(A_ , key=lambda A_ : item[0] )[0] lowerCamelCase_ = max(A_ , key=lambda A_ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class A( UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = DetrImageProcessor if is_vision_available() else None def a__ ( self : List[Any] ) -> str: """simple docstring""" lowerCamelCase_ = DetrImageProcessingTester(self ) @property def a__ ( self : List[str] ) -> List[Any]: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def a__ ( self : Dict ) -> int: """simple docstring""" lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , 'image_mean' ) ) self.assertTrue(hasattr(A_ , 'image_std' ) ) self.assertTrue(hasattr(A_ , 'do_normalize' ) ) self.assertTrue(hasattr(A_ , 'do_rescale' ) ) self.assertTrue(hasattr(A_ , 'rescale_factor' ) ) self.assertTrue(hasattr(A_ , 'do_resize' ) ) self.assertTrue(hasattr(A_ , 'size' ) ) self.assertTrue(hasattr(A_ , 'do_pad' ) ) def a__ ( self : List[str] ) -> str: """simple docstring""" lowerCamelCase_ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 18, 'longest_edge': 1333} ) self.assertEqual(image_processor.do_pad , A_ ) lowerCamelCase_ = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A_ ) self.assertEqual(image_processor.size , {'shortest_edge': 42, 'longest_edge': 84} ) self.assertEqual(image_processor.do_pad , A_ ) def a__ ( self : Dict ) -> Any: """simple docstring""" pass def a__ ( self : Tuple ) -> Tuple: """simple docstring""" lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCamelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ ) for image in image_inputs: self.assertIsInstance(A_ , Image.Image ) # Test not batched input lowerCamelCase_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) lowerCamelCase_ = image_processing(A_ , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def a__ ( self : List[str] ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCamelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , np.ndarray ) # Test not batched input lowerCamelCase_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase_ = image_processing(A_ , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def a__ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCamelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test not batched input lowerCamelCase_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase_ = image_processing(A_ , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def a__ ( self : Tuple ) -> List[Any]: """simple docstring""" lowerCamelCase_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_annotations.txt' , 'r' ) as f: lowerCamelCase_ = json.loads(f.read() ) lowerCamelCase_ = {'image_id': 39769, 'annotations': target} # encode them lowerCamelCase_ = DetrImageProcessor.from_pretrained('facebook/detr-resnet-50' ) lowerCamelCase_ = image_processing(images=A_ , annotations=A_ , return_tensors='pt' ) # verify pixel values lowerCamelCase_ = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , A_ ) lowerCamelCase_ = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area lowerCamelCase_ = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , A_ ) ) # verify boxes lowerCamelCase_ = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , A_ ) lowerCamelCase_ = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , A_ , atol=1E-3 ) ) # verify image_id lowerCamelCase_ = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , A_ ) ) # verify is_crowd lowerCamelCase_ = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , A_ ) ) # verify class_labels lowerCamelCase_ = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , A_ ) ) # verify orig_size lowerCamelCase_ = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , A_ ) ) # verify size lowerCamelCase_ = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , A_ ) ) @slow def a__ ( self : str ) -> Any: """simple docstring""" lowerCamelCase_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt' , 'r' ) as f: lowerCamelCase_ = json.loads(f.read() ) lowerCamelCase_ = {'file_name': '000000039769.png', 'image_id': 39769, 'segments_info': target} lowerCamelCase_ = pathlib.Path('./tests/fixtures/tests_samples/COCO/coco_panoptic' ) # encode them lowerCamelCase_ = DetrImageProcessor.from_pretrained('facebook/detr-resnet-50-panoptic' ) lowerCamelCase_ = image_processing(images=A_ , annotations=A_ , masks_path=A_ , return_tensors='pt' ) # verify pixel values lowerCamelCase_ = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , A_ ) lowerCamelCase_ = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area lowerCamelCase_ = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , A_ ) ) # verify boxes lowerCamelCase_ = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , A_ ) lowerCamelCase_ = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , A_ , atol=1E-3 ) ) # verify image_id lowerCamelCase_ = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , A_ ) ) # verify is_crowd lowerCamelCase_ = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , A_ ) ) # verify class_labels lowerCamelCase_ = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , A_ ) ) # verify masks lowerCamelCase_ = 822873 self.assertEqual(encoding['labels'][0]['masks'].sum().item() , A_ ) # verify orig_size lowerCamelCase_ = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , A_ ) ) # verify size lowerCamelCase_ = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , A_ ) )
70
0
'''simple docstring''' import random import unittest import torch from diffusers import IFInpaintingPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class _snake_case ( a_ , a_ , unittest.TestCase ): SCREAMING_SNAKE_CASE : Tuple = IFInpaintingPipeline SCREAMING_SNAKE_CASE : Optional[int] = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'''width''', '''height'''} SCREAMING_SNAKE_CASE : Union[str, Any] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS SCREAMING_SNAKE_CASE : Dict = PipelineTesterMixin.required_optional_params - {'''latents'''} def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' return self._get_dummy_components() def _SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=0 ): '''simple docstring''' if str(A_ ).startswith('mps' ): lowerCAmelCase = torch.manual_seed(A_ ) else: lowerCAmelCase = torch.Generator(device=A_ ).manual_seed(A_ ) lowerCAmelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(A_ ) ).to(A_ ) lowerCAmelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(A_ ) ).to(A_ ) lowerCAmelCase = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'mask_image': mask_image, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' self._test_save_load_optional_components() @unittest.skipIf(torch_device != 'cuda' , reason='float16 requires CUDA' ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' super().test_save_load_floataa(expected_max_diff=1e-1 ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' self._test_attention_slicing_forward_pass(expected_max_diff=1e-2 ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' self._test_save_load_local() def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' self._test_inference_batch_single_identical( expected_max_diff=1e-2 , )
284
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase : Dict = logging.get_logger(__name__) lowerCamelCase : int = { "microsoft/swinv2-tiny-patch4-window8-256": ( "https://huggingface.co/microsoft/swinv2-tiny-patch4-window8-256/resolve/main/config.json" ), } class A( UpperCamelCase ): '''simple docstring''' UpperCamelCase = '''swinv2''' UpperCamelCase = { '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers''', } def __init__( self : Dict , A_ : List[Any]=224 , A_ : Optional[Any]=4 , A_ : int=3 , A_ : Dict=96 , A_ : Any=[2, 2, 6, 2] , A_ : Optional[Any]=[3, 6, 12, 24] , A_ : Tuple=7 , A_ : Tuple=4.0 , A_ : str=True , A_ : str=0.0 , A_ : Union[str, Any]=0.0 , A_ : Optional[Any]=0.1 , A_ : str="gelu" , A_ : int=False , A_ : str=0.02 , A_ : List[Any]=1E-5 , A_ : Any=32 , **A_ : Tuple , ) -> Any: """simple docstring""" super().__init__(**A_ ) lowerCamelCase_ = image_size lowerCamelCase_ = patch_size lowerCamelCase_ = num_channels lowerCamelCase_ = embed_dim lowerCamelCase_ = depths lowerCamelCase_ = len(A_ ) lowerCamelCase_ = num_heads lowerCamelCase_ = window_size lowerCamelCase_ = mlp_ratio lowerCamelCase_ = qkv_bias lowerCamelCase_ = hidden_dropout_prob lowerCamelCase_ = attention_probs_dropout_prob lowerCamelCase_ = drop_path_rate lowerCamelCase_ = hidden_act lowerCamelCase_ = use_absolute_embeddings lowerCamelCase_ = layer_norm_eps lowerCamelCase_ = initializer_range lowerCamelCase_ = encoder_stride # we set the hidden_size attribute in order to make Swinv2 work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowerCamelCase_ = int(embed_dim * 2 ** (len(A_ ) - 1) ) lowerCamelCase_ = (0, 0, 0, 0)
70
0
'''simple docstring''' import unittest from transformers.testing_utils import CaptureStdout from transformers.tools.python_interpreter import evaluate def __UpperCamelCase ( lowercase_ : Union[str, Any] ): """simple docstring""" return x + 2 class __SCREAMING_SNAKE_CASE (unittest.TestCase ): """simple docstring""" def _a ( self ): """simple docstring""" a_ = 'x = 3' a_ = {} a_ = evaluate(A_ , {} , state=A_ ) assert result == 3 self.assertDictEqual(A_ , {'x': 3} ) a_ = 'x = y' a_ = {'y': 5} a_ = evaluate(A_ , {} , state=A_ ) # evaluate returns the value of the last assignment. assert result == 5 self.assertDictEqual(A_ , {'x': 5, 'y': 5} ) def _a ( self ): """simple docstring""" a_ = 'y = add_two(x)' a_ = {'x': 3} a_ = evaluate(A_ , {'add_two': add_two} , state=A_ ) assert result == 5 self.assertDictEqual(A_ , {'x': 3, 'y': 5} ) # Won't work without the tool with CaptureStdout() as out: a_ = evaluate(A_ , {} , state=A_ ) assert result is None assert "tried to execute add_two" in out.out def _a ( self ): """simple docstring""" a_ = 'x = 3' a_ = {} a_ = evaluate(A_ , {} , state=A_ ) assert result == 3 self.assertDictEqual(A_ , {'x': 3} ) def _a ( self ): """simple docstring""" a_ = 'test_dict = {\'x\': x, \'y\': add_two(x)}' a_ = {'x': 3} a_ = evaluate(A_ , {'add_two': add_two} , state=A_ ) self.assertDictEqual(A_ , {'x': 3, 'y': 5} ) self.assertDictEqual(A_ , {'x': 3, 'test_dict': {'x': 3, 'y': 5}} ) def _a ( self ): """simple docstring""" a_ = 'x = 3\ny = 5' a_ = {} a_ = evaluate(A_ , {} , state=A_ ) # evaluate returns the value of the last assignment. assert result == 5 self.assertDictEqual(A_ , {'x': 3, 'y': 5} ) def _a ( self ): """simple docstring""" a_ = 'text = f\'This is x: {x}.\'' a_ = {'x': 3} a_ = evaluate(A_ , {} , state=A_ ) # evaluate returns the value of the last assignment. assert result == "This is x: 3." self.assertDictEqual(A_ , {'x': 3, 'text': 'This is x: 3.'} ) def _a ( self ): """simple docstring""" a_ = 'if x <= 3:\n y = 2\nelse:\n y = 5' a_ = {'x': 3} a_ = evaluate(A_ , {} , state=A_ ) # evaluate returns the value of the last assignment. assert result == 2 self.assertDictEqual(A_ , {'x': 3, 'y': 2} ) a_ = {'x': 8} a_ = evaluate(A_ , {} , state=A_ ) # evaluate returns the value of the last assignment. assert result == 5 self.assertDictEqual(A_ , {'x': 8, 'y': 5} ) def _a ( self ): """simple docstring""" a_ = 'test_list = [x, add_two(x)]' a_ = {'x': 3} a_ = evaluate(A_ , {'add_two': add_two} , state=A_ ) self.assertListEqual(A_ , [3, 5] ) self.assertDictEqual(A_ , {'x': 3, 'test_list': [3, 5]} ) def _a ( self ): """simple docstring""" a_ = 'y = x' a_ = {'x': 3} a_ = evaluate(A_ , {} , state=A_ ) assert result == 3 self.assertDictEqual(A_ , {'x': 3, 'y': 3} ) def _a ( self ): """simple docstring""" a_ = 'test_list = [x, add_two(x)]\ntest_list[1]' a_ = {'x': 3} a_ = evaluate(A_ , {'add_two': add_two} , state=A_ ) assert result == 5 self.assertDictEqual(A_ , {'x': 3, 'test_list': [3, 5]} ) a_ = 'test_dict = {\'x\': x, \'y\': add_two(x)}\ntest_dict[\'y\']' a_ = {'x': 3} a_ = evaluate(A_ , {'add_two': add_two} , state=A_ ) assert result == 5 self.assertDictEqual(A_ , {'x': 3, 'test_dict': {'x': 3, 'y': 5}} ) def _a ( self ): """simple docstring""" a_ = 'x = 0\nfor i in range(3):\n x = i' a_ = {} a_ = evaluate(A_ , {'range': range} , state=A_ ) assert result == 2 self.assertDictEqual(A_ , {'x': 2, 'i': 2} )
536
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import ( AutoProcessor, BertTokenizerFast, BlipImageProcessor, GPTaTokenizer, InstructBlipProcessor, PreTrainedTokenizerFast, ) @require_vision class A( unittest.TestCase ): '''simple docstring''' def a__ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = tempfile.mkdtemp() lowerCamelCase_ = BlipImageProcessor() lowerCamelCase_ = GPTaTokenizer.from_pretrained('hf-internal-testing/tiny-random-GPT2Model' ) lowerCamelCase_ = BertTokenizerFast.from_pretrained('hf-internal-testing/tiny-random-bert' ) lowerCamelCase_ = InstructBlipProcessor(A_ , A_ , A_ ) processor.save_pretrained(self.tmpdirname ) def a__ ( self : Optional[int] , **A_ : Optional[int] ) -> Dict: """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **A_ ).tokenizer def a__ ( self : List[str] , **A_ : str ) -> Optional[Any]: """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **A_ ).image_processor def a__ ( self : Tuple , **A_ : Any ) -> Optional[int]: """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **A_ ).qformer_tokenizer def a__ ( self : str ) -> str: """simple docstring""" shutil.rmtree(self.tmpdirname ) def a__ ( self : Dict ) -> Dict: """simple docstring""" lowerCamelCase_ = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] lowerCamelCase_ = [Image.fromarray(np.moveaxis(A_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def a__ ( self : Tuple ) -> Tuple: """simple docstring""" lowerCamelCase_ = InstructBlipProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() , qformer_tokenizer=self.get_qformer_tokenizer() , ) processor.save_pretrained(self.tmpdirname ) lowerCamelCase_ = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) lowerCamelCase_ = self.get_image_processor(do_normalize=A_ , padding_value=1.0 ) lowerCamelCase_ = InstructBlipProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=A_ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , A_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , A_ ) self.assertIsInstance(processor.qformer_tokenizer , A_ ) def a__ ( self : List[str] ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = self.prepare_image_inputs() lowerCamelCase_ = image_processor(A_ , return_tensors='np' ) lowerCamelCase_ = processor(images=A_ , return_tensors='np' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def a__ ( self : Union[str, Any] ) -> Tuple: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = 'lower newer' lowerCamelCase_ = processor(text=A_ ) lowerCamelCase_ = tokenizer(A_ , return_token_type_ids=A_ ) lowerCamelCase_ = qformer_tokenizer(A_ , return_token_type_ids=A_ ) for key in encoded_tokens.keys(): self.assertListEqual(encoded_tokens[key] , encoded_processor[key] ) for key in encoded_tokens_qformer.keys(): self.assertListEqual(encoded_tokens_qformer[key] , encoded_processor['qformer_' + key] ) def a__ ( self : Tuple ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = 'lower newer' lowerCamelCase_ = self.prepare_image_inputs() lowerCamelCase_ = processor(text=A_ , images=A_ ) self.assertListEqual( list(inputs.keys() ) , ['input_ids', 'attention_mask', 'qformer_input_ids', 'qformer_attention_mask', 'pixel_values'] , ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def a__ ( self : Optional[int] ) -> int: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowerCamelCase_ = processor.batch_decode(A_ ) lowerCamelCase_ = tokenizer.batch_decode(A_ ) self.assertListEqual(A_ , A_ ) def a__ ( self : Tuple ) -> List[str]: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = 'lower newer' lowerCamelCase_ = self.prepare_image_inputs() lowerCamelCase_ = processor(text=A_ , images=A_ ) self.assertListEqual( list(inputs.keys() ) , ['input_ids', 'attention_mask', 'qformer_input_ids', 'qformer_attention_mask', 'pixel_values'] , )
70
0
import math def A ( __UpperCamelCase , __UpperCamelCase ) -> List[Any]: A__ = len(__UpperCamelCase ) A__ = int(math.floor(math.sqrt(__UpperCamelCase ) ) ) A__ = 0 while arr[min(__UpperCamelCase , __UpperCamelCase ) - 1] < x: A__ = step step += int(math.floor(math.sqrt(__UpperCamelCase ) ) ) if prev >= n: return -1 while arr[prev] < x: A__ = prev + 1 if prev == min(__UpperCamelCase , __UpperCamelCase ): return -1 if arr[prev] == x: return prev return -1 if __name__ == "__main__": SCREAMING_SNAKE_CASE__ = input('''Enter numbers separated by a comma:\n''').strip() SCREAMING_SNAKE_CASE__ = [int(item) for item in user_input.split(''',''')] SCREAMING_SNAKE_CASE__ = int(input('''Enter the number to be searched:\n''')) SCREAMING_SNAKE_CASE__ = jump_search(arr, x) if res == -1: print('''Number not found!''') else: print(f'Number {x} is at index {res}')
9
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_OBJECT_DETECTION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING lowerCamelCase : Tuple = logging.get_logger(__name__) lowerCamelCase : List[Any] = Dict[str, Any] lowerCamelCase : Dict = List[Prediction] @add_end_docstrings(UpperCamelCase ) class A( UpperCamelCase ): '''simple docstring''' def __init__( self : Tuple , *A_ : int , **A_ : int ) -> Optional[int]: """simple docstring""" super().__init__(*A_ , **A_ ) if self.framework == "tf": raise ValueError(f"""The {self.__class__} is only available in PyTorch.""" ) requires_backends(self , 'vision' ) self.check_model_type( dict(MODEL_FOR_OBJECT_DETECTION_MAPPING.items() + MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING.items() ) ) def a__ ( self : Union[str, Any] , **A_ : Union[str, Any] ) -> List[str]: """simple docstring""" lowerCamelCase_ = {} if "threshold" in kwargs: lowerCamelCase_ = kwargs['threshold'] return {}, {}, postprocess_kwargs def __call__( self : str , *A_ : Optional[int] , **A_ : Tuple ) -> Union[Predictions, List[Prediction]]: """simple docstring""" return super().__call__(*A_ , **A_ ) def a__ ( self : Union[str, Any] , A_ : Tuple ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = load_image(A_ ) lowerCamelCase_ = torch.IntTensor([[image.height, image.width]] ) lowerCamelCase_ = self.image_processor(images=[image] , return_tensors='pt' ) if self.tokenizer is not None: lowerCamelCase_ = self.tokenizer(text=inputs['words'] , boxes=inputs['boxes'] , return_tensors='pt' ) lowerCamelCase_ = target_size return inputs def a__ ( self : Union[str, Any] , A_ : List[str] ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = model_inputs.pop('target_size' ) lowerCamelCase_ = self.model(**A_ ) lowerCamelCase_ = outputs.__class__({'target_size': target_size, **outputs} ) if self.tokenizer is not None: lowerCamelCase_ = model_inputs['bbox'] return model_outputs def a__ ( self : str , A_ : Any , A_ : Tuple=0.9 ) -> str: """simple docstring""" lowerCamelCase_ = model_outputs['target_size'] if self.tokenizer is not None: # This is a LayoutLMForTokenClassification variant. # The OCR got the boxes and the model classified the words. lowerCamelCase_ , lowerCamelCase_ = target_size[0].tolist() def unnormalize(A_ : Dict ): return self._get_bounding_box( torch.Tensor( [ (width * bbox[0] / 1000), (height * bbox[1] / 1000), (width * bbox[2] / 1000), (height * bbox[3] / 1000), ] ) ) lowerCamelCase_ , lowerCamelCase_ = model_outputs['logits'].squeeze(0 ).softmax(dim=-1 ).max(dim=-1 ) lowerCamelCase_ = [self.model.config.idalabel[prediction] for prediction in classes.tolist()] lowerCamelCase_ = [unnormalize(A_ ) for bbox in model_outputs['bbox'].squeeze(0 )] lowerCamelCase_ = ['score', 'label', 'box'] lowerCamelCase_ = [dict(zip(A_ , A_ ) ) for vals in zip(scores.tolist() , A_ , A_ ) if vals[0] > threshold] else: # This is a regular ForObjectDetectionModel lowerCamelCase_ = self.image_processor.post_process_object_detection(A_ , A_ , A_ ) lowerCamelCase_ = raw_annotations[0] lowerCamelCase_ = raw_annotation['scores'] lowerCamelCase_ = raw_annotation['labels'] lowerCamelCase_ = raw_annotation['boxes'] lowerCamelCase_ = scores.tolist() lowerCamelCase_ = [self.model.config.idalabel[label.item()] for label in labels] lowerCamelCase_ = [self._get_bounding_box(A_ ) for box in boxes] # {"scores": [...], ...} --> [{"score":x, ...}, ...] lowerCamelCase_ = ['score', 'label', 'box'] lowerCamelCase_ = [ dict(zip(A_ , A_ ) ) for vals in zip(raw_annotation['scores'] , raw_annotation['labels'] , raw_annotation['boxes'] ) ] return annotation def a__ ( self : Union[str, Any] , A_ : "torch.Tensor" ) -> Dict[str, int]: """simple docstring""" if self.framework != "pt": raise ValueError('The ObjectDetectionPipeline is only available in PyTorch.' ) lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = box.int().tolist() lowerCamelCase_ = { 'xmin': xmin, 'ymin': ymin, 'xmax': xmax, 'ymax': ymax, } return bbox
70
0
'''simple docstring''' # this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under git won't be considered # since the output of this script is fed into Makefile commands it doesn't print a newline after the results import re import subprocess import sys __lowerCAmelCase =subprocess.check_output("git merge-base main HEAD".split()).decode("utf-8") __lowerCAmelCase =( subprocess.check_output(f'''git diff --diff-filter=d --name-only {fork_point_sha}'''.split()).decode("utf-8").split() ) __lowerCAmelCase ="|".join(sys.argv[1:]) __lowerCAmelCase =re.compile(rf'''^({joined_dirs}).*?\.py$''') __lowerCAmelCase =[x for x in modified_files if regex.match(x)] print(" ".join(relevant_modified_files), end="")
697
from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
70
0
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging A : str = logging.get_logger(__name__) A : List[str] = { "vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_config_file": "tokenizer_config.json", } A : Tuple = { "vocab_file": { "facebook/blenderbot_small-90M": "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json" }, "merges_file": { "facebook/blenderbot_small-90M": "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt" }, "tokenizer_config_file": { "facebook/blenderbot_small-90M": ( "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json" ) }, } A : str = {"facebook/blenderbot_small-90M": 5_1_2} def __lowerCAmelCase ( a__ ) -> int: __a = set() __a = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __a = char __a = set(a__ ) return pairs class __A( a ): snake_case_ = VOCAB_FILES_NAMES snake_case_ = PRETRAINED_VOCAB_FILES_MAP snake_case_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case_ = ['''input_ids''', '''attention_mask'''] def __init__( self , _snake_case , _snake_case , _snake_case="__start__" , _snake_case="__end__" , _snake_case="__unk__" , _snake_case="__null__" , **_snake_case , ) -> Optional[Any]: '''simple docstring''' super().__init__(unk_token=A_ , bos_token=A_ , eos_token=A_ , pad_token=A_ , **A_ ) with open(A_ , encoding='''utf-8''' ) as vocab_handle: __a = json.load(A_ ) __a = {v: k for k, v in self.encoder.items()} with open(A_ , encoding='''utf-8''' ) as merges_handle: __a = merges_handle.read().split('''\n''' )[1:-1] __a = [tuple(merge.split() ) for merge in merges] __a = dict(zip(A_ , range(len(A_ ) ) ) ) __a = {} @property def SCREAMING_SNAKE_CASE_ ( self ) -> int: '''simple docstring''' return len(self.encoder ) def SCREAMING_SNAKE_CASE_ ( self ) -> Dict: '''simple docstring''' return dict(self.encoder , **self.added_tokens_encoder ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> str: '''simple docstring''' if token in self.cache: return self.cache[token] __a = re.sub('''([.,!?()])''' , r''' \1''' , A_ ) __a = re.sub('''(\')''' , r''' \1 ''' , A_ ) __a = re.sub(r'''\s{2,}''' , ''' ''' , A_ ) if "\n" in token: __a = token.replace('''\n''' , ''' __newln__''' ) __a = token.split(''' ''' ) __a = [] for token in tokens: if not len(A_ ): continue __a = token.lower() __a = tuple(A_ ) __a = tuple(list(word[:-1] ) + [word[-1] + '''</w>'''] ) __a = get_pairs(A_ ) if not pairs: words.append(A_ ) continue while True: __a = min(A_ , key=lambda _snake_case : self.bpe_ranks.get(A_ , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break __a , __a = bigram __a = [] __a = 0 while i < len(A_ ): try: __a = word.index(A_ , A_ ) new_word.extend(word[i:j] ) __a = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(A_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __a = tuple(A_ ) __a = new_word if len(A_ ) == 1: break else: __a = get_pairs(A_ ) __a = '''@@ '''.join(A_ ) __a = word[:-4] __a = word words.append(A_ ) return " ".join(A_ ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> List[str]: '''simple docstring''' __a = [] __a = re.findall(r'''\S+\n?''' , A_ ) for token in words: split_tokens.extend(list(self.bpe(A_ ).split(''' ''' ) ) ) return split_tokens def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> int: '''simple docstring''' __a = token.lower() return self.encoder.get(A_ , self.encoder.get(self.unk_token ) ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> str: '''simple docstring''' return self.decoder.get(A_ , self.unk_token ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> str: '''simple docstring''' __a = ''' '''.join(A_ ).replace('''@@ ''' , '''''' ).strip() return out_string def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case = None ) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(A_ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return __a = os.path.join( A_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) __a = os.path.join( A_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) with open(A_ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=A_ , ensure_ascii=A_ ) + '''\n''' ) __a = 0 with open(A_ , '''w''' , encoding='''utf-8''' ) as writer: writer.write('''#version: 0.2\n''' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _snake_case : kv[1] ): if index != token_index: logger.warning( F"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ''' Please check that the tokenizer is not corrupted!''' ) __a = token_index writer.write(''' '''.join(A_ ) + '''\n''' ) index += 1 return vocab_file, merge_file
219
from collections import Counter from timeit import timeit def _SCREAMING_SNAKE_CASE ( lowercase : str = "" , ): '''simple docstring''' return sum(c % 2 for c in Counter(input_str.replace(' ' , '' ).lower() ).values() ) < 2 def _SCREAMING_SNAKE_CASE ( lowercase : str = "" ): '''simple docstring''' if len(lowercase ) == 0: return True lowerCamelCase_ = input_str.replace(' ' , '' ).lower() # character_freq_dict: Stores the frequency of every character in the input string lowerCamelCase_ = {} for character in lower_case_input_str: lowerCamelCase_ = character_freq_dict.get(lowercase , 0 ) + 1 lowerCamelCase_ = 0 for character_count in character_freq_dict.values(): if character_count % 2: odd_char += 1 if odd_char > 1: return False return True def _SCREAMING_SNAKE_CASE ( lowercase : str = "" ): '''simple docstring''' print('\nFor string = ' , lowercase , ':' ) print( '> can_string_be_rearranged_as_palindrome_counter()' , '\tans =' , can_string_be_rearranged_as_palindrome_counter(lowercase ) , '\ttime =' , timeit( 'z.can_string_be_rearranged_as_palindrome_counter(z.check_str)' , setup='import __main__ as z' , ) , 'seconds' , ) print( '> can_string_be_rearranged_as_palindrome()' , '\tans =' , can_string_be_rearranged_as_palindrome(lowercase ) , '\ttime =' , timeit( 'z.can_string_be_rearranged_as_palindrome(z.check_str)' , setup='import __main__ as z' , ) , 'seconds' , ) if __name__ == "__main__": lowerCamelCase : Optional[Any] = input( "Enter string to determine if it can be rearranged as a palindrome or not: " ).strip() benchmark(check_str) lowerCamelCase : int = can_string_be_rearranged_as_palindrome_counter(check_str) print(F"""{check_str} can {'' if status else 'not '}be rearranged as a palindrome""")
70
0
'''simple docstring''' from __future__ import annotations from decimal import Decimal from math import * # noqa: F403 from sympy import diff def UpperCamelCase__ ( _lowercase : str , _lowercase : float | Decimal , _lowercase : float = 1_0**-1_0 ) -> Optional[Any]: __UpperCAmelCase: Optional[int] = a while True: __UpperCAmelCase: Union[str, Any] = Decimal(_lowercase ) - ( Decimal(eval(_lowercase ) ) / Decimal(eval(str(diff(_lowercase ) ) ) ) # noqa: S307 ) # This number dictates the accuracy of the answer if abs(eval(_lowercase ) ) < precision: # noqa: S307 return float(_lowercase ) # Let's Execute if __name__ == "__main__": # Find root of trigonometric function # Find value of pi print(F"""The root of sin(x) = 0 is {newton_raphson("sin(x)", 2)}""") # Find root of polynomial print(F"""The root of x**2 - 5*x + 2 = 0 is {newton_raphson("x**2 - 5*x + 2", 0.4)}""") # Find Square Root of 5 print(F"""The root of log(x) - 1 = 0 is {newton_raphson("log(x) - 1", 2)}""") # Exponential Roots print(F"""The root of exp(x) - 1 = 0 is {newton_raphson("exp(x) - 1", 0)}""")
523
from __future__ import annotations from decimal import Decimal from math import * # noqa: F403 from sympy import diff def _SCREAMING_SNAKE_CASE ( lowercase : str , lowercase : float | Decimal , lowercase : float = 10**-10 ): '''simple docstring''' lowerCamelCase_ = a while True: lowerCamelCase_ = Decimal(lowercase ) - ( Decimal(eval(lowercase ) ) / Decimal(eval(str(diff(lowercase ) ) ) ) # noqa: S307 ) # This number dictates the accuracy of the answer if abs(eval(lowercase ) ) < precision: # noqa: S307 return float(lowercase ) # Let's Execute if __name__ == "__main__": # Find root of trigonometric function # Find value of pi print(F"""The root of sin(x) = 0 is {newton_raphson('sin(x)', 2)}""") # Find root of polynomial print(F"""The root of x**2 - 5*x + 2 = 0 is {newton_raphson('x**2 - 5*x + 2', 0.4)}""") # Find Square Root of 5 print(F"""The root of log(x) - 1 = 0 is {newton_raphson('log(x) - 1', 2)}""") # Exponential Roots print(F"""The root of exp(x) - 1 = 0 is {newton_raphson('exp(x) - 1', 0)}""")
70
0
from typing import Optional, Tuple, Union import tensorflow as tf from ...activations_tf import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_tf_outputs import ( TFBaseModelOutputWithNoAttention, TFBaseModelOutputWithPoolingAndNoAttention, TFSequenceClassifierOutput, ) from ...modeling_tf_utils import TFPreTrainedModel, TFSequenceClassificationLoss, keras_serializable, unpack_inputs from ...tf_utils import shape_list from ...utils import logging from .configuration_regnet import RegNetConfig UpperCamelCase_ = logging.get_logger(__name__) # General docstring UpperCamelCase_ = "RegNetConfig" # Base docstring UpperCamelCase_ = "facebook/regnet-y-040" UpperCamelCase_ = [1, 10_88, 7, 7] # Image classification docstring UpperCamelCase_ = "facebook/regnet-y-040" UpperCamelCase_ = "tabby, tabby cat" UpperCamelCase_ = [ "facebook/regnet-y-040", # See all regnet models at https://huggingface.co/models?filter=regnet ] class _SCREAMING_SNAKE_CASE ( tf.keras.layers.Layer ): def __init__( self : List[str] , snake_case_ : int , snake_case_ : int = 3 , snake_case_ : int = 1 , snake_case_ : int = 1 , snake_case_ : Optional[str] = "relu" , **snake_case_ : int , ): """simple docstring""" super().__init__(**A_ ) # The padding and conv has been verified in # https://colab.research.google.com/gist/sayakpaul/854bc10eeaf21c9ee2119e0b9f3841a7/scratchpad.ipynb A : Dict = tf.keras.layers.ZeroPaddingaD(padding=kernel_size // 2 ) A : Dict = tf.keras.layers.ConvaD( filters=A_ , kernel_size=A_ , strides=A_ , padding='''VALID''' , groups=A_ , use_bias=A_ , name='''convolution''' , ) A : List[str] = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name='''normalization''' ) A : Optional[int] = ACTaFN[activation] if activation is not None else tf.identity def _UpperCAmelCase ( self : Dict , snake_case_ : Any ): """simple docstring""" A : Tuple = self.convolution(self.padding(A_ ) ) A : int = self.normalization(A_ ) A : Union[str, Any] = self.activation(A_ ) return hidden_state class _SCREAMING_SNAKE_CASE ( tf.keras.layers.Layer ): def __init__( self : Optional[int] , snake_case_ : RegNetConfig , **snake_case_ : List[str] ): """simple docstring""" super().__init__(**A_ ) A : int = config.num_channels A : Tuple = TFRegNetConvLayer( out_channels=config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act , name='''embedder''' , ) def _UpperCAmelCase ( self : Any , snake_case_ : str ): """simple docstring""" A : Tuple = shape_list(A_ )[1] if tf.executing_eagerly() and num_channels != self.num_channels: raise ValueError( '''Make sure that the channel dimension of the pixel values match with the one set in the configuration.''' ) # When running on CPU, `tf.keras.layers.Conv2D` doesn't support `NCHW` format. # So change the input format from `NCHW` to `NHWC`. # shape = (batch_size, in_height, in_width, in_channels=num_channels) A : Optional[int] = tf.transpose(A_ , perm=(0, 2, 3, 1) ) A : List[Any] = self.embedder(A_ ) return hidden_state class _SCREAMING_SNAKE_CASE ( tf.keras.layers.Layer ): def __init__( self : Tuple , snake_case_ : int , snake_case_ : int = 2 , **snake_case_ : str ): """simple docstring""" super().__init__(**A_ ) A : Dict = tf.keras.layers.ConvaD( filters=A_ , kernel_size=1 , strides=A_ , use_bias=A_ , name='''convolution''' ) A : int = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name='''normalization''' ) def _UpperCAmelCase ( self : Any , snake_case_ : tf.Tensor , snake_case_ : bool = False ): """simple docstring""" return self.normalization(self.convolution(A_ ) , training=A_ ) class _SCREAMING_SNAKE_CASE ( tf.keras.layers.Layer ): def __init__( self : Any , snake_case_ : int , snake_case_ : int , **snake_case_ : Dict ): """simple docstring""" super().__init__(**A_ ) A : int = tf.keras.layers.GlobalAveragePoolingaD(keepdims=A_ , name='''pooler''' ) A : Optional[Any] = [ tf.keras.layers.ConvaD(filters=A_ , kernel_size=1 , activation='''relu''' , name='''attention.0''' ), tf.keras.layers.ConvaD(filters=A_ , kernel_size=1 , activation='''sigmoid''' , name='''attention.2''' ), ] def _UpperCAmelCase ( self : List[Any] , snake_case_ : Any ): """simple docstring""" A : str = self.pooler(A_ ) for layer_module in self.attention: A : Union[str, Any] = layer_module(A_ ) A : List[str] = hidden_state * pooled return hidden_state class _SCREAMING_SNAKE_CASE ( tf.keras.layers.Layer ): def __init__( self : List[Any] , snake_case_ : RegNetConfig , snake_case_ : int , snake_case_ : int , snake_case_ : int = 1 , **snake_case_ : Optional[Any] ): """simple docstring""" super().__init__(**A_ ) A : List[str] = in_channels != out_channels or stride != 1 A : int = max(1 , out_channels // config.groups_width ) A : Optional[Any] = ( TFRegNetShortCut(A_ , stride=A_ , name='''shortcut''' ) if should_apply_shortcut else tf.keras.layers.Activation('''linear''' , name='''shortcut''' ) ) # `self.layers` instead of `self.layer` because that is a reserved argument. A : Union[str, Any] = [ TFRegNetConvLayer(A_ , kernel_size=1 , activation=config.hidden_act , name='''layer.0''' ), TFRegNetConvLayer( A_ , stride=A_ , groups=A_ , activation=config.hidden_act , name='''layer.1''' ), TFRegNetConvLayer(A_ , kernel_size=1 , activation=A_ , name='''layer.2''' ), ] A : Any = ACTaFN[config.hidden_act] def _UpperCAmelCase ( self : Any , snake_case_ : List[str] ): """simple docstring""" A : Optional[Any] = hidden_state for layer_module in self.layers: A : Tuple = layer_module(A_ ) A : int = self.shortcut(A_ ) hidden_state += residual A : str = self.activation(A_ ) return hidden_state class _SCREAMING_SNAKE_CASE ( tf.keras.layers.Layer ): def __init__( self : Dict , snake_case_ : RegNetConfig , snake_case_ : int , snake_case_ : int , snake_case_ : int = 1 , **snake_case_ : Optional[Any] ): """simple docstring""" super().__init__(**A_ ) A : List[Any] = in_channels != out_channels or stride != 1 A : Optional[Any] = max(1 , out_channels // config.groups_width ) A : Any = ( TFRegNetShortCut(A_ , stride=A_ , name='''shortcut''' ) if should_apply_shortcut else tf.keras.layers.Activation('''linear''' , name='''shortcut''' ) ) A : Union[str, Any] = [ TFRegNetConvLayer(A_ , kernel_size=1 , activation=config.hidden_act , name='''layer.0''' ), TFRegNetConvLayer( A_ , stride=A_ , groups=A_ , activation=config.hidden_act , name='''layer.1''' ), TFRegNetSELayer(A_ , reduced_channels=int(round(in_channels / 4 ) ) , name='''layer.2''' ), TFRegNetConvLayer(A_ , kernel_size=1 , activation=A_ , name='''layer.3''' ), ] A : Dict = ACTaFN[config.hidden_act] def _UpperCAmelCase ( self : Any , snake_case_ : Union[str, Any] ): """simple docstring""" A : List[str] = hidden_state for layer_module in self.layers: A : int = layer_module(A_ ) A : str = self.shortcut(A_ ) hidden_state += residual A : int = self.activation(A_ ) return hidden_state class _SCREAMING_SNAKE_CASE ( tf.keras.layers.Layer ): def __init__( self : Any , snake_case_ : RegNetConfig , snake_case_ : int , snake_case_ : int , snake_case_ : int = 2 , snake_case_ : int = 2 , **snake_case_ : Tuple ): """simple docstring""" super().__init__(**A_ ) A : List[str] = TFRegNetXLayer if config.layer_type == '''x''' else TFRegNetYLayer A : List[str] = [ # downsampling is done in the first layer with stride of 2 layer(A_ , A_ , A_ , stride=A_ , name='''layers.0''' ), *[layer(A_ , A_ , A_ , name=f"""layers.{i+1}""" ) for i in range(depth - 1 )], ] def _UpperCAmelCase ( self : str , snake_case_ : List[Any] ): """simple docstring""" for layer_module in self.layers: A : str = layer_module(A_ ) return hidden_state class _SCREAMING_SNAKE_CASE ( tf.keras.layers.Layer ): def __init__( self : Tuple , snake_case_ : RegNetConfig , **snake_case_ : Any ): """simple docstring""" super().__init__(**A_ ) A : Optional[int] = [] # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( TFRegNetStage( A_ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , name='''stages.0''' , ) ) A : List[Any] = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for i, ((in_channels, out_channels), depth) in enumerate(zip(A_ , config.depths[1:] ) ): self.stages.append(TFRegNetStage(A_ , A_ , A_ , depth=A_ , name=f"""stages.{i+1}""" ) ) def _UpperCAmelCase ( self : Dict , snake_case_ : tf.Tensor , snake_case_ : bool = False , snake_case_ : bool = True ): """simple docstring""" A : Optional[Any] = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: A : int = hidden_states + (hidden_state,) A : List[Any] = stage_module(A_ ) if output_hidden_states: A : List[str] = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return TFBaseModelOutputWithNoAttention(last_hidden_state=A_ , hidden_states=A_ ) @keras_serializable class _SCREAMING_SNAKE_CASE ( tf.keras.layers.Layer ): lowerCamelCase_ = RegNetConfig def __init__( self : List[str] , snake_case_ : Optional[Any] , **snake_case_ : int ): """simple docstring""" super().__init__(**A_ ) A : Tuple = config A : Optional[Any] = TFRegNetEmbeddings(A_ , name='''embedder''' ) A : List[str] = TFRegNetEncoder(A_ , name='''encoder''' ) A : Union[str, Any] = tf.keras.layers.GlobalAveragePoolingaD(keepdims=A_ , name='''pooler''' ) @unpack_inputs def _UpperCAmelCase ( self : int , snake_case_ : tf.Tensor , snake_case_ : Optional[bool] = None , snake_case_ : Optional[bool] = None , snake_case_ : bool = False , ): """simple docstring""" A : Tuple = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) A : List[str] = return_dict if return_dict is not None else self.config.use_return_dict A : Dict = self.embedder(A_ , training=A_ ) A : List[str] = self.encoder( A_ , output_hidden_states=A_ , return_dict=A_ , training=A_ ) A : Any = encoder_outputs[0] A : Any = self.pooler(A_ ) # Change to NCHW output format have uniformity in the modules A : Any = tf.transpose(A_ , perm=(0, 3, 1, 2) ) A : Any = tf.transpose(A_ , perm=(0, 3, 1, 2) ) # Change the other hidden state outputs to NCHW as well if output_hidden_states: A : List[Any] = tuple([tf.transpose(A_ , perm=(0, 3, 1, 2) ) for h in encoder_outputs[1]] ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=A_ , pooler_output=A_ , hidden_states=hidden_states if output_hidden_states else encoder_outputs.hidden_states , ) class _SCREAMING_SNAKE_CASE ( snake_case ): lowerCamelCase_ = RegNetConfig lowerCamelCase_ = 'regnet' lowerCamelCase_ = 'pixel_values' @property def _UpperCAmelCase ( self : int ): """simple docstring""" return {"pixel_values": tf.TensorSpec(shape=(None, self.config.num_channels, 224, 224) , dtype=tf.floataa )} UpperCamelCase_ = r"\n Parameters:\n This model is a Tensorflow\n [tf.keras.layers.Layer](https://www.tensorflow.org/api_docs/python/tf/keras/layers/Layer) sub-class. Use it as a\n regular Tensorflow Module and refer to the Tensorflow documentation for all matter related to general usage and\n behavior.\n config ([`RegNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~TFPreTrainedModel.from_pretrained`] method to load the model weights.\n" UpperCamelCase_ = r"\n Args:\n pixel_values (`tf.Tensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`ConveNextImageProcessor.__call__`] for details.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for\n more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n" @add_start_docstrings( 'The bare RegNet model outputting raw features without any specific head on top.', snake_case, ) class _SCREAMING_SNAKE_CASE ( snake_case ): def __init__( self : Tuple , snake_case_ : RegNetConfig , *snake_case_ : Tuple , **snake_case_ : Any ): """simple docstring""" super().__init__(A_ , *A_ , **A_ ) A : Tuple = TFRegNetMainLayer(A_ , name='''regnet''' ) @unpack_inputs @add_start_docstrings_to_model_forward(A_ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=A_ , config_class=_CONFIG_FOR_DOC , modality='''vision''' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def _UpperCAmelCase ( self : List[str] , snake_case_ : tf.Tensor , snake_case_ : Optional[bool] = None , snake_case_ : Optional[bool] = None , snake_case_ : int=False , ): """simple docstring""" A : List[str] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) A : List[str] = return_dict if return_dict is not None else self.config.use_return_dict A : List[str] = self.regnet( pixel_values=A_ , output_hidden_states=A_ , return_dict=A_ , training=A_ , ) if not return_dict: return (outputs[0],) + outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=outputs.last_hidden_state , pooler_output=outputs.pooler_output , hidden_states=outputs.hidden_states , ) @add_start_docstrings( '\n RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n ', snake_case, ) class _SCREAMING_SNAKE_CASE ( snake_case, snake_case ): def __init__( self : List[str] , snake_case_ : RegNetConfig , *snake_case_ : Optional[int] , **snake_case_ : int ): """simple docstring""" super().__init__(A_ , *A_ , **A_ ) A : Any = config.num_labels A : Any = TFRegNetMainLayer(A_ , name='''regnet''' ) # classification head A : Union[str, Any] = [ tf.keras.layers.Flatten(), tf.keras.layers.Dense(config.num_labels , name='''classifier.1''' ) if config.num_labels > 0 else tf.identity, ] @unpack_inputs @add_start_docstrings_to_model_forward(A_ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=A_ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def _UpperCAmelCase ( self : Optional[int] , snake_case_ : tf.Tensor = None , snake_case_ : tf.Tensor = None , snake_case_ : bool = None , snake_case_ : bool = None , snake_case_ : List[Any]=False , ): """simple docstring""" A : Optional[int] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) A : Tuple = return_dict if return_dict is not None else self.config.use_return_dict A : int = self.regnet( A_ , output_hidden_states=A_ , return_dict=A_ , training=A_ ) A : Optional[int] = outputs.pooler_output if return_dict else outputs[1] A : Tuple = self.classifier[0](A_ ) A : int = self.classifier[1](A_ ) A : List[Any] = None if labels is None else self.hf_compute_loss(labels=A_ , logits=A_ ) if not return_dict: A : Optional[Any] = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return TFSequenceClassifierOutput(loss=A_ , logits=A_ , hidden_states=outputs.hidden_states )
256
from __future__ import annotations from typing import Any class A( UpperCamelCase ): '''simple docstring''' pass class A: '''simple docstring''' def __init__( self : List[str] , A_ : Any ) -> None: """simple docstring""" lowerCamelCase_ = data lowerCamelCase_ = None def __iter__( self : int ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self lowerCamelCase_ = [] while node: if node in visited: raise ContainsLoopError visited.append(A_ ) yield node.data lowerCamelCase_ = node.next_node @property def a__ ( self : List[str] ) -> bool: """simple docstring""" try: list(self ) return False except ContainsLoopError: return True if __name__ == "__main__": lowerCamelCase : int = Node(1) lowerCamelCase : Optional[int] = Node(2) lowerCamelCase : Union[str, Any] = Node(3) lowerCamelCase : List[Any] = Node(4) print(root_node.has_loop) # False lowerCamelCase : int = root_node.next_node print(root_node.has_loop) # True lowerCamelCase : Dict = Node(5) lowerCamelCase : Optional[int] = Node(6) lowerCamelCase : str = Node(5) lowerCamelCase : Union[str, Any] = Node(6) print(root_node.has_loop) # False lowerCamelCase : List[str] = Node(1) print(root_node.has_loop) # False
70
0
"""simple docstring""" from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class snake_case ( __UpperCAmelCase ): lowerCamelCase__ = CustomTokenizer pass
674
import unittest import torch from diffusers import DDIMScheduler, DDPMScheduler, UNetaDModel from diffusers.training_utils import set_seed from diffusers.utils.testing_utils import slow lowerCamelCase : int = False class A( unittest.TestCase ): '''simple docstring''' def a__ ( self : int , A_ : Dict=32 ) -> Any: """simple docstring""" set_seed(0 ) lowerCamelCase_ = UNetaDModel(sample_size=A_ , in_channels=3 , out_channels=3 ) lowerCamelCase_ = torch.optim.SGD(model.parameters() , lr=0.0001 ) return model, optimizer @slow def a__ ( self : int ) -> str: """simple docstring""" lowerCamelCase_ = 'cpu' # ensure full determinism without setting the CUBLAS_WORKSPACE_CONFIG env variable lowerCamelCase_ = DDPMScheduler( num_train_timesteps=1000 , beta_start=0.0001 , beta_end=0.02 , beta_schedule='linear' , clip_sample=A_ , ) lowerCamelCase_ = DDIMScheduler( num_train_timesteps=1000 , beta_start=0.0001 , beta_end=0.02 , beta_schedule='linear' , clip_sample=A_ , ) assert ddpm_scheduler.config.num_train_timesteps == ddim_scheduler.config.num_train_timesteps # shared batches for DDPM and DDIM set_seed(0 ) lowerCamelCase_ = [torch.randn((4, 3, 32, 32) ).clip(-1 , 1 ).to(A_ ) for _ in range(4 )] lowerCamelCase_ = [torch.randn((4, 3, 32, 32) ).to(A_ ) for _ in range(4 )] lowerCamelCase_ = [torch.randint(0 , 1000 , (4,) ).long().to(A_ ) for _ in range(4 )] # train with a DDPM scheduler lowerCamelCase_ , lowerCamelCase_ = self.get_model_optimizer(resolution=32 ) model.train().to(A_ ) for i in range(4 ): optimizer.zero_grad() lowerCamelCase_ = ddpm_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) lowerCamelCase_ = model(A_ , timesteps[i] ).sample lowerCamelCase_ = torch.nn.functional.mse_loss(A_ , noise[i] ) loss.backward() optimizer.step() del model, optimizer # recreate the model and optimizer, and retry with DDIM lowerCamelCase_ , lowerCamelCase_ = self.get_model_optimizer(resolution=32 ) model.train().to(A_ ) for i in range(4 ): optimizer.zero_grad() lowerCamelCase_ = ddim_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) lowerCamelCase_ = model(A_ , timesteps[i] ).sample lowerCamelCase_ = torch.nn.functional.mse_loss(A_ , noise[i] ) loss.backward() optimizer.step() del model, optimizer self.assertTrue(torch.allclose(A_ , A_ , atol=1E-5 ) ) self.assertTrue(torch.allclose(A_ , A_ , atol=1E-5 ) )
70
0
'''simple docstring''' from __future__ import annotations import numpy as np from numpy import floataa from numpy.typing import NDArray def SCREAMING_SNAKE_CASE_ ( __A : NDArray[floataa] , __A : NDArray[floataa] , __A : list[int] , __A : int , ) -> int: _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE = coefficient_matrix.shape _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE = constant_matrix.shape if rowsa != colsa: _SCREAMING_SNAKE_CASE = f"""Coefficient matrix dimensions must be nxn but received {rowsa}x{colsa}""" raise ValueError(__A ) if colsa != 1: _SCREAMING_SNAKE_CASE = f"""Constant matrix must be nx1 but received {rowsa}x{colsa}""" raise ValueError(__A ) if rowsa != rowsa: _SCREAMING_SNAKE_CASE = ( "Coefficient and constant matrices dimensions must be nxn and nx1 but " f"""received {rowsa}x{colsa} and {rowsa}x{colsa}""" ) raise ValueError(__A ) if len(__A ) != rowsa: _SCREAMING_SNAKE_CASE = ( "Number of initial values must be equal to number of rows in coefficient " f"""matrix but received {len(__A )} and {rowsa}""" ) raise ValueError(__A ) if iterations <= 0: raise ValueError("Iterations must be at least 1" ) _SCREAMING_SNAKE_CASE = np.concatenate( (coefficient_matrix, constant_matrix) , axis=1 ) _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE = table.shape strictly_diagonally_dominant(__A ) # Iterates the whole matrix for given number of times for _ in range(__A ): _SCREAMING_SNAKE_CASE = [] for row in range(__A ): _SCREAMING_SNAKE_CASE = 0 for col in range(__A ): if col == row: _SCREAMING_SNAKE_CASE = table[row][col] elif col == cols - 1: _SCREAMING_SNAKE_CASE = table[row][col] else: temp += (-1) * table[row][col] * init_val[col] _SCREAMING_SNAKE_CASE = (temp + val) / denom new_val.append(__A ) _SCREAMING_SNAKE_CASE = new_val return [float(__A ) for i in new_val] def SCREAMING_SNAKE_CASE_ ( __A : NDArray[floataa] ) -> Tuple: _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE = table.shape _SCREAMING_SNAKE_CASE = True for i in range(0 , __A ): _SCREAMING_SNAKE_CASE = 0 for j in range(0 , cols - 1 ): if i == j: continue else: total += table[i][j] if table[i][i] <= total: raise ValueError("Coefficient matrix is not strictly diagonally dominant" ) return is_diagonally_dominant # Test Cases if __name__ == "__main__": import doctest doctest.testmod()
418
def _SCREAMING_SNAKE_CASE ( lowercase : str , lowercase : str ): '''simple docstring''' if len(lowercase ) != len(lowercase ): raise ValueError('String lengths must match!' ) lowerCamelCase_ = 0 for chara, chara in zip(lowercase , lowercase ): if chara != chara: count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
70
0
'''simple docstring''' from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase_ = { "configuration_trajectory_transformer": [ "TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "TrajectoryTransformerConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ "TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "TrajectoryTransformerModel", "TrajectoryTransformerPreTrainedModel", "load_tf_weights_in_trajectory_transformer", ] if TYPE_CHECKING: from .configuration_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TrajectoryTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TrajectoryTransformerModel, TrajectoryTransformerPreTrainedModel, load_tf_weights_in_trajectory_transformer, ) else: import sys lowerCAmelCase_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
531
def _SCREAMING_SNAKE_CASE ( lowercase : int = 10 ): '''simple docstring''' if not isinstance(lowercase , lowercase ) or n < 0: raise ValueError('Invalid input' ) lowerCamelCase_ = 10**n lowerCamelCase_ = 2_84_33 * (pow(2 , 7_83_04_57 , lowercase )) + 1 return str(number % modulus ) if __name__ == "__main__": from doctest import testmod testmod() print(F"""{solution(10) = }""")
70
0
'''simple docstring''' import unittest from accelerate import debug_launcher from accelerate.test_utils import require_cpu, test_ops, test_script @require_cpu class A__ ( unittest.TestCase ): def A ( self : Union[str, Any] ) -> int: '''simple docstring''' debug_launcher(test_script.main ) def A ( self : Tuple ) -> str: '''simple docstring''' debug_launcher(test_ops.main )
405
from maths.prime_check import is_prime def _SCREAMING_SNAKE_CASE ( lowercase : int ): '''simple docstring''' if not isinstance(lowercase , lowercase ): lowerCamelCase_ = f"""Input value of [number={number}] must be an integer""" raise TypeError(lowercase ) if is_prime(lowercase ) and is_prime(number + 2 ): return number + 2 else: return -1 if __name__ == "__main__": import doctest doctest.testmod()
70
0
'''simple docstring''' from __future__ import annotations from typing import Any class _snake_case ( a_ ): pass class _snake_case : def __init__( self , _SCREAMING_SNAKE_CASE ): '''simple docstring''' lowerCAmelCase = data lowerCAmelCase = None def __iter__( self ): '''simple docstring''' lowerCAmelCase = self lowerCAmelCase = [] while node: if node in visited: raise ContainsLoopError visited.append(A_ ) yield node.data lowerCAmelCase = node.next_node @property def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' try: list(self ) return False except ContainsLoopError: return True if __name__ == "__main__": _UpperCamelCase : int = Node(1) _UpperCamelCase : Optional[int] = Node(2) _UpperCamelCase : Union[str, Any] = Node(3) _UpperCamelCase : List[Any] = Node(4) print(root_node.has_loop) # False _UpperCamelCase : int = root_node.next_node print(root_node.has_loop) # True _UpperCamelCase : Dict = Node(5) _UpperCamelCase : Optional[int] = Node(6) _UpperCamelCase : str = Node(5) _UpperCamelCase : Union[str, Any] = Node(6) print(root_node.has_loop) # False _UpperCamelCase : List[str] = Node(1) print(root_node.has_loop) # False
284
# Algorithm for the pigeonhole sorting def _SCREAMING_SNAKE_CASE ( lowercase : str ): '''simple docstring''' lowerCamelCase_ = min(lowercase ) # min() finds the minimum value lowerCamelCase_ = max(lowercase ) # max() finds the maximum value lowerCamelCase_ = max_val - min_val + 1 # size is difference of max and min values plus one # list of pigeonholes of size equal to the variable size lowerCamelCase_ = [0] * size # Populate the pigeonholes. for x in a: assert isinstance(lowercase , lowercase ), "integers only please" holes[x - min_val] += 1 # Putting the elements back into the array in an order. lowerCamelCase_ = 0 for count in range(lowercase ): while holes[count] > 0: holes[count] -= 1 lowerCamelCase_ = count + min_val i += 1 def _SCREAMING_SNAKE_CASE ( ): '''simple docstring''' lowerCamelCase_ = [8, 3, 2, 7, 4, 6, 8] pigeonhole_sort(lowercase ) print('Sorted order is:' , ' '.join(lowercase ) ) if __name__ == "__main__": main()
70
0
'''simple docstring''' def __UpperCamelCase ( lowercase_ : str ): """simple docstring""" a_ = len(lowercase_ ) a_ = sum(lowercase_ ) a_ = [[False for x in range(s + 1 )] for y in range(n + 1 )] for i in range(1 , n + 1 ): a_ = True for i in range(1 , s + 1 ): a_ = False for i in range(1 , n + 1 ): for j in range(1 , s + 1 ): a_ = dp[i][j - 1] if arr[i - 1] <= j: a_ = dp[i][j] or dp[i - 1][j - arr[i - 1]] for j in range(int(s / 2 ) , -1 , -1 ): if dp[n][j] is True: a_ = s - 2 * j break return diff
536
import os import unittest from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, BertTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class A( UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = BertTokenizer UpperCamelCase = BertTokenizerFast UpperCamelCase = True UpperCamelCase = True UpperCamelCase = filter_non_english def a__ ( self : List[Any] ) -> int: """simple docstring""" super().setUp() lowerCamelCase_ = [ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] lowerCamelCase_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def a__ ( self : Tuple , A_ : List[str] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = 'UNwant\u00E9d,running' lowerCamelCase_ = 'unwanted, running' return input_text, output_text def a__ ( self : Any ) -> Tuple: """simple docstring""" lowerCamelCase_ = self.tokenizer_class(self.vocab_file ) lowerCamelCase_ = tokenizer.tokenize('UNwant\u00E9d,running' ) self.assertListEqual(A_ , ['un', '##want', '##ed', ',', 'runn', '##ing'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A_ ) , [9, 6, 7, 12, 10, 11] ) def a__ ( self : Tuple ) -> List[str]: """simple docstring""" if not self.test_rust_tokenizer: return lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_rust_tokenizer() lowerCamelCase_ = 'UNwant\u00E9d,running' lowerCamelCase_ = tokenizer.tokenize(A_ ) lowerCamelCase_ = rust_tokenizer.tokenize(A_ ) self.assertListEqual(A_ , A_ ) lowerCamelCase_ = tokenizer.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = rust_tokenizer.encode(A_ , add_special_tokens=A_ ) self.assertListEqual(A_ , A_ ) lowerCamelCase_ = self.get_rust_tokenizer() lowerCamelCase_ = tokenizer.encode(A_ ) lowerCamelCase_ = rust_tokenizer.encode(A_ ) self.assertListEqual(A_ , A_ ) # With lower casing lowerCamelCase_ = self.get_tokenizer(do_lower_case=A_ ) lowerCamelCase_ = self.get_rust_tokenizer(do_lower_case=A_ ) lowerCamelCase_ = 'UNwant\u00E9d,running' lowerCamelCase_ = tokenizer.tokenize(A_ ) lowerCamelCase_ = rust_tokenizer.tokenize(A_ ) self.assertListEqual(A_ , A_ ) lowerCamelCase_ = tokenizer.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = rust_tokenizer.encode(A_ , add_special_tokens=A_ ) self.assertListEqual(A_ , A_ ) lowerCamelCase_ = self.get_rust_tokenizer() lowerCamelCase_ = tokenizer.encode(A_ ) lowerCamelCase_ = rust_tokenizer.encode(A_ ) self.assertListEqual(A_ , A_ ) def a__ ( self : Any ) -> Dict: """simple docstring""" lowerCamelCase_ = BasicTokenizer() self.assertListEqual(tokenizer.tokenize('ah\u535A\u63A8zz' ) , ['ah', '\u535A', '\u63A8', 'zz'] ) def a__ ( self : Dict ) -> int: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? ' ) , ['hello', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['hello'] ) def a__ ( self : str ) -> int: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ , strip_accents=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['hällo', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['h\u00E9llo'] ) def a__ ( self : Optional[int] ) -> Any: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ , strip_accents=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['hallo', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['hello'] ) def a__ ( self : Tuple ) -> str: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['hallo', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['hello'] ) def a__ ( self : int ) -> List[Any]: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? ' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?'] ) def a__ ( self : str ) -> List[str]: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ , strip_accents=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['HäLLo', '!', 'how', 'Are', 'yoU', '?'] ) def a__ ( self : Dict ) -> Any: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ , strip_accents=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['HaLLo', '!', 'how', 'Are', 'yoU', '?'] ) def a__ ( self : int ) -> Any: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ , never_split=['[UNK]'] ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? [UNK]' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?', '[UNK]'] ) def a__ ( self : List[Any] ) -> int: """simple docstring""" lowerCamelCase_ = BasicTokenizer() lowerCamelCase_ = 'a\n\'ll !!to?\'d of, can\'t.' lowerCamelCase_ = ['a', '\'', 'll', '!', '!', 'to', '?', '\'', 'd', 'of', ',', 'can', '\'', 't', '.'] self.assertListEqual(tokenizer.tokenize(A_ ) , A_ ) def a__ ( self : Optional[int] ) -> Dict: """simple docstring""" lowerCamelCase_ = ['[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing'] lowerCamelCase_ = {} for i, token in enumerate(A_ ): lowerCamelCase_ = i lowerCamelCase_ = WordpieceTokenizer(vocab=A_ , unk_token='[UNK]' ) self.assertListEqual(tokenizer.tokenize('' ) , [] ) self.assertListEqual(tokenizer.tokenize('unwanted running' ) , ['un', '##want', '##ed', 'runn', '##ing'] ) self.assertListEqual(tokenizer.tokenize('unwantedX running' ) , ['[UNK]', 'runn', '##ing'] ) def a__ ( self : Optional[Any] ) -> str: """simple docstring""" self.assertTrue(_is_whitespace(' ' ) ) self.assertTrue(_is_whitespace('\t' ) ) self.assertTrue(_is_whitespace('\r' ) ) self.assertTrue(_is_whitespace('\n' ) ) self.assertTrue(_is_whitespace('\u00A0' ) ) self.assertFalse(_is_whitespace('A' ) ) self.assertFalse(_is_whitespace('-' ) ) def a__ ( self : List[Any] ) -> int: """simple docstring""" self.assertTrue(_is_control('\u0005' ) ) self.assertFalse(_is_control('A' ) ) self.assertFalse(_is_control(' ' ) ) self.assertFalse(_is_control('\t' ) ) self.assertFalse(_is_control('\r' ) ) def a__ ( self : List[str] ) -> Optional[Any]: """simple docstring""" self.assertTrue(_is_punctuation('-' ) ) self.assertTrue(_is_punctuation('$' ) ) self.assertTrue(_is_punctuation('`' ) ) self.assertTrue(_is_punctuation('.' ) ) self.assertFalse(_is_punctuation('A' ) ) self.assertFalse(_is_punctuation(' ' ) ) def a__ ( self : int ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_rust_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(A_ ) for t in ['Test', '\xad', 'test']] , [['[UNK]'], [], ['[UNK]']] ) self.assertListEqual( [rust_tokenizer.tokenize(A_ ) for t in ['Test', '\xad', 'test']] , [['[UNK]'], [], ['[UNK]']] ) @slow def a__ ( self : Any ) -> int: """simple docstring""" lowerCamelCase_ = self.tokenizer_class.from_pretrained('bert-base-uncased' ) lowerCamelCase_ = tokenizer.encode('sequence builders' , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer.encode('multi-sequence build' , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer.build_inputs_with_special_tokens(A_ ) lowerCamelCase_ = tokenizer.build_inputs_with_special_tokens(A_ , A_ ) assert encoded_sentence == [101] + text + [102] assert encoded_pair == [101] + text + [102] + text_a + [102] def a__ ( self : str ) -> str: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCamelCase_ = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) lowerCamelCase_ = f"""A, naïve {tokenizer_r.mask_token} AllenNLP sentence.""" lowerCamelCase_ = tokenizer_r.encode_plus( A_ , return_attention_mask=A_ , return_token_type_ids=A_ , return_offsets_mapping=A_ , add_special_tokens=A_ , ) lowerCamelCase_ = tokenizer_r.do_lower_case if hasattr(A_ , 'do_lower_case' ) else False lowerCamelCase_ = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), 'A'), ((1, 2), ','), ((3, 5), 'na'), ((5, 6), '##ï'), ((6, 8), '##ve'), ((9, 15), tokenizer_r.mask_token), ((16, 21), 'Allen'), ((21, 23), '##NL'), ((23, 24), '##P'), ((25, 33), 'sentence'), ((33, 34), '.'), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), 'a'), ((1, 2), ','), ((3, 8), 'naive'), ((9, 15), tokenizer_r.mask_token), ((16, 21), 'allen'), ((21, 23), '##nl'), ((23, 24), '##p'), ((25, 33), 'sentence'), ((33, 34), '.'), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens['input_ids'] ) ) self.assertEqual([e[0] for e in expected_results] , tokens['offset_mapping'] ) def a__ ( self : List[Any] ) -> Dict: """simple docstring""" lowerCamelCase_ = ['的', '人', '有'] lowerCamelCase_ = ''.join(A_ ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCamelCase_ = True lowerCamelCase_ = self.tokenizer_class.from_pretrained(A_ , **A_ ) lowerCamelCase_ = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) lowerCamelCase_ = tokenizer_p.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer_r.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer_r.convert_ids_to_tokens(A_ ) lowerCamelCase_ = tokenizer_p.convert_ids_to_tokens(A_ ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(A_ , A_ ) self.assertListEqual(A_ , A_ ) lowerCamelCase_ = False lowerCamelCase_ = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) lowerCamelCase_ = self.tokenizer_class.from_pretrained(A_ , **A_ ) lowerCamelCase_ = tokenizer_r.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer_p.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer_r.convert_ids_to_tokens(A_ ) lowerCamelCase_ = tokenizer_p.convert_ids_to_tokens(A_ ) # it is expected that only the first Chinese character is not preceded by "##". lowerCamelCase_ = [ f"""##{token}""" if idx != 0 else token for idx, token in enumerate(A_ ) ] self.assertListEqual(A_ , A_ ) self.assertListEqual(A_ , A_ )
70
0
from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class __lowerCAmelCase ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): """simple docstring""" A__ : List[Any] = [R"h\.\d+\.attn\.bias", R"h\.\d+\.attn\.masked_bias"] @register_to_config def __init__( self : List[str] , _snake_case : int , _snake_case : int , _snake_case : Optional[int] = None , _snake_case : int = 5_02_57 , _snake_case : int = 10_24 , _snake_case : int = 7_68 , _snake_case : int = 12 , _snake_case : int = 12 , _snake_case : Optional[int] = None , _snake_case : str = "gelu_new" , _snake_case : float = 0.1 , _snake_case : float = 0.1 , _snake_case : float = 0.1 , _snake_case : float = 1E-5 , _snake_case : float = 0.02 , _snake_case : bool = True , _snake_case : bool = True , _snake_case : bool = False , _snake_case : bool = False , ): """simple docstring""" super().__init__() A__ = prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( F'''`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and''' F''' `n_embd`: {n_embd} are not equal.''' ) A__ = prefix_inner_dim A__ = prefix_hidden_dim A__ = ( nn.Linear(self.prefix_inner_dim , self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) A__ = ( nn.Linear(self.prefix_hidden_dim , A_ ) if self.prefix_hidden_dim is not None else nn.Identity() ) A__ = GPTaConfig( vocab_size=A_ , n_positions=A_ , n_embd=A_ , n_layer=A_ , n_head=A_ , n_inner=A_ , activation_function=A_ , resid_pdrop=A_ , embd_pdrop=A_ , attn_pdrop=A_ , layer_norm_epsilon=A_ , initializer_range=A_ , scale_attn_weights=A_ , use_cache=A_ , scale_attn_by_inverse_layer_idx=A_ , reorder_and_upcast_attn=A_ , ) A__ = GPTaLMHeadModel(A_ ) def _a ( self : Optional[int] , _snake_case : torch.Tensor , _snake_case : torch.Tensor , _snake_case : Optional[torch.Tensor] = None , _snake_case : Optional[torch.Tensor] = None , ): """simple docstring""" A__ = self.transformer.transformer.wte(A_ ) A__ = self.encode_prefix(A_ ) A__ = self.decode_prefix(A_ ) A__ = torch.cat((prefix_embeds, embedding_text) , dim=1 ) if labels is not None: A__ = self.get_dummy_token(input_ids.shape[0] , input_ids.device ) A__ = torch.cat((dummy_token, input_ids) , dim=1 ) A__ = self.transformer(inputs_embeds=A_ , labels=A_ , attention_mask=A_ ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def _a ( self : Tuple , _snake_case : int , _snake_case : torch.device ): """simple docstring""" return torch.zeros(A_ , self.prefix_length , dtype=torch.intaa , device=A_ ) def _a ( self : Tuple , _snake_case : Dict ): """simple docstring""" return self.encode_prefix(A_ ) @torch.no_grad() def _a ( self : List[str] , _snake_case : List[str] , _snake_case : Union[str, Any] , _snake_case : Any ): """simple docstring""" A__ = torch.split(A_ , 1 , dim=0 ) A__ = [] A__ = [] for feature in features: A__ = self.decode_prefix(feature.to(A_ ) ) # back to the clip feature # Only support beam search for now A__ , A__ = self.generate_beam( input_embeds=A_ , device=A_ , eos_token_id=A_ ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) A__ = torch.stack(A_ ) A__ = torch.stack(A_ ) return generated_tokens, generated_seq_lengths @torch.no_grad() def _a ( self : int , _snake_case : Dict=None , _snake_case : Optional[int]=None , _snake_case : str=None , _snake_case : int = 5 , _snake_case : int = 67 , _snake_case : float = 1.0 , _snake_case : Optional[int] = None , ): """simple docstring""" A__ = eos_token_id A__ = None A__ = None A__ = torch.ones(A_ , device=A_ , dtype=torch.int ) A__ = torch.zeros(A_ , device=A_ , dtype=torch.bool ) if input_embeds is not None: A__ = input_embeds else: A__ = self.transformer.transformer.wte(A_ ) for i in range(A_ ): A__ = self.transformer(inputs_embeds=A_ ) A__ = outputs.logits A__ = logits[:, -1, :] / (temperature if temperature > 0 else 1.0) A__ = logits.softmax(-1 ).log() if scores is None: A__ , A__ = logits.topk(A_ , -1 ) A__ = generated.expand(A_ , *generated.shape[1:] ) A__ , A__ = next_tokens.permute(1 , 0 ), scores.squeeze(0 ) if tokens is None: A__ = next_tokens else: A__ = tokens.expand(A_ , *tokens.shape[1:] ) A__ = torch.cat((tokens, next_tokens) , dim=1 ) else: A__ = -float(np.inf ) A__ = 0 A__ = scores[:, None] + logits seq_lengths[~is_stopped] += 1 A__ = scores_sum / seq_lengths[:, None] A__ , A__ = scores_sum_average.view(-1 ).topk(A_ , -1 ) A__ = next_tokens // scores_sum.shape[1] A__ = seq_lengths[next_tokens_source] A__ = next_tokens % scores_sum.shape[1] A__ = next_tokens.unsqueeze(1 ) A__ = tokens[next_tokens_source] A__ = torch.cat((tokens, next_tokens) , dim=1 ) A__ = generated[next_tokens_source] A__ = scores_sum_average * seq_lengths A__ = is_stopped[next_tokens_source] A__ = self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] , 1 , -1 ) A__ = torch.cat((generated, next_token_embed) , dim=1 ) A__ = is_stopped + next_tokens.eq(A_ ).squeeze() if is_stopped.all(): break A__ = scores / seq_lengths A__ = scores.argsort(descending=A_ ) # tokens tensors are already padded to max_seq_length A__ = [tokens[i] for i in order] A__ = torch.stack(A_ , dim=0 ) A__ = torch.tensor([seq_lengths[i] for i in order] , dtype=seq_lengths.dtype ) return output_texts, seq_lengths
9
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) lowerCamelCase : Dict = logging.get_logger(__name__) # pylint: disable=invalid-name lowerCamelCase : List[str] = "\n Examples:\n ```py\n >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline\n >>> import torch\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-prior\")\n >>> pipe_prior.to(\"cuda\")\n >>> prompt = \"red cat, 4k photo\"\n >>> out = pipe_prior(prompt)\n >>> image_emb = out.image_embeds\n >>> zero_image_emb = out.negative_image_embeds\n >>> pipe = KandinskyV22Pipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-decoder\")\n >>> pipe.to(\"cuda\")\n >>> image = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=50,\n ... ).images\n >>> image[0].save(\"cat.png\")\n ```\n" def _SCREAMING_SNAKE_CASE ( lowercase : Any , lowercase : str , lowercase : Any=8 ): '''simple docstring''' lowerCamelCase_ = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 lowerCamelCase_ = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class A( UpperCamelCase ): '''simple docstring''' def __init__( self : str , A_ : UNetaDConditionModel , A_ : DDPMScheduler , A_ : VQModel , ) -> List[str]: """simple docstring""" super().__init__() self.register_modules( unet=A_ , scheduler=A_ , movq=A_ , ) lowerCamelCase_ = 2 ** (len(self.movq.config.block_out_channels ) - 1) def a__ ( self : List[Any] , A_ : Tuple , A_ : Dict , A_ : List[Any] , A_ : int , A_ : Any , A_ : Tuple ) -> Any: """simple docstring""" if latents is None: lowerCamelCase_ = randn_tensor(A_ , generator=A_ , device=A_ , dtype=A_ ) else: if latents.shape != shape: raise ValueError(f"""Unexpected latents shape, got {latents.shape}, expected {shape}""" ) lowerCamelCase_ = latents.to(A_ ) lowerCamelCase_ = latents * scheduler.init_noise_sigma return latents def a__ ( self : int , A_ : str=0 ) -> Optional[int]: """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) lowerCamelCase_ = torch.device(f"""cuda:{gpu_id}""" ) lowerCamelCase_ = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(A_ , A_ ) def a__ ( self : Tuple , A_ : Union[str, Any]=0 ) -> Dict: """simple docstring""" if is_accelerate_available() and is_accelerate_version('>=' , '0.17.0.dev0' ): from accelerate import cpu_offload_with_hook else: raise ImportError('`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.' ) lowerCamelCase_ = torch.device(f"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to('cpu' , silence_dtype_warnings=A_ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) lowerCamelCase_ = None for cpu_offloaded_model in [self.unet, self.movq]: lowerCamelCase_ , lowerCamelCase_ = cpu_offload_with_hook(A_ , A_ , prev_module_hook=A_ ) # We'll offload the last model manually. lowerCamelCase_ = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def a__ ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" if not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(A_ , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(A_ ) def __call__( self : List[Any] , A_ : Union[torch.FloatTensor, List[torch.FloatTensor]] , A_ : Union[torch.FloatTensor, List[torch.FloatTensor]] , A_ : int = 512 , A_ : int = 512 , A_ : int = 100 , A_ : float = 4.0 , A_ : int = 1 , A_ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , A_ : Optional[torch.FloatTensor] = None , A_ : Optional[str] = "pil" , A_ : bool = True , ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self._execution_device lowerCamelCase_ = guidance_scale > 1.0 if isinstance(A_ , A_ ): lowerCamelCase_ = torch.cat(A_ , dim=0 ) lowerCamelCase_ = image_embeds.shape[0] * num_images_per_prompt if isinstance(A_ , A_ ): lowerCamelCase_ = torch.cat(A_ , dim=0 ) if do_classifier_free_guidance: lowerCamelCase_ = image_embeds.repeat_interleave(A_ , dim=0 ) lowerCamelCase_ = negative_image_embeds.repeat_interleave(A_ , dim=0 ) lowerCamelCase_ = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=A_ ) self.scheduler.set_timesteps(A_ , device=A_ ) lowerCamelCase_ = self.scheduler.timesteps lowerCamelCase_ = self.unet.config.in_channels lowerCamelCase_ , lowerCamelCase_ = downscale_height_and_width(A_ , A_ , self.movq_scale_factor ) # create initial latent lowerCamelCase_ = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , A_ , A_ , A_ , self.scheduler , ) for i, t in enumerate(self.progress_bar(A_ ) ): # expand the latents if we are doing classifier free guidance lowerCamelCase_ = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowerCamelCase_ = {'image_embeds': image_embeds} lowerCamelCase_ = self.unet( sample=A_ , timestep=A_ , encoder_hidden_states=A_ , added_cond_kwargs=A_ , return_dict=A_ , )[0] if do_classifier_free_guidance: lowerCamelCase_ , lowerCamelCase_ = noise_pred.split(latents.shape[1] , dim=1 ) lowerCamelCase_ , lowerCamelCase_ = noise_pred.chunk(2 ) lowerCamelCase_ , lowerCamelCase_ = variance_pred.chunk(2 ) lowerCamelCase_ = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) lowerCamelCase_ = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , 'variance_type' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): lowerCamelCase_ , lowerCamelCase_ = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 lowerCamelCase_ = self.scheduler.step( A_ , A_ , A_ , generator=A_ , )[0] # post-processing lowerCamelCase_ = self.movq.decode(A_ , force_not_quantize=A_ )['sample'] if output_type not in ["pt", "np", "pil"]: raise ValueError(f"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: lowerCamelCase_ = image * 0.5 + 0.5 lowerCamelCase_ = image.clamp(0 , 1 ) lowerCamelCase_ = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowerCamelCase_ = self.numpy_to_pil(A_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=A_ )
70
0
'''simple docstring''' def a ( _UpperCAmelCase ) -> Any: """simple docstring""" a_ = len(_UpperCAmelCase ) for _ in range(_UpperCAmelCase ): for i in range(_ % 2 , arr_size - 1 , 2 ): if arr[i + 1] < arr[i]: a_ , a_ = arr[i + 1], arr[i] return arr if __name__ == "__main__": __lowerCAmelCase =list(range(10, 0, -1)) print(f'''Original: {arr}. Sorted: {odd_even_transposition(arr)}''')
697
from PIL import Image def _SCREAMING_SNAKE_CASE ( lowercase : Image ): '''simple docstring''' lowerCamelCase_ , lowerCamelCase_ = image.size lowerCamelCase_ = 0 lowerCamelCase_ = image.load() for i in range(lowercase ): for j in range(lowercase ): lowerCamelCase_ = pixels[j, i] mean += pixel mean //= width * height for j in range(lowercase ): for i in range(lowercase ): lowerCamelCase_ = 2_55 if pixels[i, j] > mean else 0 return image if __name__ == "__main__": lowerCamelCase : Optional[Any] = mean_threshold(Image.open("path_to_image").convert("L")) image.save("output_image_path")
70
0
import csv import tweepy # Twitter API credentials A : Tuple = "" A : str = "" A : Any = "" A : Dict = "" def __lowerCAmelCase ( a__ ) -> str: __a = tweepy.OAuthHandler(a__ , a__ ) auth.set_access_token(a__ , a__ ) __a = tweepy.API(a__ ) # initialize a list to hold all the tweepy Tweets __a = [] # make initial request for most recent tweets (200 is the maximum allowed count) __a = api.user_timeline(screen_name=a__ , count=200 ) # save most recent tweets alltweets.extend(a__ ) # save the id of the oldest tweet less one __a = alltweets[-1].id - 1 # keep grabbing tweets until there are no tweets left to grab while len(a__ ) > 0: print(F"""getting tweets before {oldest}""" ) # all subsequent requests use the max_id param to prevent duplicates __a = api.user_timeline( screen_name=a__ , count=200 , max_id=a__ ) # save most recent tweets alltweets.extend(a__ ) # update the id of the oldest tweet less one __a = alltweets[-1].id - 1 print(F"""...{len(a__ )} tweets downloaded so far""" ) # transform the tweepy tweets into a 2D array that will populate the csv __a = [[tweet.id_str, tweet.created_at, tweet.text] for tweet in alltweets] # write the csv with open(F"""new_{screen_name}_tweets.csv""" , '''w''' ) as f: __a = csv.writer(a__ ) writer.writerow(['''id''', '''created_at''', '''text'''] ) writer.writerows(a__ ) if __name__ == "__main__": # pass in the username of the account you want to download get_all_tweets('FirePing32')
219
# this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under git won't be considered # since the output of this script is fed into Makefile commands it doesn't print a newline after the results import re import subprocess import sys lowerCamelCase : List[Any] = subprocess.check_output("git merge-base main HEAD".split()).decode("utf-8") lowerCamelCase : Tuple = ( subprocess.check_output(F"""git diff --diff-filter=d --name-only {fork_point_sha}""".split()).decode("utf-8").split() ) lowerCamelCase : Tuple = "|".join(sys.argv[1:]) lowerCamelCase : Any = re.compile(rF"""^({joined_dirs}).*?\.py$""") lowerCamelCase : List[str] = [x for x in modified_files if regex.match(x)] print(" ".join(relevant_modified_files), end="")
70
0
'''simple docstring''' import unittest from parameterized import parameterized from transformers import AutoTokenizer, GPTNeoXConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXModel, ) class a : """simple docstring""" def __init__( self , snake_case_ , snake_case_=13 , snake_case_=7 , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=99 , snake_case_=64 , snake_case_=5 , snake_case_=4 , snake_case_=37 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=512 , snake_case_=16 , snake_case_=2 , snake_case_=0.0_2 , snake_case_=3 , snake_case_=4 , snake_case_=None , ): '''simple docstring''' __UpperCAmelCase: str = parent __UpperCAmelCase: Any = batch_size __UpperCAmelCase: Union[str, Any] = seq_length __UpperCAmelCase: Optional[int] = is_training __UpperCAmelCase: List[str] = use_input_mask __UpperCAmelCase: Optional[int] = use_token_type_ids __UpperCAmelCase: List[str] = use_labels __UpperCAmelCase: int = vocab_size __UpperCAmelCase: Optional[int] = hidden_size __UpperCAmelCase: Optional[Any] = num_hidden_layers __UpperCAmelCase: Optional[Any] = num_attention_heads __UpperCAmelCase: str = intermediate_size __UpperCAmelCase: List[str] = hidden_act __UpperCAmelCase: Tuple = hidden_dropout_prob __UpperCAmelCase: Any = attention_probs_dropout_prob __UpperCAmelCase: int = max_position_embeddings __UpperCAmelCase: Any = type_vocab_size __UpperCAmelCase: Any = type_sequence_label_size __UpperCAmelCase: List[str] = initializer_range __UpperCAmelCase: int = num_labels __UpperCAmelCase: Dict = num_choices __UpperCAmelCase: List[Any] = scope __UpperCAmelCase: int = vocab_size - 1 def lowercase_ ( self ): '''simple docstring''' __UpperCAmelCase: str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __UpperCAmelCase: int = None if self.use_input_mask: __UpperCAmelCase: Optional[int] = random_attention_mask([self.batch_size, self.seq_length] ) __UpperCAmelCase: Dict = None if self.use_labels: __UpperCAmelCase: Tuple = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __UpperCAmelCase: Dict = self.get_config() return config, input_ids, input_mask, token_labels def lowercase_ ( self ): '''simple docstring''' return GPTNeoXConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=A_ , initializer_range=self.initializer_range , pad_token_id=self.pad_token_id , ) def lowercase_ ( self ): '''simple docstring''' __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase: List[Any] = self.prepare_config_and_inputs() __UpperCAmelCase: Optional[int] = True return config, input_ids, input_mask, token_labels def lowercase_ ( self , snake_case_ , snake_case_ , snake_case_ ): '''simple docstring''' __UpperCAmelCase: Any = GPTNeoXModel(config=A_ ) model.to(A_ ) model.eval() __UpperCAmelCase: Dict = model(A_ , attention_mask=A_ ) __UpperCAmelCase: str = model(A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase_ ( self , snake_case_ , snake_case_ , snake_case_ ): '''simple docstring''' __UpperCAmelCase: Any = True __UpperCAmelCase: Optional[Any] = GPTNeoXModel(A_ ) model.to(A_ ) model.eval() __UpperCAmelCase: List[str] = model(A_ , attention_mask=A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase_ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): '''simple docstring''' __UpperCAmelCase: int = GPTNeoXForCausalLM(config=A_ ) model.to(A_ ) model.eval() __UpperCAmelCase: List[str] = model(A_ , attention_mask=A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowercase_ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): '''simple docstring''' __UpperCAmelCase: Any = self.num_labels __UpperCAmelCase: Tuple = GPTNeoXForQuestionAnswering(A_ ) model.to(A_ ) model.eval() __UpperCAmelCase: Optional[int] = model(A_ , attention_mask=A_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowercase_ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): '''simple docstring''' __UpperCAmelCase: Optional[Any] = self.num_labels __UpperCAmelCase: Dict = GPTNeoXForSequenceClassification(A_ ) model.to(A_ ) model.eval() __UpperCAmelCase: Union[str, Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __UpperCAmelCase: List[str] = model(A_ , attention_mask=A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowercase_ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): '''simple docstring''' __UpperCAmelCase: Dict = self.num_labels __UpperCAmelCase: Union[str, Any] = GPTNeoXForTokenClassification(A_ ) model.to(A_ ) model.eval() __UpperCAmelCase: Tuple = model(A_ , attention_mask=A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowercase_ ( self , snake_case_ , snake_case_ , snake_case_ ): '''simple docstring''' __UpperCAmelCase: List[str] = True __UpperCAmelCase: Optional[int] = GPTNeoXForCausalLM(config=A_ ) model.to(A_ ) model.eval() # first forward pass __UpperCAmelCase: Any = model(A_ , attention_mask=A_ , use_cache=A_ ) __UpperCAmelCase: Dict = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids __UpperCAmelCase: Optional[Any] = ids_tensor((self.batch_size, 3) , config.vocab_size ) __UpperCAmelCase: Optional[Any] = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and __UpperCAmelCase: Dict = torch.cat([input_ids, next_tokens] , dim=-1 ) __UpperCAmelCase: Optional[Any] = torch.cat([input_mask, next_mask] , dim=-1 ) __UpperCAmelCase: Tuple = model(A_ , attention_mask=A_ , output_hidden_states=A_ ) __UpperCAmelCase: str = output_from_no_past["""hidden_states"""][0] __UpperCAmelCase: Any = model( A_ , attention_mask=A_ , past_key_values=A_ , output_hidden_states=A_ , )["""hidden_states"""][0] # select random slice __UpperCAmelCase: Union[str, Any] = ids_tensor((1,) , output_from_past.shape[-1] ).item() __UpperCAmelCase: Optional[Any] = output_from_no_past[:, -3:, random_slice_idx].detach() __UpperCAmelCase: str = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(A_ , A_ , atol=1e-3 ) ) def lowercase_ ( self ): '''simple docstring''' __UpperCAmelCase: str = self.prepare_config_and_inputs() __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase: Optional[Any] = config_and_inputs __UpperCAmelCase: Optional[int] = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class a ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): """simple docstring""" __lowerCAmelCase = ( ( GPTNeoXModel, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, ) if is_torch_available() else () ) __lowerCAmelCase = (GPTNeoXForCausalLM,) if is_torch_available() else () __lowerCAmelCase = ( { """feature-extraction""": GPTNeoXModel, """question-answering""": GPTNeoXForQuestionAnswering, """text-classification""": GPTNeoXForSequenceClassification, """text-generation""": GPTNeoXForCausalLM, """token-classification""": GPTNeoXForTokenClassification, """zero-shot""": GPTNeoXForSequenceClassification, } if is_torch_available() else {} ) __lowerCAmelCase = False __lowerCAmelCase = False __lowerCAmelCase = False __lowerCAmelCase = False def lowercase_ ( self ): '''simple docstring''' __UpperCAmelCase: Optional[int] = GPTNeoXModelTester(self ) __UpperCAmelCase: int = ConfigTester(self , config_class=A_ , hidden_size=64 , num_attention_heads=8 ) def lowercase_ ( self ): '''simple docstring''' self.config_tester.run_common_tests() def lowercase_ ( self ): '''simple docstring''' __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase: Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(A_ , A_ , A_ ) def lowercase_ ( self ): '''simple docstring''' __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase: str = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(A_ , A_ , A_ ) def lowercase_ ( self ): '''simple docstring''' __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase: Dict = self.model_tester.prepare_config_and_inputs_for_decoder() __UpperCAmelCase: Dict = None self.model_tester.create_and_check_model_as_decoder(A_ , A_ , A_ ) def lowercase_ ( self ): '''simple docstring''' __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase: int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(A_ , A_ , A_ ) def lowercase_ ( self ): '''simple docstring''' __UpperCAmelCase: Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*A_ ) def lowercase_ ( self ): '''simple docstring''' __UpperCAmelCase: Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*A_ ) def lowercase_ ( self ): '''simple docstring''' __UpperCAmelCase: str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*A_ ) def lowercase_ ( self ): '''simple docstring''' __UpperCAmelCase: Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*A_ ) @unittest.skip(reason="""Feed forward chunking is not implemented""" ) def lowercase_ ( self ): '''simple docstring''' pass @parameterized.expand([("""linear""",), ("""dynamic""",)] ) def lowercase_ ( self , snake_case_ ): '''simple docstring''' __UpperCAmelCase, __UpperCAmelCase: str = self.model_tester.prepare_config_and_inputs_for_common() __UpperCAmelCase: Optional[int] = ids_tensor([1, 10] , config.vocab_size ) __UpperCAmelCase: Tuple = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights __UpperCAmelCase: List[str] = GPTNeoXModel(A_ ) original_model.to(A_ ) original_model.eval() __UpperCAmelCase: List[str] = original_model(A_ ).last_hidden_state __UpperCAmelCase: Any = original_model(A_ ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights __UpperCAmelCase: Optional[int] = {"""type""": scaling_type, """factor""": 1_0.0} __UpperCAmelCase: Optional[int] = GPTNeoXModel(A_ ) scaled_model.to(A_ ) scaled_model.eval() __UpperCAmelCase: Any = scaled_model(A_ ).last_hidden_state __UpperCAmelCase: Any = scaled_model(A_ ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(A_ , A_ , atol=1e-5 ) ) else: self.assertFalse(torch.allclose(A_ , A_ , atol=1e-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(A_ , A_ , atol=1e-5 ) ) @require_torch class a ( unittest.TestCase ): """simple docstring""" @slow def lowercase_ ( self ): '''simple docstring''' __UpperCAmelCase: Tuple = AutoTokenizer.from_pretrained("""EleutherAI/pythia-410m-deduped""" ) for checkpointing in [True, False]: __UpperCAmelCase: Dict = GPTNeoXForCausalLM.from_pretrained("""EleutherAI/pythia-410m-deduped""" ) if checkpointing: model.gradient_checkpointing_enable() else: model.gradient_checkpointing_disable() model.to(A_ ) __UpperCAmelCase: List[Any] = tokenizer("""My favorite food is""" , return_tensors="""pt""" ).to(A_ ) # The hub repo. is updated on 2023-04-04, resulting in poor outputs. # See: https://github.com/huggingface/transformers/pull/24193 __UpperCAmelCase: Tuple = """My favorite food is a good old-fashioned, old-fashioned, old-fashioned.\n\nI\'m not sure""" __UpperCAmelCase: Optional[int] = model.generate(**A_ , do_sample=A_ , max_new_tokens=20 ) __UpperCAmelCase: Optional[Any] = tokenizer.batch_decode(A_ )[0] self.assertEqual(A_ , A_ )
523
import argparse import json import subprocess def _SCREAMING_SNAKE_CASE ( lowercase : Dict , lowercase : List[str] ): '''simple docstring''' lowerCamelCase_ = [] lowerCamelCase_ = ( f"""curl -H \"Accept: application/vnd.github+json\" -H \"Authorization: Bearer {token}\"""" ' https://api.github.com/repos/huggingface/transformers/actions/runners' ) lowerCamelCase_ = subprocess.run(lowercase , shell=lowercase , stdout=subprocess.PIPE ) lowerCamelCase_ = output.stdout.decode('utf-8' ) lowerCamelCase_ = json.loads(lowercase ) lowerCamelCase_ = status['runners'] for runner in runners: if runner["name"] in target_runners: if runner["status"] == "offline": offline_runners.append(lowercase ) # save the result so we can report them on Slack with open('offline_runners.txt' , 'w' ) as fp: fp.write(json.dumps(lowercase ) ) if len(lowercase ) > 0: lowerCamelCase_ = '\n'.join([x['name'] for x in offline_runners] ) raise ValueError(f"""The following runners are offline:\n{failed}""" ) if __name__ == "__main__": def _SCREAMING_SNAKE_CASE ( lowercase : List[str] ): '''simple docstring''' return values.split(',' ) lowerCamelCase : str = argparse.ArgumentParser() # Required parameters parser.add_argument( "--target_runners", default=None, type=list_str, required=True, help="Comma-separated list of runners to check status.", ) parser.add_argument( "--token", default=None, type=str, required=True, help="A token that has actions:read permission." ) lowerCamelCase : Optional[int] = parser.parse_args() get_runner_status(args.target_runners, args.token)
70
0
def _lowerCamelCase ( lowerCamelCase_: Union[str, Any] , lowerCamelCase_: Dict ): '''simple docstring''' return (pointa[0] - pointa[0]) ** 2 + (pointa[1] - pointa[1]) ** 2 def _lowerCamelCase ( lowerCamelCase_: Optional[Any] , lowerCamelCase_: Union[str, Any]=0 ): '''simple docstring''' return sorted(lowerCamelCase_ , key=lambda lowerCamelCase_ : x[column] ) def _lowerCamelCase ( lowerCamelCase_: Tuple , lowerCamelCase_: List[str] , lowerCamelCase_: str=float('''inf''' ) ): '''simple docstring''' for i in range(points_counts - 1 ): for j in range(i + 1 , lowerCamelCase_ ): A : Union[str, Any] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: A : Optional[Any] = current_dis return min_dis def _lowerCamelCase ( lowerCamelCase_: int , lowerCamelCase_: Any , lowerCamelCase_: int=float('''inf''' ) ): '''simple docstring''' for i in range(min(6 , points_counts - 1 ) , lowerCamelCase_ ): for j in range(max(0 , i - 6 ) , lowerCamelCase_ ): A : Any = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: A : Optional[Any] = current_dis return min_dis def _lowerCamelCase ( lowerCamelCase_: str , lowerCamelCase_: Dict , lowerCamelCase_: str ): '''simple docstring''' if points_counts <= 3: return dis_between_closest_pair(lowerCamelCase_ , lowerCamelCase_ ) # recursion A : int = points_counts // 2 A : Optional[Any] = closest_pair_of_points_sqr( lowerCamelCase_ , points_sorted_on_y[:mid] , lowerCamelCase_ ) A : Union[str, Any] = closest_pair_of_points_sqr( lowerCamelCase_ , points_sorted_on_y[mid:] , points_counts - mid ) A : List[str] = min(lowerCamelCase_ , lowerCamelCase_ ) A : Dict = [] for point in points_sorted_on_x: if abs(point[0] - points_sorted_on_x[mid][0] ) < closest_pair_dis: cross_strip.append(lowerCamelCase_ ) A : Optional[int] = dis_between_closest_in_strip( lowerCamelCase_ , len(lowerCamelCase_ ) , lowerCamelCase_ ) return min(lowerCamelCase_ , lowerCamelCase_ ) def _lowerCamelCase ( lowerCamelCase_: Any , lowerCamelCase_: Optional[int] ): '''simple docstring''' A : Dict = column_based_sort(lowerCamelCase_ , column=0 ) A : Tuple = column_based_sort(lowerCamelCase_ , column=1 ) return ( closest_pair_of_points_sqr( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) ) ** 0.5 if __name__ == "__main__": UpperCamelCase_ = [(2, 3), (12, 30), (40, 50), (5, 1), (12, 10), (3, 4)] print("Distance:", closest_pair_of_points(points, len(points)))
256
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from timm import create_model from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import BitConfig, BitForImageClassification, BitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase : Optional[Any] = logging.get_logger(__name__) def _SCREAMING_SNAKE_CASE ( lowercase : List[Any] ): '''simple docstring''' lowerCamelCase_ = 'huggingface/label-files' lowerCamelCase_ = 'imagenet-1k-id2label.json' lowerCamelCase_ = json.load(open(hf_hub_download(lowercase , lowercase , repo_type='dataset' ) , 'r' ) ) lowerCamelCase_ = {int(lowercase ): v for k, v in idalabel.items()} lowerCamelCase_ = {v: k for k, v in idalabel.items()} lowerCamelCase_ = 'std_conv' if 'bit' in model_name else False # note that when using BiT as backbone for ViT-hybrid checkpoints, # one needs to additionally set config.layer_type = "bottleneck", config.stem_type = "same", # config.conv_layer = "std_conv_same" lowerCamelCase_ = BitConfig( conv_layer=lowercase , num_labels=10_00 , idalabel=lowercase , labelaid=lowercase , ) return config def _SCREAMING_SNAKE_CASE ( lowercase : Any ): '''simple docstring''' if "stem.conv" in name: lowerCamelCase_ = name.replace('stem.conv' , 'bit.embedder.convolution' ) if "blocks" in name: lowerCamelCase_ = name.replace('blocks' , 'layers' ) if "head.fc" in name: lowerCamelCase_ = name.replace('head.fc' , 'classifier.1' ) if name.startswith('norm' ): lowerCamelCase_ = 'bit.' + name if "bit" not in name and "classifier" not in name: lowerCamelCase_ = 'bit.encoder.' + name return name def _SCREAMING_SNAKE_CASE ( ): '''simple docstring''' lowerCamelCase_ = 'http://images.cocodataset.org/val2017/000000039769.jpg' lowerCamelCase_ = Image.open(requests.get(lowercase , stream=lowercase ).raw ) return im @torch.no_grad() def _SCREAMING_SNAKE_CASE ( lowercase : str , lowercase : int , lowercase : Union[str, Any]=False ): '''simple docstring''' lowerCamelCase_ = get_config(lowercase ) # load original model from timm lowerCamelCase_ = create_model(lowercase , pretrained=lowercase ) timm_model.eval() # load state_dict of original model lowerCamelCase_ = timm_model.state_dict() for key in state_dict.copy().keys(): lowerCamelCase_ = state_dict.pop(lowercase ) lowerCamelCase_ = val.squeeze() if 'head' in key else val # load HuggingFace model lowerCamelCase_ = BitForImageClassification(lowercase ) model.eval() model.load_state_dict(lowercase ) # create image processor lowerCamelCase_ = create_transform(**resolve_data_config({} , model=lowercase ) ) lowerCamelCase_ = transform.transforms lowerCamelCase_ = { 'bilinear': PILImageResampling.BILINEAR, 'bicubic': PILImageResampling.BICUBIC, 'nearest': PILImageResampling.NEAREST, } lowerCamelCase_ = BitImageProcessor( do_resize=lowercase , size={'shortest_edge': timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=lowercase , crop_size={'height': timm_transforms[1].size[0], 'width': timm_transforms[1].size[1]} , do_normalize=lowercase , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) lowerCamelCase_ = prepare_img() lowerCamelCase_ = transform(lowercase ).unsqueeze(0 ) lowerCamelCase_ = processor(lowercase , return_tensors='pt' ).pixel_values # verify pixel values assert torch.allclose(lowercase , lowercase ) # verify logits with torch.no_grad(): lowerCamelCase_ = model(lowercase ) lowerCamelCase_ = outputs.logits print('Logits:' , logits[0, :3] ) print('Predicted class:' , model.config.idalabel[logits.argmax(-1 ).item()] ) lowerCamelCase_ = timm_model(lowercase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(lowercase , outputs.logits , atol=1e-3 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: Path(lowercase ).mkdir(exist_ok=lowercase ) print(f"""Saving model {model_name} and processor to {pytorch_dump_folder_path}""" ) model.save_pretrained(lowercase ) processor.save_pretrained(lowercase ) if push_to_hub: print(f"""Pushing model {model_name} and processor to the hub""" ) model.push_to_hub(f"""ybelkada/{model_name}""" ) processor.push_to_hub(f"""ybelkada/{model_name}""" ) if __name__ == "__main__": lowerCamelCase : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="resnetv2_50x1_bitm", type=str, help="Name of the BiT timm model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to push the model to the hub.", ) lowerCamelCase : Optional[int] = parser.parse_args() convert_bit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
70
0
"""simple docstring""" import collections.abc from typing import Optional, Tuple, Union import torch import torch.utils.checkpoint from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_poolformer import PoolFormerConfig _lowerCamelCase = logging.get_logger(__name__) # General docstring _lowerCamelCase = "PoolFormerConfig" # Base docstring _lowerCamelCase = "sail/poolformer_s12" _lowerCamelCase = [1, 5_12, 7, 7] # Image classification docstring _lowerCamelCase = "sail/poolformer_s12" _lowerCamelCase = "tabby, tabby cat" _lowerCamelCase = [ "sail/poolformer_s12", # See all PoolFormer models at https://huggingface.co/models?filter=poolformer ] def lowerCAmelCase_ ( lowercase_ : Union[str, Any] , lowercase_ : float = 0.0 , lowercase_ : bool = False ): '''simple docstring''' if drop_prob == 0.0 or not training: return input __SCREAMING_SNAKE_CASE : str = 1 - drop_prob __SCREAMING_SNAKE_CASE : int = (input.shape[0],) + (1,) * (input.ndim - 1) # work with diff dim tensors, not just 2D ConvNets __SCREAMING_SNAKE_CASE : Optional[int] = keep_prob + torch.rand(lowercase_ , dtype=input.dtype , device=input.device ) random_tensor.floor_() # binarize __SCREAMING_SNAKE_CASE : Union[str, Any] = input.div(lowercase_ ) * random_tensor return output class snake_case ( nn.Module ): def __init__( self :Tuple , _lowerCamelCase :Optional[float] = None ): super().__init__() __SCREAMING_SNAKE_CASE : Any = drop_prob def SCREAMING_SNAKE_CASE_ ( self :List[Any] , _lowerCamelCase :torch.Tensor ): return drop_path(A_ , self.drop_prob , self.training ) def SCREAMING_SNAKE_CASE_ ( self :Tuple ): return "p={}".format(self.drop_prob ) class snake_case ( nn.Module ): def __init__( self :List[str] , _lowerCamelCase :Optional[Any] , _lowerCamelCase :List[Any] , _lowerCamelCase :int , _lowerCamelCase :List[str] , _lowerCamelCase :Any , _lowerCamelCase :Tuple=None ): super().__init__() __SCREAMING_SNAKE_CASE : Union[str, Any] = patch_size if isinstance(A_ , collections.abc.Iterable ) else (patch_size, patch_size) __SCREAMING_SNAKE_CASE : Union[str, Any] = stride if isinstance(A_ , collections.abc.Iterable ) else (stride, stride) __SCREAMING_SNAKE_CASE : int = padding if isinstance(A_ , collections.abc.Iterable ) else (padding, padding) __SCREAMING_SNAKE_CASE : Optional[Any] = nn.Convad(A_ , A_ , kernel_size=A_ , stride=A_ , padding=A_ ) __SCREAMING_SNAKE_CASE : Dict = norm_layer(A_ ) if norm_layer else nn.Identity() def SCREAMING_SNAKE_CASE_ ( self :Tuple , _lowerCamelCase :str ): __SCREAMING_SNAKE_CASE : int = self.projection(A_ ) __SCREAMING_SNAKE_CASE : Optional[Any] = self.norm(A_ ) return embeddings class snake_case ( nn.GroupNorm ): def __init__( self :Tuple , _lowerCamelCase :Optional[int] , **_lowerCamelCase :Dict ): super().__init__(1 , A_ , **A_ ) class snake_case ( nn.Module ): def __init__( self :Union[str, Any] , _lowerCamelCase :Optional[int] ): super().__init__() __SCREAMING_SNAKE_CASE : int = nn.AvgPoolad(A_ , stride=1 , padding=pool_size // 2 , count_include_pad=A_ ) def SCREAMING_SNAKE_CASE_ ( self :List[str] , _lowerCamelCase :Union[str, Any] ): return self.pool(A_ ) - hidden_states class snake_case ( nn.Module ): def __init__( self :Union[str, Any] , _lowerCamelCase :Optional[Any] , _lowerCamelCase :List[Any] , _lowerCamelCase :Dict , _lowerCamelCase :str ): super().__init__() __SCREAMING_SNAKE_CASE : List[str] = nn.Convad(A_ , A_ , 1 ) __SCREAMING_SNAKE_CASE : List[Any] = nn.Convad(A_ , A_ , 1 ) __SCREAMING_SNAKE_CASE : str = PoolFormerDropPath(A_ ) if isinstance(config.hidden_act , A_ ): __SCREAMING_SNAKE_CASE : List[Any] = ACTaFN[config.hidden_act] else: __SCREAMING_SNAKE_CASE : List[str] = config.hidden_act def SCREAMING_SNAKE_CASE_ ( self :List[str] , _lowerCamelCase :int ): __SCREAMING_SNAKE_CASE : Optional[int] = self.conva(A_ ) __SCREAMING_SNAKE_CASE : Dict = self.act_fn(A_ ) __SCREAMING_SNAKE_CASE : Dict = self.drop(A_ ) __SCREAMING_SNAKE_CASE : Any = self.conva(A_ ) __SCREAMING_SNAKE_CASE : List[Any] = self.drop(A_ ) return hidden_states class snake_case ( nn.Module ): def __init__( self :Union[str, Any] , _lowerCamelCase :List[str] , _lowerCamelCase :Union[str, Any] , _lowerCamelCase :Dict , _lowerCamelCase :Dict , _lowerCamelCase :int , _lowerCamelCase :int ): super().__init__() __SCREAMING_SNAKE_CASE : List[str] = PoolFormerPooling(A_ ) __SCREAMING_SNAKE_CASE : Optional[Any] = PoolFormerOutput(A_ , A_ , A_ , A_ ) __SCREAMING_SNAKE_CASE : Tuple = PoolFormerGroupNorm(A_ ) __SCREAMING_SNAKE_CASE : str = PoolFormerGroupNorm(A_ ) # Useful for training neural nets __SCREAMING_SNAKE_CASE : Optional[int] = PoolFormerDropPath(A_ ) if drop_path > 0.0 else nn.Identity() __SCREAMING_SNAKE_CASE : Optional[int] = config.use_layer_scale if config.use_layer_scale: __SCREAMING_SNAKE_CASE : int = nn.Parameter( config.layer_scale_init_value * torch.ones((A_) ) , requires_grad=A_ ) __SCREAMING_SNAKE_CASE : str = nn.Parameter( config.layer_scale_init_value * torch.ones((A_) ) , requires_grad=A_ ) def SCREAMING_SNAKE_CASE_ ( self :Dict , _lowerCamelCase :str ): if self.use_layer_scale: __SCREAMING_SNAKE_CASE : Any = self.pooling(self.before_norm(A_ ) ) __SCREAMING_SNAKE_CASE : Optional[Any] = self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * pooling_output # First residual connection __SCREAMING_SNAKE_CASE : Union[str, Any] = hidden_states + self.drop_path(A_ ) __SCREAMING_SNAKE_CASE : List[Any] = () __SCREAMING_SNAKE_CASE : Optional[Any] = self.output(self.after_norm(A_ ) ) __SCREAMING_SNAKE_CASE : Tuple = self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * layer_output # Second residual connection __SCREAMING_SNAKE_CASE : Union[str, Any] = hidden_states + self.drop_path(A_ ) __SCREAMING_SNAKE_CASE : Tuple = (output,) + outputs return outputs else: __SCREAMING_SNAKE_CASE : Any = self.drop_path(self.pooling(self.before_norm(A_ ) ) ) # First residual connection __SCREAMING_SNAKE_CASE : Tuple = pooling_output + hidden_states __SCREAMING_SNAKE_CASE : Tuple = () # Second residual connection inside the PoolFormerOutput block __SCREAMING_SNAKE_CASE : List[Any] = self.drop_path(self.output(self.after_norm(A_ ) ) ) __SCREAMING_SNAKE_CASE : Tuple = hidden_states + layer_output __SCREAMING_SNAKE_CASE : Dict = (output,) + outputs return outputs class snake_case ( nn.Module ): def __init__( self :Tuple , _lowerCamelCase :List[str] ): super().__init__() __SCREAMING_SNAKE_CASE : List[str] = config # stochastic depth decay rule __SCREAMING_SNAKE_CASE : Union[str, Any] = [x.item() for x in torch.linspace(0 , config.drop_path_rate , sum(config.depths ) )] # patch embeddings __SCREAMING_SNAKE_CASE : Union[str, Any] = [] for i in range(config.num_encoder_blocks ): embeddings.append( PoolFormerEmbeddings( patch_size=config.patch_sizes[i] , stride=config.strides[i] , padding=config.padding[i] , num_channels=config.num_channels if i == 0 else config.hidden_sizes[i - 1] , hidden_size=config.hidden_sizes[i] , ) ) __SCREAMING_SNAKE_CASE : Any = nn.ModuleList(A_ ) # Transformer blocks __SCREAMING_SNAKE_CASE : Any = [] __SCREAMING_SNAKE_CASE : int = 0 for i in range(config.num_encoder_blocks ): # each block consists of layers __SCREAMING_SNAKE_CASE : List[Any] = [] if i != 0: cur += config.depths[i - 1] for j in range(config.depths[i] ): layers.append( PoolFormerLayer( A_ , num_channels=config.hidden_sizes[i] , pool_size=config.pool_size , hidden_size=config.hidden_sizes[i] , intermediate_size=int(config.hidden_sizes[i] * config.mlp_ratio ) , drop_path=dpr[cur + j] , ) ) blocks.append(nn.ModuleList(A_ ) ) __SCREAMING_SNAKE_CASE : List[str] = nn.ModuleList(A_ ) def SCREAMING_SNAKE_CASE_ ( self :str , _lowerCamelCase :Union[str, Any] , _lowerCamelCase :Dict=False , _lowerCamelCase :Union[str, Any]=True ): __SCREAMING_SNAKE_CASE : Tuple = () if output_hidden_states else None __SCREAMING_SNAKE_CASE : Union[str, Any] = pixel_values for idx, layers in enumerate(zip(self.patch_embeddings , self.block ) ): __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Dict = layers # Get patch embeddings from hidden_states __SCREAMING_SNAKE_CASE : Optional[Any] = embedding_layer(A_ ) # Send the embeddings through the blocks for _, blk in enumerate(A_ ): __SCREAMING_SNAKE_CASE : Any = blk(A_ ) __SCREAMING_SNAKE_CASE : str = layer_outputs[0] if output_hidden_states: __SCREAMING_SNAKE_CASE : Union[str, Any] = all_hidden_states + (hidden_states,) if not return_dict: return tuple(v for v in [hidden_states, all_hidden_states] if v is not None ) return BaseModelOutputWithNoAttention(last_hidden_state=A_ , hidden_states=A_ ) class snake_case ( __UpperCAmelCase ): lowerCamelCase__ = PoolFormerConfig lowerCamelCase__ = '''poolformer''' lowerCamelCase__ = '''pixel_values''' lowerCamelCase__ = True def SCREAMING_SNAKE_CASE_ ( self :Optional[Any] , _lowerCamelCase :Any ): if isinstance(A_ , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(A_ , nn.LayerNorm ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) def SCREAMING_SNAKE_CASE_ ( self :Dict , _lowerCamelCase :Union[str, Any] , _lowerCamelCase :Dict=False ): if isinstance(A_ , A_ ): __SCREAMING_SNAKE_CASE : Tuple = value _lowerCamelCase = r"\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n\n Parameters:\n config ([`PoolFormerConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" _lowerCamelCase = r"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`PoolFormerImageProcessor.__call__`] for details.\n" @add_start_docstrings( '''The bare PoolFormer Model transformer outputting raw hidden-states without any specific head on top.''' , __UpperCAmelCase , ) class snake_case ( __UpperCAmelCase ): def __init__( self :Dict , _lowerCamelCase :Any ): super().__init__(A_ ) __SCREAMING_SNAKE_CASE : Optional[Any] = config __SCREAMING_SNAKE_CASE : Union[str, Any] = PoolFormerEncoder(A_ ) # Initialize weights and apply final processing self.post_init() def SCREAMING_SNAKE_CASE_ ( self :Optional[Any] ): return self.embeddings.patch_embeddings @add_start_docstrings_to_model_forward(A_ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=A_ , config_class=_CONFIG_FOR_DOC , modality='''vision''' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def SCREAMING_SNAKE_CASE_ ( self :Union[str, Any] , _lowerCamelCase :Optional[torch.FloatTensor] = None , _lowerCamelCase :Optional[bool] = None , _lowerCamelCase :Optional[bool] = None , ): __SCREAMING_SNAKE_CASE : Dict = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __SCREAMING_SNAKE_CASE : str = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError('''You have to specify pixel_values''' ) __SCREAMING_SNAKE_CASE : List[str] = self.encoder( A_ , output_hidden_states=A_ , return_dict=A_ , ) __SCREAMING_SNAKE_CASE : Dict = encoder_outputs[0] if not return_dict: return (sequence_output, None) + encoder_outputs[1:] return BaseModelOutputWithNoAttention( last_hidden_state=A_ , hidden_states=encoder_outputs.hidden_states , ) class snake_case ( nn.Module ): def __init__( self :List[Any] , _lowerCamelCase :Optional[Any] ): super().__init__() __SCREAMING_SNAKE_CASE : Optional[Any] = nn.Linear(config.hidden_size , config.hidden_size ) def SCREAMING_SNAKE_CASE_ ( self :Any , _lowerCamelCase :str ): __SCREAMING_SNAKE_CASE : Dict = self.dense(A_ ) return output @add_start_docstrings( ''' PoolFormer Model transformer with an image classification head on top ''' , __UpperCAmelCase , ) class snake_case ( __UpperCAmelCase ): def __init__( self :Dict , _lowerCamelCase :Dict ): super().__init__(A_ ) __SCREAMING_SNAKE_CASE : Dict = config.num_labels __SCREAMING_SNAKE_CASE : List[Any] = PoolFormerModel(A_ ) # Final norm __SCREAMING_SNAKE_CASE : Dict = PoolFormerGroupNorm(config.hidden_sizes[-1] ) # Classifier head __SCREAMING_SNAKE_CASE : Union[str, Any] = ( nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(A_ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=A_ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def SCREAMING_SNAKE_CASE_ ( self :Optional[Any] , _lowerCamelCase :Optional[torch.FloatTensor] = None , _lowerCamelCase :Optional[torch.LongTensor] = None , _lowerCamelCase :Optional[bool] = None , _lowerCamelCase :Optional[bool] = None , ): __SCREAMING_SNAKE_CASE : Dict = return_dict if return_dict is not None else self.config.use_return_dict __SCREAMING_SNAKE_CASE : int = self.poolformer( A_ , output_hidden_states=A_ , return_dict=A_ , ) __SCREAMING_SNAKE_CASE : List[Any] = outputs[0] __SCREAMING_SNAKE_CASE : Dict = self.classifier(self.norm(A_ ).mean([-2, -1] ) ) __SCREAMING_SNAKE_CASE : Any = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: __SCREAMING_SNAKE_CASE : List[str] = '''regression''' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): __SCREAMING_SNAKE_CASE : Optional[Any] = '''single_label_classification''' else: __SCREAMING_SNAKE_CASE : Tuple = '''multi_label_classification''' if self.config.problem_type == "regression": __SCREAMING_SNAKE_CASE : Dict = MSELoss() if self.num_labels == 1: __SCREAMING_SNAKE_CASE : str = loss_fct(logits.squeeze() , labels.squeeze() ) else: __SCREAMING_SNAKE_CASE : str = loss_fct(A_ , A_ ) elif self.config.problem_type == "single_label_classification": __SCREAMING_SNAKE_CASE : Tuple = CrossEntropyLoss() __SCREAMING_SNAKE_CASE : Any = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": __SCREAMING_SNAKE_CASE : Dict = BCEWithLogitsLoss() __SCREAMING_SNAKE_CASE : List[str] = loss_fct(A_ , A_ ) if not return_dict: __SCREAMING_SNAKE_CASE : Dict = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=A_ , logits=A_ , hidden_states=outputs.hidden_states )
674
from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class A: '''simple docstring''' def __init__( self : List[Any] , A_ : Dict , A_ : Union[str, Any]=13 , A_ : List[Any]=30 , A_ : Optional[Any]=2 , A_ : List[str]=3 , A_ : List[str]=True , A_ : Dict=True , A_ : List[Any]=32 , A_ : Any=2 , A_ : Any=4 , A_ : Optional[int]=37 , A_ : Dict="gelu" , A_ : List[Any]=0.1 , A_ : Optional[int]=0.1 , A_ : Union[str, Any]=10 , A_ : Optional[Any]=0.02 , A_ : List[Any]=3 , A_ : str=None , ) -> str: """simple docstring""" lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = image_size lowerCamelCase_ = patch_size lowerCamelCase_ = num_channels lowerCamelCase_ = is_training lowerCamelCase_ = use_labels lowerCamelCase_ = hidden_size lowerCamelCase_ = num_hidden_layers lowerCamelCase_ = num_attention_heads lowerCamelCase_ = intermediate_size lowerCamelCase_ = hidden_act lowerCamelCase_ = hidden_dropout_prob lowerCamelCase_ = attention_probs_dropout_prob lowerCamelCase_ = type_sequence_label_size lowerCamelCase_ = initializer_range lowerCamelCase_ = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) lowerCamelCase_ = (image_size // patch_size) ** 2 lowerCamelCase_ = num_patches + 1 def a__ ( self : List[str] ) -> Dict: """simple docstring""" lowerCamelCase_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase_ = None if self.use_labels: lowerCamelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase_ = self.get_config() return config, pixel_values, labels def a__ ( self : List[Any] ) -> Any: """simple docstring""" return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=A_ , initializer_range=self.initializer_range , ) def a__ ( self : Any , A_ : int , A_ : int , A_ : int ) -> List[Any]: """simple docstring""" lowerCamelCase_ = TFViTModel(config=A_ ) lowerCamelCase_ = model(A_ , training=A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # Test with an image with different size than the one specified in config. lowerCamelCase_ = self.image_size // 2 lowerCamelCase_ = pixel_values[:, :, :image_size, :image_size] lowerCamelCase_ = model(A_ , interpolate_pos_encoding=A_ , training=A_ ) lowerCamelCase_ = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, seq_length, self.hidden_size) ) def a__ ( self : List[Any] , A_ : List[Any] , A_ : Any , A_ : Any ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.type_sequence_label_size lowerCamelCase_ = TFViTForImageClassification(A_ ) lowerCamelCase_ = model(A_ , labels=A_ , training=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # Test with an image with different size than the one specified in config. lowerCamelCase_ = self.image_size // 2 lowerCamelCase_ = pixel_values[:, :, :image_size, :image_size] lowerCamelCase_ = model(A_ , interpolate_pos_encoding=A_ , training=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowerCamelCase_ = 1 lowerCamelCase_ = TFViTForImageClassification(A_ ) lowerCamelCase_ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCamelCase_ = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def a__ ( self : List[Any] ) -> Any: """simple docstring""" lowerCamelCase_ = self.prepare_config_and_inputs() lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = config_and_inputs lowerCamelCase_ = {'pixel_values': pixel_values} return config, inputs_dict @require_tf class A( UpperCamelCase , UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () UpperCamelCase = ( {'''feature-extraction''': TFViTModel, '''image-classification''': TFViTForImageClassification} if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def a__ ( self : int ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = TFViTModelTester(self ) lowerCamelCase_ = ConfigTester(self , config_class=A_ , has_text_modality=A_ , hidden_size=37 ) def a__ ( self : int ) -> List[str]: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='ViT does not use inputs_embeds' ) def a__ ( self : List[str] ) -> Tuple: """simple docstring""" pass @unittest.skip(reason='ViT does not use inputs_embeds' ) def a__ ( self : int ) -> Optional[Any]: """simple docstring""" pass def a__ ( self : str ) -> List[str]: """simple docstring""" lowerCamelCase_ , lowerCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase_ = model_class(A_ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) lowerCamelCase_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(A_ , tf.keras.layers.Layer ) ) def a__ ( self : List[Any] ) -> Dict: """simple docstring""" lowerCamelCase_ , lowerCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase_ = model_class(A_ ) lowerCamelCase_ = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase_ = [*signature.parameters.keys()] lowerCamelCase_ = ['pixel_values'] self.assertListEqual(arg_names[:1] , A_ ) def a__ ( self : Tuple ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def a__ ( self : Tuple ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A_ ) @slow def a__ ( self : Any ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = TFViTModel.from_pretrained('google/vit-base-patch16-224' ) self.assertIsNotNone(A_ ) def _SCREAMING_SNAKE_CASE ( ): '''simple docstring''' lowerCamelCase_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class A( unittest.TestCase ): '''simple docstring''' @cached_property def a__ ( self : int ) -> Tuple: """simple docstring""" return ViTImageProcessor.from_pretrained('google/vit-base-patch16-224' ) if is_vision_available() else None @slow def a__ ( self : List[str] ) -> str: """simple docstring""" lowerCamelCase_ = TFViTForImageClassification.from_pretrained('google/vit-base-patch16-224' ) lowerCamelCase_ = self.default_image_processor lowerCamelCase_ = prepare_img() lowerCamelCase_ = image_processor(images=A_ , return_tensors='tf' ) # forward pass lowerCamelCase_ = model(**A_ ) # verify the logits lowerCamelCase_ = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , A_ ) lowerCamelCase_ = tf.constant([-0.2744, 0.8215, -0.0836] ) tf.debugging.assert_near(outputs.logits[0, :3] , A_ , atol=1E-4 )
70
0
'''simple docstring''' import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetrImageProcessor class lowercase_ ( unittest.TestCase ): """simple docstring""" def __init__( self : Optional[Any] , __lowerCamelCase : Dict , __lowerCamelCase : int=7 , __lowerCamelCase : Any=3 , __lowerCamelCase : List[str]=3_0 , __lowerCamelCase : Union[str, Any]=4_0_0 , __lowerCamelCase : List[str]=True , __lowerCamelCase : int=None , __lowerCamelCase : Any=True , __lowerCamelCase : str=1 / 2_5_5 , __lowerCamelCase : int=True , __lowerCamelCase : List[Any]=[0.5, 0.5, 0.5] , __lowerCamelCase : Union[str, Any]=[0.5, 0.5, 0.5] , __lowerCamelCase : Union[str, Any]=True , ): """simple docstring""" _SCREAMING_SNAKE_CASE = size if size is not None else {"shortest_edge": 1_8, "longest_edge": 1_3_3_3} _SCREAMING_SNAKE_CASE = parent _SCREAMING_SNAKE_CASE = batch_size _SCREAMING_SNAKE_CASE = num_channels _SCREAMING_SNAKE_CASE = min_resolution _SCREAMING_SNAKE_CASE = max_resolution _SCREAMING_SNAKE_CASE = do_resize _SCREAMING_SNAKE_CASE = size _SCREAMING_SNAKE_CASE = do_rescale _SCREAMING_SNAKE_CASE = rescale_factor _SCREAMING_SNAKE_CASE = do_normalize _SCREAMING_SNAKE_CASE = image_mean _SCREAMING_SNAKE_CASE = image_std _SCREAMING_SNAKE_CASE = do_pad def lowerCAmelCase_ ( self : Tuple ): """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_pad": self.do_pad, } def lowerCAmelCase_ ( self : Union[str, Any] , __lowerCamelCase : Dict , __lowerCamelCase : Any=False ): """simple docstring""" if not batched: _SCREAMING_SNAKE_CASE = image_inputs[0] if isinstance(A_ , Image.Image ): _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE = image.size else: _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE = image.shape[1], image.shape[2] if w < h: _SCREAMING_SNAKE_CASE = int(self.size["shortest_edge"] * h / w ) _SCREAMING_SNAKE_CASE = self.size["shortest_edge"] elif w > h: _SCREAMING_SNAKE_CASE = self.size["shortest_edge"] _SCREAMING_SNAKE_CASE = int(self.size["shortest_edge"] * w / h ) else: _SCREAMING_SNAKE_CASE = self.size["shortest_edge"] _SCREAMING_SNAKE_CASE = self.size["shortest_edge"] else: _SCREAMING_SNAKE_CASE = [] for image in image_inputs: _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) _SCREAMING_SNAKE_CASE = max(A_ , key=lambda __lowerCamelCase : item[0] )[0] _SCREAMING_SNAKE_CASE = max(A_ , key=lambda __lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class lowercase_ ( A , unittest.TestCase ): """simple docstring""" lowerCamelCase_ = DetrImageProcessor if is_vision_available() else None def lowerCAmelCase_ ( self : List[Any] ): """simple docstring""" _SCREAMING_SNAKE_CASE = DetrImageProcessingTester(self ) @property def lowerCAmelCase_ ( self : List[str] ): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def lowerCAmelCase_ ( self : Dict ): """simple docstring""" _SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , "image_mean" ) ) self.assertTrue(hasattr(A_ , "image_std" ) ) self.assertTrue(hasattr(A_ , "do_normalize" ) ) self.assertTrue(hasattr(A_ , "do_rescale" ) ) self.assertTrue(hasattr(A_ , "rescale_factor" ) ) self.assertTrue(hasattr(A_ , "do_resize" ) ) self.assertTrue(hasattr(A_ , "size" ) ) self.assertTrue(hasattr(A_ , "do_pad" ) ) def lowerCAmelCase_ ( self : List[str] ): """simple docstring""" _SCREAMING_SNAKE_CASE = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 1_8, "longest_edge": 1_3_3_3} ) self.assertEqual(image_processor.do_pad , A_ ) _SCREAMING_SNAKE_CASE = self.image_processing_class.from_dict( self.image_processor_dict , size=4_2 , max_size=8_4 , pad_and_return_pixel_mask=A_ ) self.assertEqual(image_processor.size , {"shortest_edge": 4_2, "longest_edge": 8_4} ) self.assertEqual(image_processor.do_pad , A_ ) def lowerCAmelCase_ ( self : Dict ): """simple docstring""" pass def lowerCAmelCase_ ( self : Tuple ): """simple docstring""" _SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _SCREAMING_SNAKE_CASE = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ ) for image in image_inputs: self.assertIsInstance(A_ , Image.Image ) # Test not batched input _SCREAMING_SNAKE_CASE = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) _SCREAMING_SNAKE_CASE = image_processing(A_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def lowerCAmelCase_ ( self : List[str] ): """simple docstring""" _SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _SCREAMING_SNAKE_CASE = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , np.ndarray ) # Test not batched input _SCREAMING_SNAKE_CASE = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _SCREAMING_SNAKE_CASE = image_processing(A_ , return_tensors="pt" ).pixel_values _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def lowerCAmelCase_ ( self : List[Any] ): """simple docstring""" _SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _SCREAMING_SNAKE_CASE = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test not batched input _SCREAMING_SNAKE_CASE = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _SCREAMING_SNAKE_CASE = image_processing(A_ , return_tensors="pt" ).pixel_values _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def lowerCAmelCase_ ( self : Tuple ): """simple docstring""" _SCREAMING_SNAKE_CASE = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: _SCREAMING_SNAKE_CASE = json.loads(f.read() ) _SCREAMING_SNAKE_CASE = {"image_id": 3_9_7_6_9, "annotations": target} # encode them _SCREAMING_SNAKE_CASE = DetrImageProcessor.from_pretrained("facebook/detr-resnet-50" ) _SCREAMING_SNAKE_CASE = image_processing(images=A_ , annotations=A_ , return_tensors="pt" ) # verify pixel values _SCREAMING_SNAKE_CASE = torch.Size([1, 3, 8_0_0, 1_0_6_6] ) self.assertEqual(encoding["pixel_values"].shape , A_ ) _SCREAMING_SNAKE_CASE = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , A_ , atol=1e-4 ) ) # verify area _SCREAMING_SNAKE_CASE = torch.tensor([5_8_8_7.9_6_0_0, 1_1_2_5_0.2_0_6_1, 4_8_9_3_5_3.8_4_3_8, 8_3_7_1_2_2.7_5_0_0, 1_4_7_9_6_7.5_1_5_6, 1_6_5_7_3_2.3_4_3_8] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , A_ ) ) # verify boxes _SCREAMING_SNAKE_CASE = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , A_ ) _SCREAMING_SNAKE_CASE = torch.tensor([0.5_5_0_3, 0.2_7_6_5, 0.0_6_0_4, 0.2_2_1_5] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , A_ , atol=1e-3 ) ) # verify image_id _SCREAMING_SNAKE_CASE = torch.tensor([3_9_7_6_9] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , A_ ) ) # verify is_crowd _SCREAMING_SNAKE_CASE = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , A_ ) ) # verify class_labels _SCREAMING_SNAKE_CASE = torch.tensor([7_5, 7_5, 6_3, 6_5, 1_7, 1_7] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , A_ ) ) # verify orig_size _SCREAMING_SNAKE_CASE = torch.tensor([4_8_0, 6_4_0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , A_ ) ) # verify size _SCREAMING_SNAKE_CASE = torch.tensor([8_0_0, 1_0_6_6] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , A_ ) ) @slow def lowerCAmelCase_ ( self : str ): """simple docstring""" _SCREAMING_SNAKE_CASE = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: _SCREAMING_SNAKE_CASE = json.loads(f.read() ) _SCREAMING_SNAKE_CASE = {"file_name": "000000039769.png", "image_id": 3_9_7_6_9, "segments_info": target} _SCREAMING_SNAKE_CASE = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them _SCREAMING_SNAKE_CASE = DetrImageProcessor.from_pretrained("facebook/detr-resnet-50-panoptic" ) _SCREAMING_SNAKE_CASE = image_processing(images=A_ , annotations=A_ , masks_path=A_ , return_tensors="pt" ) # verify pixel values _SCREAMING_SNAKE_CASE = torch.Size([1, 3, 8_0_0, 1_0_6_6] ) self.assertEqual(encoding["pixel_values"].shape , A_ ) _SCREAMING_SNAKE_CASE = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , A_ , atol=1e-4 ) ) # verify area _SCREAMING_SNAKE_CASE = torch.tensor([1_4_7_9_7_9.6_8_7_5, 1_6_5_5_2_7.0_4_6_9, 4_8_4_6_3_8.5_9_3_8, 1_1_2_9_2.9_3_7_5, 5_8_7_9.6_5_6_2, 7_6_3_4.1_1_4_7] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , A_ ) ) # verify boxes _SCREAMING_SNAKE_CASE = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , A_ ) _SCREAMING_SNAKE_CASE = torch.tensor([0.2_6_2_5, 0.5_4_3_7, 0.4_6_8_8, 0.8_6_2_5] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , A_ , atol=1e-3 ) ) # verify image_id _SCREAMING_SNAKE_CASE = torch.tensor([3_9_7_6_9] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , A_ ) ) # verify is_crowd _SCREAMING_SNAKE_CASE = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , A_ ) ) # verify class_labels _SCREAMING_SNAKE_CASE = torch.tensor([1_7, 1_7, 6_3, 7_5, 7_5, 9_3] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , A_ ) ) # verify masks _SCREAMING_SNAKE_CASE = 8_2_2_8_7_3 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , A_ ) # verify orig_size _SCREAMING_SNAKE_CASE = torch.tensor([4_8_0, 6_4_0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , A_ ) ) # verify size _SCREAMING_SNAKE_CASE = torch.tensor([8_0_0, 1_0_6_6] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , A_ ) )
418
import itertools import random import unittest import numpy as np from transformers import is_speech_available from transformers.testing_utils import require_torch, require_torchaudio from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import SpeechaTextFeatureExtractor lowerCamelCase : Any = random.Random() def _SCREAMING_SNAKE_CASE ( lowercase : Union[str, Any] , lowercase : int=1.0 , lowercase : List[str]=None , lowercase : str=None ): '''simple docstring''' if rng is None: lowerCamelCase_ = global_rng lowerCamelCase_ = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class A( unittest.TestCase ): '''simple docstring''' def __init__( self : List[Any] , A_ : Dict , A_ : int=7 , A_ : str=400 , A_ : Dict=2000 , A_ : List[Any]=24 , A_ : List[Any]=24 , A_ : int=0.0 , A_ : Dict=16000 , A_ : List[Any]=True , A_ : str=True , ) -> Dict: """simple docstring""" lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = min_seq_length lowerCamelCase_ = max_seq_length lowerCamelCase_ = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) lowerCamelCase_ = feature_size lowerCamelCase_ = num_mel_bins lowerCamelCase_ = padding_value lowerCamelCase_ = sampling_rate lowerCamelCase_ = return_attention_mask lowerCamelCase_ = do_normalize def a__ ( self : List[str] ) -> Optional[int]: """simple docstring""" return { "feature_size": self.feature_size, "num_mel_bins": self.num_mel_bins, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def a__ ( self : List[Any] , A_ : str=False , A_ : Union[str, Any]=False ) -> str: """simple docstring""" def _flatten(A_ : List[Any] ): return list(itertools.chain(*A_ ) ) if equal_length: lowerCamelCase_ = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size lowerCamelCase_ = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: lowerCamelCase_ = [np.asarray(A_ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class A( UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = SpeechaTextFeatureExtractor if is_speech_available() else None def a__ ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = SpeechaTextFeatureExtractionTester(self ) def a__ ( self : str , A_ : Dict ) -> Dict: """simple docstring""" self.assertTrue(np.all(np.mean(A_ , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(A_ , axis=0 ) - 1 ) < 1E-3 ) ) def a__ ( self : int ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = [np.asarray(A_ ) for speech_input in speech_inputs] # Test feature size lowerCamelCase_ = feature_extractor(A_ , padding=A_ , return_tensors='np' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.feature_size ) # Test not batched input lowerCamelCase_ = feature_extractor(speech_inputs[0] , return_tensors='np' ).input_features lowerCamelCase_ = feature_extractor(np_speech_inputs[0] , return_tensors='np' ).input_features self.assertTrue(np.allclose(A_ , A_ , atol=1E-3 ) ) # Test batched lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. lowerCamelCase_ = [floats_list((1, x) )[0] for x in (800, 800, 800)] lowerCamelCase_ = np.asarray(A_ ) lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1E-3 ) ) def a__ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = ['longest', 'max_length', 'do_not_pad'] lowerCamelCase_ = [None, 16, None] for max_length, padding in zip(A_ , A_ ): lowerCamelCase_ = feature_extractor( A_ , padding=A_ , max_length=A_ , return_attention_mask=A_ ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = [np.sum(A_ ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def a__ ( self : List[Any] ) -> Any: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = ['longest', 'max_length', 'do_not_pad'] lowerCamelCase_ = [None, 16, None] for max_length, padding in zip(A_ , A_ ): lowerCamelCase_ = feature_extractor( A_ , max_length=A_ , padding=A_ , return_tensors='np' , return_attention_mask=A_ ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = [np.sum(A_ ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self.assertTrue(input_features[0][fbank_feat_lengths[0] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self.assertTrue(input_features[0][fbank_feat_lengths[1] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def a__ ( self : List[str] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = feature_extractor( A_ , padding='max_length' , max_length=4 , truncation=A_ , return_tensors='np' , return_attention_mask=A_ , ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1] ) self._check_zero_mean_unit_variance(input_features[2] ) def a__ ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = feature_extractor( A_ , padding='longest' , max_length=4 , truncation=A_ , return_tensors='np' , return_attention_mask=A_ , ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 4, 24) ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = feature_extractor( A_ , padding='longest' , max_length=16 , truncation=A_ , return_tensors='np' , return_attention_mask=A_ , ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 6, 24) ) def a__ ( self : List[Any] ) -> List[str]: """simple docstring""" import torch lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = np.random.rand(100 , 32 ).astype(np.floataa ) lowerCamelCase_ = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: lowerCamelCase_ = feature_extractor.pad([{'input_features': inputs}] , return_tensors='np' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) lowerCamelCase_ = feature_extractor.pad([{'input_features': inputs}] , return_tensors='pt' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def a__ ( self : List[str] , A_ : Union[str, Any] ) -> List[Any]: """simple docstring""" from datasets import load_dataset lowerCamelCase_ = load_dataset('hf-internal-testing/librispeech_asr_dummy' , 'clean' , split='validation' ) # automatic decoding with librispeech lowerCamelCase_ = ds.sort('id' ).select(range(A_ ) )[:num_samples]['audio'] return [x["array"] for x in speech_samples] def a__ ( self : str ) -> Tuple: """simple docstring""" lowerCamelCase_ = np.array([ -1.5745, -1.7713, -1.7020, -1.6069, -1.2250, -1.1105, -0.9072, -0.8241, -1.2310, -0.8098, -0.3320, -0.4101, -0.7985, -0.4996, -0.8213, -0.9128, -1.0420, -1.1286, -1.0440, -0.7999, -0.8405, -1.2275, -1.5443, -1.4625, ] ) # fmt: on lowerCamelCase_ = self._load_datasamples(1 ) lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = feature_extractor(A_ , return_tensors='pt' ).input_features self.assertEquals(input_features.shape , (1, 584, 24) ) self.assertTrue(np.allclose(input_features[0, 0, :30] , A_ , atol=1E-4 ) )
70
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available lowerCAmelCase_ = {"tokenization_herbert": ["HerbertTokenizer"]} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = ["HerbertTokenizerFast"] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys lowerCAmelCase_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
531
import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class A( UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = TransfoXLTokenizer UpperCamelCase = False UpperCamelCase = False def a__ ( self : Optional[Any] ) -> int: """simple docstring""" super().setUp() lowerCamelCase_ = [ '<unk>', '[CLS]', '[SEP]', 'want', 'unwanted', 'wa', 'un', 'running', ',', 'low', 'l', ] lowerCamelCase_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def a__ ( self : Optional[Any] , **A_ : Tuple ) -> Any: """simple docstring""" lowerCamelCase_ = True return TransfoXLTokenizer.from_pretrained(self.tmpdirname , **A_ ) def a__ ( self : List[str] , A_ : Dict ) -> Any: """simple docstring""" lowerCamelCase_ = '<unk> UNwanted , running' lowerCamelCase_ = '<unk> unwanted, running' return input_text, output_text def a__ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = TransfoXLTokenizer(vocab_file=self.vocab_file , lower_case=A_ ) lowerCamelCase_ = tokenizer.tokenize('<unk> UNwanted , running' ) self.assertListEqual(A_ , ['<unk>', 'unwanted', ',', 'running'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A_ ) , [0, 4, 8, 7] ) def a__ ( self : Any ) -> str: """simple docstring""" lowerCamelCase_ = TransfoXLTokenizer(lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['hello', '!', 'how', 'are', 'you', '?'] ) def a__ ( self : int ) -> Dict: """simple docstring""" lowerCamelCase_ = TransfoXLTokenizer(lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?'] ) def a__ ( self : List[Any] ) -> Tuple: """simple docstring""" lowerCamelCase_ = TransfoXLTokenizer(lower_case=A_ ) lowerCamelCase_ = 'Hello (bracket) and side-scrolled [and] Henry\'s $5,000 with 3.34 m. What\'s up!?' lowerCamelCase_ = [ 'Hello', '(', 'bracket', ')', 'and', 'side', '@-@', 'scrolled', '[', 'and', ']', 'Henry', '\'s', '$', '5', '@,@', '000', 'with', '3', '@.@', '34', 'm', '.', 'What', '\'s', 'up', '!', '?', ] self.assertListEqual(tokenizer.tokenize(A_ ) , A_ ) self.assertEqual(tokenizer.convert_tokens_to_string(A_ ) , A_ ) def a__ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = len(A_ ) tokenizer.add_tokens(['new1', 'new2'] ) tokenizer.move_added_token('new1' , 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(A_ ) , original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode('new1' ) , [1] ) self.assertEqual(tokenizer.decode([1] ) , 'new1' )
70
0
'''simple docstring''' from argparse import ArgumentParser from .add_new_model import AddNewModelCommand from .add_new_model_like import AddNewModelLikeCommand from .convert import ConvertCommand from .download import DownloadCommand from .env import EnvironmentCommand from .lfs import LfsCommands from .pt_to_tf import PTtoTFCommand from .run import RunCommand from .serving import ServeCommand from .user import UserCommands def _lowerCAmelCase ( ) -> Any: """simple docstring""" _SCREAMING_SNAKE_CASE =ArgumentParser('Transformers CLI tool' , usage='transformers-cli <command> [<args>]' ) _SCREAMING_SNAKE_CASE =parser.add_subparsers(help='transformers-cli command helpers' ) # Register commands ConvertCommand.register_subcommand(_UpperCamelCase ) DownloadCommand.register_subcommand(_UpperCamelCase ) EnvironmentCommand.register_subcommand(_UpperCamelCase ) RunCommand.register_subcommand(_UpperCamelCase ) ServeCommand.register_subcommand(_UpperCamelCase ) UserCommands.register_subcommand(_UpperCamelCase ) AddNewModelCommand.register_subcommand(_UpperCamelCase ) AddNewModelLikeCommand.register_subcommand(_UpperCamelCase ) LfsCommands.register_subcommand(_UpperCamelCase ) PTtoTFCommand.register_subcommand(_UpperCamelCase ) # Let's go _SCREAMING_SNAKE_CASE =parser.parse_args() if not hasattr(_UpperCamelCase , 'func' ): parser.print_help() exit(1 ) # Run _SCREAMING_SNAKE_CASE =args.func(_UpperCamelCase ) service.run() if __name__ == "__main__": main()
405
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetrImageProcessor class A( unittest.TestCase ): '''simple docstring''' def __init__( self : Optional[Any] , A_ : Dict , A_ : int=7 , A_ : Any=3 , A_ : List[str]=30 , A_ : Union[str, Any]=400 , A_ : List[str]=True , A_ : int=None , A_ : Any=True , A_ : str=1 / 255 , A_ : int=True , A_ : List[Any]=[0.5, 0.5, 0.5] , A_ : Union[str, Any]=[0.5, 0.5, 0.5] , A_ : Union[str, Any]=True , ) -> List[str]: """simple docstring""" lowerCamelCase_ = size if size is not None else {'shortest_edge': 18, 'longest_edge': 1333} lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = num_channels lowerCamelCase_ = min_resolution lowerCamelCase_ = max_resolution lowerCamelCase_ = do_resize lowerCamelCase_ = size lowerCamelCase_ = do_rescale lowerCamelCase_ = rescale_factor lowerCamelCase_ = do_normalize lowerCamelCase_ = image_mean lowerCamelCase_ = image_std lowerCamelCase_ = do_pad def a__ ( self : Tuple ) -> Dict: """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_pad": self.do_pad, } def a__ ( self : Union[str, Any] , A_ : Dict , A_ : Any=False ) -> Union[str, Any]: """simple docstring""" if not batched: lowerCamelCase_ = image_inputs[0] if isinstance(A_ , Image.Image ): lowerCamelCase_ , lowerCamelCase_ = image.size else: lowerCamelCase_ , lowerCamelCase_ = image.shape[1], image.shape[2] if w < h: lowerCamelCase_ = int(self.size['shortest_edge'] * h / w ) lowerCamelCase_ = self.size['shortest_edge'] elif w > h: lowerCamelCase_ = self.size['shortest_edge'] lowerCamelCase_ = int(self.size['shortest_edge'] * w / h ) else: lowerCamelCase_ = self.size['shortest_edge'] lowerCamelCase_ = self.size['shortest_edge'] else: lowerCamelCase_ = [] for image in image_inputs: lowerCamelCase_ , lowerCamelCase_ = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) lowerCamelCase_ = max(A_ , key=lambda A_ : item[0] )[0] lowerCamelCase_ = max(A_ , key=lambda A_ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class A( UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = DetrImageProcessor if is_vision_available() else None def a__ ( self : List[Any] ) -> str: """simple docstring""" lowerCamelCase_ = DetrImageProcessingTester(self ) @property def a__ ( self : List[str] ) -> List[Any]: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def a__ ( self : Dict ) -> int: """simple docstring""" lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , 'image_mean' ) ) self.assertTrue(hasattr(A_ , 'image_std' ) ) self.assertTrue(hasattr(A_ , 'do_normalize' ) ) self.assertTrue(hasattr(A_ , 'do_rescale' ) ) self.assertTrue(hasattr(A_ , 'rescale_factor' ) ) self.assertTrue(hasattr(A_ , 'do_resize' ) ) self.assertTrue(hasattr(A_ , 'size' ) ) self.assertTrue(hasattr(A_ , 'do_pad' ) ) def a__ ( self : List[str] ) -> str: """simple docstring""" lowerCamelCase_ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 18, 'longest_edge': 1333} ) self.assertEqual(image_processor.do_pad , A_ ) lowerCamelCase_ = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A_ ) self.assertEqual(image_processor.size , {'shortest_edge': 42, 'longest_edge': 84} ) self.assertEqual(image_processor.do_pad , A_ ) def a__ ( self : Dict ) -> Any: """simple docstring""" pass def a__ ( self : Tuple ) -> Tuple: """simple docstring""" lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCamelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ ) for image in image_inputs: self.assertIsInstance(A_ , Image.Image ) # Test not batched input lowerCamelCase_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) lowerCamelCase_ = image_processing(A_ , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def a__ ( self : List[str] ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCamelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , np.ndarray ) # Test not batched input lowerCamelCase_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase_ = image_processing(A_ , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def a__ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCamelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test not batched input lowerCamelCase_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase_ = image_processing(A_ , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def a__ ( self : Tuple ) -> List[Any]: """simple docstring""" lowerCamelCase_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_annotations.txt' , 'r' ) as f: lowerCamelCase_ = json.loads(f.read() ) lowerCamelCase_ = {'image_id': 39769, 'annotations': target} # encode them lowerCamelCase_ = DetrImageProcessor.from_pretrained('facebook/detr-resnet-50' ) lowerCamelCase_ = image_processing(images=A_ , annotations=A_ , return_tensors='pt' ) # verify pixel values lowerCamelCase_ = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , A_ ) lowerCamelCase_ = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area lowerCamelCase_ = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , A_ ) ) # verify boxes lowerCamelCase_ = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , A_ ) lowerCamelCase_ = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , A_ , atol=1E-3 ) ) # verify image_id lowerCamelCase_ = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , A_ ) ) # verify is_crowd lowerCamelCase_ = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , A_ ) ) # verify class_labels lowerCamelCase_ = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , A_ ) ) # verify orig_size lowerCamelCase_ = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , A_ ) ) # verify size lowerCamelCase_ = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , A_ ) ) @slow def a__ ( self : str ) -> Any: """simple docstring""" lowerCamelCase_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt' , 'r' ) as f: lowerCamelCase_ = json.loads(f.read() ) lowerCamelCase_ = {'file_name': '000000039769.png', 'image_id': 39769, 'segments_info': target} lowerCamelCase_ = pathlib.Path('./tests/fixtures/tests_samples/COCO/coco_panoptic' ) # encode them lowerCamelCase_ = DetrImageProcessor.from_pretrained('facebook/detr-resnet-50-panoptic' ) lowerCamelCase_ = image_processing(images=A_ , annotations=A_ , masks_path=A_ , return_tensors='pt' ) # verify pixel values lowerCamelCase_ = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , A_ ) lowerCamelCase_ = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area lowerCamelCase_ = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , A_ ) ) # verify boxes lowerCamelCase_ = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , A_ ) lowerCamelCase_ = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , A_ , atol=1E-3 ) ) # verify image_id lowerCamelCase_ = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , A_ ) ) # verify is_crowd lowerCamelCase_ = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , A_ ) ) # verify class_labels lowerCamelCase_ = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , A_ ) ) # verify masks lowerCamelCase_ = 822873 self.assertEqual(encoding['labels'][0]['masks'].sum().item() , A_ ) # verify orig_size lowerCamelCase_ = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , A_ ) ) # verify size lowerCamelCase_ = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , A_ ) )
70
0
'''simple docstring''' def snake_case ( snake_case : int , snake_case : int , snake_case : int ) -> Tuple: """simple docstring""" lowerCAmelCase = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) # formula for sum of series return total def snake_case ( ) -> List[Any]: """simple docstring""" print(sum_of_series(1 , 1 , 10 ) ) if __name__ == "__main__": import doctest doctest.testmod()
284
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase : Dict = logging.get_logger(__name__) lowerCamelCase : int = { "microsoft/swinv2-tiny-patch4-window8-256": ( "https://huggingface.co/microsoft/swinv2-tiny-patch4-window8-256/resolve/main/config.json" ), } class A( UpperCamelCase ): '''simple docstring''' UpperCamelCase = '''swinv2''' UpperCamelCase = { '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers''', } def __init__( self : Dict , A_ : List[Any]=224 , A_ : Optional[Any]=4 , A_ : int=3 , A_ : Dict=96 , A_ : Any=[2, 2, 6, 2] , A_ : Optional[Any]=[3, 6, 12, 24] , A_ : Tuple=7 , A_ : Tuple=4.0 , A_ : str=True , A_ : str=0.0 , A_ : Union[str, Any]=0.0 , A_ : Optional[Any]=0.1 , A_ : str="gelu" , A_ : int=False , A_ : str=0.02 , A_ : List[Any]=1E-5 , A_ : Any=32 , **A_ : Tuple , ) -> Any: """simple docstring""" super().__init__(**A_ ) lowerCamelCase_ = image_size lowerCamelCase_ = patch_size lowerCamelCase_ = num_channels lowerCamelCase_ = embed_dim lowerCamelCase_ = depths lowerCamelCase_ = len(A_ ) lowerCamelCase_ = num_heads lowerCamelCase_ = window_size lowerCamelCase_ = mlp_ratio lowerCamelCase_ = qkv_bias lowerCamelCase_ = hidden_dropout_prob lowerCamelCase_ = attention_probs_dropout_prob lowerCamelCase_ = drop_path_rate lowerCamelCase_ = hidden_act lowerCamelCase_ = use_absolute_embeddings lowerCamelCase_ = layer_norm_eps lowerCamelCase_ = initializer_range lowerCamelCase_ = encoder_stride # we set the hidden_size attribute in order to make Swinv2 work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowerCamelCase_ = int(embed_dim * 2 ** (len(A_ ) - 1) ) lowerCamelCase_ = (0, 0, 0, 0)
70
0
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_realm import RealmTokenizer __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} __lowerCAmelCase = { "vocab_file": { "google/realm-cc-news-pretrained-embedder": ( "https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/vocab.txt" ), "google/realm-cc-news-pretrained-encoder": ( "https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/vocab.txt" ), "google/realm-cc-news-pretrained-scorer": ( "https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/vocab.txt" ), "google/realm-cc-news-pretrained-openqa": ( "https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/vocab.txt" ), "google/realm-orqa-nq-openqa": "https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/vocab.txt", "google/realm-orqa-nq-reader": "https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/vocab.txt", "google/realm-orqa-wq-openqa": "https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/vocab.txt", "google/realm-orqa-wq-reader": "https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/vocab.txt", }, "tokenizer_file": { "google/realm-cc-news-pretrained-embedder": ( "https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/tokenizer.jsont" ), "google/realm-cc-news-pretrained-encoder": ( "https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/tokenizer.json" ), "google/realm-cc-news-pretrained-scorer": ( "https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/tokenizer.json" ), "google/realm-cc-news-pretrained-openqa": ( "https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/tokenizer.json" ), "google/realm-orqa-nq-openqa": ( "https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/tokenizer.json" ), "google/realm-orqa-nq-reader": ( "https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/tokenizer.json" ), "google/realm-orqa-wq-openqa": ( "https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/tokenizer.json" ), "google/realm-orqa-wq-reader": ( "https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/tokenizer.json" ), }, } __lowerCAmelCase = { "google/realm-cc-news-pretrained-embedder": 5_12, "google/realm-cc-news-pretrained-encoder": 5_12, "google/realm-cc-news-pretrained-scorer": 5_12, "google/realm-cc-news-pretrained-openqa": 5_12, "google/realm-orqa-nq-openqa": 5_12, "google/realm-orqa-nq-reader": 5_12, "google/realm-orqa-wq-openqa": 5_12, "google/realm-orqa-wq-reader": 5_12, } __lowerCAmelCase = { "google/realm-cc-news-pretrained-embedder": {"do_lower_case": True}, "google/realm-cc-news-pretrained-encoder": {"do_lower_case": True}, "google/realm-cc-news-pretrained-scorer": {"do_lower_case": True}, "google/realm-cc-news-pretrained-openqa": {"do_lower_case": True}, "google/realm-orqa-nq-openqa": {"do_lower_case": True}, "google/realm-orqa-nq-reader": {"do_lower_case": True}, "google/realm-orqa-wq-openqa": {"do_lower_case": True}, "google/realm-orqa-wq-reader": {"do_lower_case": True}, } class __SCREAMING_SNAKE_CASE (__A ): """simple docstring""" _a : str = VOCAB_FILES_NAMES _a : Tuple = PRETRAINED_VOCAB_FILES_MAP _a : Tuple = PRETRAINED_INIT_CONFIGURATION _a : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _a : Optional[int] = RealmTokenizer def __init__( self , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=True , UpperCamelCase__="[UNK]" , UpperCamelCase__="[SEP]" , UpperCamelCase__="[PAD]" , UpperCamelCase__="[CLS]" , UpperCamelCase__="[MASK]" , UpperCamelCase__=True , UpperCamelCase__=None , **UpperCamelCase__ , ): """simple docstring""" super().__init__( A_ , tokenizer_file=A_ , do_lower_case=A_ , unk_token=A_ , sep_token=A_ , pad_token=A_ , cls_token=A_ , mask_token=A_ , tokenize_chinese_chars=A_ , strip_accents=A_ , **A_ , ) a_ = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('lowercase' , A_ ) != do_lower_case or normalizer_state.get('strip_accents' , A_ ) != strip_accents or normalizer_state.get('handle_chinese_chars' , A_ ) != tokenize_chinese_chars ): a_ = getattr(A_ , normalizer_state.pop('type' ) ) a_ = do_lower_case a_ = strip_accents a_ = tokenize_chinese_chars a_ = normalizer_class(**A_ ) a_ = do_lower_case def _a ( self , UpperCamelCase__ , **UpperCamelCase__ ): """simple docstring""" a_ = PaddingStrategy.MAX_LENGTH a_ = text a_ = kwargs.pop('text_pair' , A_ ) a_ = kwargs.pop('return_tensors' , A_ ) a_ = { 'input_ids': [], 'attention_mask': [], 'token_type_ids': [], } for idx, candidate_text in enumerate(A_ ): if batch_text_pair is not None: a_ = batch_text_pair[idx] else: a_ = None a_ = super().__call__(A_ , A_ , return_tensors=A_ , **A_ ) a_ = encoded_candidates.get('input_ids' ) a_ = encoded_candidates.get('attention_mask' ) a_ = encoded_candidates.get('token_type_ids' ) if encoded_input_ids is not None: output_data["input_ids"].append(A_ ) if encoded_attention_mask is not None: output_data["attention_mask"].append(A_ ) if encoded_token_type_ids is not None: output_data["token_type_ids"].append(A_ ) a_ = {key: item for key, item in output_data.items() if len(A_ ) != 0} return BatchEncoding(A_ , tensor_type=A_ ) def _a ( self , UpperCamelCase__ , UpperCamelCase__=None ): """simple docstring""" a_ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _a ( self , UpperCamelCase__ , UpperCamelCase__ = None ): """simple docstring""" a_ = [self.sep_token_id] a_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _a ( self , UpperCamelCase__ , UpperCamelCase__ = None ): """simple docstring""" a_ = self._tokenizer.model.save(A_ , name=A_ ) return tuple(A_ )
536
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import ( AutoProcessor, BertTokenizerFast, BlipImageProcessor, GPTaTokenizer, InstructBlipProcessor, PreTrainedTokenizerFast, ) @require_vision class A( unittest.TestCase ): '''simple docstring''' def a__ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = tempfile.mkdtemp() lowerCamelCase_ = BlipImageProcessor() lowerCamelCase_ = GPTaTokenizer.from_pretrained('hf-internal-testing/tiny-random-GPT2Model' ) lowerCamelCase_ = BertTokenizerFast.from_pretrained('hf-internal-testing/tiny-random-bert' ) lowerCamelCase_ = InstructBlipProcessor(A_ , A_ , A_ ) processor.save_pretrained(self.tmpdirname ) def a__ ( self : Optional[int] , **A_ : Optional[int] ) -> Dict: """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **A_ ).tokenizer def a__ ( self : List[str] , **A_ : str ) -> Optional[Any]: """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **A_ ).image_processor def a__ ( self : Tuple , **A_ : Any ) -> Optional[int]: """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **A_ ).qformer_tokenizer def a__ ( self : str ) -> str: """simple docstring""" shutil.rmtree(self.tmpdirname ) def a__ ( self : Dict ) -> Dict: """simple docstring""" lowerCamelCase_ = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] lowerCamelCase_ = [Image.fromarray(np.moveaxis(A_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def a__ ( self : Tuple ) -> Tuple: """simple docstring""" lowerCamelCase_ = InstructBlipProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() , qformer_tokenizer=self.get_qformer_tokenizer() , ) processor.save_pretrained(self.tmpdirname ) lowerCamelCase_ = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) lowerCamelCase_ = self.get_image_processor(do_normalize=A_ , padding_value=1.0 ) lowerCamelCase_ = InstructBlipProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=A_ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , A_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , A_ ) self.assertIsInstance(processor.qformer_tokenizer , A_ ) def a__ ( self : List[str] ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = self.prepare_image_inputs() lowerCamelCase_ = image_processor(A_ , return_tensors='np' ) lowerCamelCase_ = processor(images=A_ , return_tensors='np' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def a__ ( self : Union[str, Any] ) -> Tuple: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = 'lower newer' lowerCamelCase_ = processor(text=A_ ) lowerCamelCase_ = tokenizer(A_ , return_token_type_ids=A_ ) lowerCamelCase_ = qformer_tokenizer(A_ , return_token_type_ids=A_ ) for key in encoded_tokens.keys(): self.assertListEqual(encoded_tokens[key] , encoded_processor[key] ) for key in encoded_tokens_qformer.keys(): self.assertListEqual(encoded_tokens_qformer[key] , encoded_processor['qformer_' + key] ) def a__ ( self : Tuple ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = 'lower newer' lowerCamelCase_ = self.prepare_image_inputs() lowerCamelCase_ = processor(text=A_ , images=A_ ) self.assertListEqual( list(inputs.keys() ) , ['input_ids', 'attention_mask', 'qformer_input_ids', 'qformer_attention_mask', 'pixel_values'] , ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def a__ ( self : Optional[int] ) -> int: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowerCamelCase_ = processor.batch_decode(A_ ) lowerCamelCase_ = tokenizer.batch_decode(A_ ) self.assertListEqual(A_ , A_ ) def a__ ( self : Tuple ) -> List[str]: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = 'lower newer' lowerCamelCase_ = self.prepare_image_inputs() lowerCamelCase_ = processor(text=A_ , images=A_ ) self.assertListEqual( list(inputs.keys() ) , ['input_ids', 'attention_mask', 'qformer_input_ids', 'qformer_attention_mask', 'pixel_values'] , )
70
0
import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.local_sgd import LocalSGD ######################################################################## # This is a fully working simple example to use Accelerate # with LocalSGD, which is a method to synchronize model # parameters every K batches. It is different, but complementary # to gradient accumulation. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## SCREAMING_SNAKE_CASE__ = 1_6 SCREAMING_SNAKE_CASE__ = 3_2 def A ( __UpperCamelCase , __UpperCamelCase = 16 ) -> Dict: A__ = AutoTokenizer.from_pretrained('bert-base-cased' ) A__ = load_dataset('glue' , 'mrpc' ) def tokenize_function(__UpperCamelCase ): # max_length=None => use the model max length (it's actually the default) A__ = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=__UpperCamelCase , max_length=__UpperCamelCase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): A__ = datasets.map( __UpperCamelCase , batched=__UpperCamelCase , remove_columns=['idx', 'sentence1', 'sentence2'] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library A__ = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(__UpperCamelCase ): # On TPU it's best to pad everything to the same length or training will be very slow. A__ = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": A__ = 16 elif accelerator.mixed_precision != "no": A__ = 8 else: A__ = None return tokenizer.pad( __UpperCamelCase , padding='longest' , max_length=__UpperCamelCase , pad_to_multiple_of=__UpperCamelCase , return_tensors='pt' , ) # Instantiate dataloaders. A__ = DataLoader( tokenized_datasets['train'] , shuffle=__UpperCamelCase , collate_fn=__UpperCamelCase , batch_size=__UpperCamelCase ) A__ = DataLoader( tokenized_datasets['validation'] , shuffle=__UpperCamelCase , collate_fn=__UpperCamelCase , batch_size=__UpperCamelCase ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('''TESTING_MOCKED_DATALOADERS''', None) == "1": from accelerate.test_utils.training import mocked_dataloaders SCREAMING_SNAKE_CASE__ = mocked_dataloaders # noqa: F811 def A ( __UpperCamelCase , __UpperCamelCase ) -> Dict: if os.environ.get('TESTING_MOCKED_DATALOADERS' , __UpperCamelCase ) == "1": A__ = 2 # New Code # A__ = int(args.gradient_accumulation_steps ) A__ = int(args.local_sgd_steps ) # Initialize accelerator A__ = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , gradient_accumulation_steps=__UpperCamelCase ) if accelerator.distributed_type not in [DistributedType.NO, DistributedType.MULTI_CPU, DistributedType.MULTI_GPU]: raise NotImplementedError('LocalSGD is supported only for CPUs and GPUs (no DeepSpeed or MegatronLM)' ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs A__ = config['lr'] A__ = int(config['num_epochs'] ) A__ = int(config['seed'] ) A__ = int(config['batch_size'] ) A__ = evaluate.load('glue' , 'mrpc' ) set_seed(__UpperCamelCase ) A__ , A__ = get_dataloaders(__UpperCamelCase , __UpperCamelCase ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) A__ = AutoModelForSequenceClassification.from_pretrained('bert-base-cased' , return_dict=__UpperCamelCase ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). A__ = model.to(accelerator.device ) # Instantiate optimizer A__ = AdamW(params=model.parameters() , lr=__UpperCamelCase ) # Instantiate scheduler A__ = get_linear_schedule_with_warmup( optimizer=__UpperCamelCase , num_warmup_steps=100 , num_training_steps=(len(__UpperCamelCase ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. A__ , A__ , A__ , A__ , A__ = accelerator.prepare( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # Now we train the model for epoch in range(__UpperCamelCase ): model.train() with LocalSGD( accelerator=__UpperCamelCase , model=__UpperCamelCase , local_sgd_steps=__UpperCamelCase , enabled=local_sgd_steps is not None ) as local_sgd: for step, batch in enumerate(__UpperCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) # New code # # We use the new `accumulate` context manager to perform gradient accumulation # We also currently do not support TPUs nor advise it as bugs were found on the XLA side when running our tests. with accelerator.accumulate(__UpperCamelCase ): A__ = model(**__UpperCamelCase ) A__ = output.loss accelerator.backward(__UpperCamelCase ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() # LocalSGD-specific line local_sgd.step() model.eval() for step, batch in enumerate(__UpperCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): A__ = model(**__UpperCamelCase ) A__ = outputs.logits.argmax(dim=-1 ) A__ , A__ = accelerator.gather_for_metrics((predictions, batch['labels']) ) metric.add_batch( predictions=__UpperCamelCase , references=__UpperCamelCase , ) A__ = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f'''epoch {epoch}:''' , __UpperCamelCase ) def A ( ) -> str: A__ = argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument( '--mixed_precision' , type=__UpperCamelCase , default=__UpperCamelCase , choices=['no', 'fp16', 'bf16', 'fp8'] , help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' , ) # New Code # parser.add_argument( '--gradient_accumulation_steps' , type=__UpperCamelCase , default=1 , help='The number of minibatches to be ran before gradients are accumulated.' , ) parser.add_argument( '--local_sgd_steps' , type=__UpperCamelCase , default=8 , help='Number of local SGD steps or None to disable local SGD' ) parser.add_argument('--cpu' , action='store_true' , help='If passed, will train on the CPU.' ) A__ = parser.parse_args() A__ = {'lr': 2E-5, 'num_epochs': 3, 'seed': 42, 'batch_size': 16} training_function(__UpperCamelCase , __UpperCamelCase ) if __name__ == "__main__": main()
9
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_OBJECT_DETECTION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING lowerCamelCase : Tuple = logging.get_logger(__name__) lowerCamelCase : List[Any] = Dict[str, Any] lowerCamelCase : Dict = List[Prediction] @add_end_docstrings(UpperCamelCase ) class A( UpperCamelCase ): '''simple docstring''' def __init__( self : Tuple , *A_ : int , **A_ : int ) -> Optional[int]: """simple docstring""" super().__init__(*A_ , **A_ ) if self.framework == "tf": raise ValueError(f"""The {self.__class__} is only available in PyTorch.""" ) requires_backends(self , 'vision' ) self.check_model_type( dict(MODEL_FOR_OBJECT_DETECTION_MAPPING.items() + MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING.items() ) ) def a__ ( self : Union[str, Any] , **A_ : Union[str, Any] ) -> List[str]: """simple docstring""" lowerCamelCase_ = {} if "threshold" in kwargs: lowerCamelCase_ = kwargs['threshold'] return {}, {}, postprocess_kwargs def __call__( self : str , *A_ : Optional[int] , **A_ : Tuple ) -> Union[Predictions, List[Prediction]]: """simple docstring""" return super().__call__(*A_ , **A_ ) def a__ ( self : Union[str, Any] , A_ : Tuple ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = load_image(A_ ) lowerCamelCase_ = torch.IntTensor([[image.height, image.width]] ) lowerCamelCase_ = self.image_processor(images=[image] , return_tensors='pt' ) if self.tokenizer is not None: lowerCamelCase_ = self.tokenizer(text=inputs['words'] , boxes=inputs['boxes'] , return_tensors='pt' ) lowerCamelCase_ = target_size return inputs def a__ ( self : Union[str, Any] , A_ : List[str] ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = model_inputs.pop('target_size' ) lowerCamelCase_ = self.model(**A_ ) lowerCamelCase_ = outputs.__class__({'target_size': target_size, **outputs} ) if self.tokenizer is not None: lowerCamelCase_ = model_inputs['bbox'] return model_outputs def a__ ( self : str , A_ : Any , A_ : Tuple=0.9 ) -> str: """simple docstring""" lowerCamelCase_ = model_outputs['target_size'] if self.tokenizer is not None: # This is a LayoutLMForTokenClassification variant. # The OCR got the boxes and the model classified the words. lowerCamelCase_ , lowerCamelCase_ = target_size[0].tolist() def unnormalize(A_ : Dict ): return self._get_bounding_box( torch.Tensor( [ (width * bbox[0] / 1000), (height * bbox[1] / 1000), (width * bbox[2] / 1000), (height * bbox[3] / 1000), ] ) ) lowerCamelCase_ , lowerCamelCase_ = model_outputs['logits'].squeeze(0 ).softmax(dim=-1 ).max(dim=-1 ) lowerCamelCase_ = [self.model.config.idalabel[prediction] for prediction in classes.tolist()] lowerCamelCase_ = [unnormalize(A_ ) for bbox in model_outputs['bbox'].squeeze(0 )] lowerCamelCase_ = ['score', 'label', 'box'] lowerCamelCase_ = [dict(zip(A_ , A_ ) ) for vals in zip(scores.tolist() , A_ , A_ ) if vals[0] > threshold] else: # This is a regular ForObjectDetectionModel lowerCamelCase_ = self.image_processor.post_process_object_detection(A_ , A_ , A_ ) lowerCamelCase_ = raw_annotations[0] lowerCamelCase_ = raw_annotation['scores'] lowerCamelCase_ = raw_annotation['labels'] lowerCamelCase_ = raw_annotation['boxes'] lowerCamelCase_ = scores.tolist() lowerCamelCase_ = [self.model.config.idalabel[label.item()] for label in labels] lowerCamelCase_ = [self._get_bounding_box(A_ ) for box in boxes] # {"scores": [...], ...} --> [{"score":x, ...}, ...] lowerCamelCase_ = ['score', 'label', 'box'] lowerCamelCase_ = [ dict(zip(A_ , A_ ) ) for vals in zip(raw_annotation['scores'] , raw_annotation['labels'] , raw_annotation['boxes'] ) ] return annotation def a__ ( self : Union[str, Any] , A_ : "torch.Tensor" ) -> Dict[str, int]: """simple docstring""" if self.framework != "pt": raise ValueError('The ObjectDetectionPipeline is only available in PyTorch.' ) lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = box.int().tolist() lowerCamelCase_ = { 'xmin': xmin, 'ymin': ymin, 'xmax': xmax, 'ymax': ymax, } return bbox
70
0
'''simple docstring''' import math import qiskit def a ( _UpperCAmelCase = 1 , _UpperCAmelCase = 1 , _UpperCAmelCase = 1 ) -> Dict: """simple docstring""" if ( isinstance(_UpperCAmelCase , _UpperCAmelCase ) or isinstance(_UpperCAmelCase , _UpperCAmelCase ) or isinstance(_UpperCAmelCase , _UpperCAmelCase ) ): raise TypeError('inputs must be integers.' ) if (input_a < 0) or (input_a < 0) or (carry_in < 0): raise ValueError('inputs must be positive.' ) if ( (math.floor(_UpperCAmelCase ) != input_a) or (math.floor(_UpperCAmelCase ) != input_a) or (math.floor(_UpperCAmelCase ) != carry_in) ): raise ValueError('inputs must be exact integers.' ) if (input_a > 2) or (input_a > 2) or (carry_in > 2): raise ValueError('inputs must be less or equal to 2.' ) # build registers a_ = qiskit.QuantumRegister(4 , 'qr' ) a_ = qiskit.ClassicalRegister(2 , 'cr' ) # list the entries a_ = [input_a, input_a, carry_in] a_ = qiskit.QuantumCircuit(_UpperCAmelCase , _UpperCAmelCase ) for i in range(0 , 3 ): if entry[i] == 2: quantum_circuit.h(_UpperCAmelCase ) # for hadamard entries elif entry[i] == 1: quantum_circuit.x(_UpperCAmelCase ) # for 1 entries elif entry[i] == 0: quantum_circuit.i(_UpperCAmelCase ) # for 0 entries # build the circuit quantum_circuit.ccx(0 , 1 , 3 ) # ccx = toffoli gate quantum_circuit.cx(0 , 1 ) quantum_circuit.ccx(1 , 2 , 3 ) quantum_circuit.cx(1 , 2 ) quantum_circuit.cx(0 , 1 ) quantum_circuit.measure([2, 3] , _UpperCAmelCase ) # measure the last two qbits a_ = qiskit.Aer.get_backend('aer_simulator' ) a_ = qiskit.execute(_UpperCAmelCase , _UpperCAmelCase , shots=1_0_0_0 ) return job.result().get_counts(_UpperCAmelCase ) if __name__ == "__main__": print(f'''Total sum count for state is: {quantum_full_adder(1, 1, 1)}''')
697
from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
70
0
from __future__ import annotations from typing import TypedDict class __A( a ): snake_case_ = 4_2 snake_case_ = 4_2 def __lowerCAmelCase ( a__ ) -> Optional[Any]: if not isinstance(a__ , a__ ): raise TypeError('''The parameter s type must be str.''' ) return [s[i:] + s[:i] for i in range(len(a__ ) )] def __lowerCAmelCase ( a__ ) -> Optional[Any]: if not isinstance(a__ , a__ ): raise TypeError('''The parameter s type must be str.''' ) if not s: raise ValueError('''The parameter s must not be empty.''' ) __a = all_rotations(a__ ) rotations.sort() # sort the list of rotations in alphabetically order # make a string composed of the last char of each rotation __a = { '''bwt_string''': ''''''.join([word[-1] for word in rotations] ), '''idx_original_string''': rotations.index(a__ ), } return response def __lowerCAmelCase ( a__ , a__ ) -> Union[str, Any]: if not isinstance(a__ , a__ ): raise TypeError('''The parameter bwt_string type must be str.''' ) if not bwt_string: raise ValueError('''The parameter bwt_string must not be empty.''' ) try: __a = int(a__ ) except ValueError: raise TypeError( '''The parameter idx_original_string type must be int or passive''' ''' of cast to int.''' ) if idx_original_string < 0: raise ValueError('''The parameter idx_original_string must not be lower than 0.''' ) if idx_original_string >= len(a__ ): raise ValueError( '''The parameter idx_original_string must be lower than''' ''' len(bwt_string).''' ) __a = [''''''] * len(a__ ) for _ in range(len(a__ ) ): for i in range(len(a__ ) ): __a = bwt_string[i] + ordered_rotations[i] ordered_rotations.sort() return ordered_rotations[idx_original_string] if __name__ == "__main__": A : Any = "Provide a string that I will generate its BWT transform: " A : Tuple = input(entry_msg).strip() A : Dict = bwt_transform(s) print( F"Burrows Wheeler transform for string '{s}' results " F"in '{result['bwt_string']}'" ) A : Dict = reverse_bwt(result['bwt_string'], result['idx_original_string']) print( F"Reversing Burrows Wheeler transform for entry '{result['bwt_string']}' " F"we get original string '{original_string}'" )
219
from collections import Counter from timeit import timeit def _SCREAMING_SNAKE_CASE ( lowercase : str = "" , ): '''simple docstring''' return sum(c % 2 for c in Counter(input_str.replace(' ' , '' ).lower() ).values() ) < 2 def _SCREAMING_SNAKE_CASE ( lowercase : str = "" ): '''simple docstring''' if len(lowercase ) == 0: return True lowerCamelCase_ = input_str.replace(' ' , '' ).lower() # character_freq_dict: Stores the frequency of every character in the input string lowerCamelCase_ = {} for character in lower_case_input_str: lowerCamelCase_ = character_freq_dict.get(lowercase , 0 ) + 1 lowerCamelCase_ = 0 for character_count in character_freq_dict.values(): if character_count % 2: odd_char += 1 if odd_char > 1: return False return True def _SCREAMING_SNAKE_CASE ( lowercase : str = "" ): '''simple docstring''' print('\nFor string = ' , lowercase , ':' ) print( '> can_string_be_rearranged_as_palindrome_counter()' , '\tans =' , can_string_be_rearranged_as_palindrome_counter(lowercase ) , '\ttime =' , timeit( 'z.can_string_be_rearranged_as_palindrome_counter(z.check_str)' , setup='import __main__ as z' , ) , 'seconds' , ) print( '> can_string_be_rearranged_as_palindrome()' , '\tans =' , can_string_be_rearranged_as_palindrome(lowercase ) , '\ttime =' , timeit( 'z.can_string_be_rearranged_as_palindrome(z.check_str)' , setup='import __main__ as z' , ) , 'seconds' , ) if __name__ == "__main__": lowerCamelCase : Optional[Any] = input( "Enter string to determine if it can be rearranged as a palindrome or not: " ).strip() benchmark(check_str) lowerCamelCase : int = can_string_be_rearranged_as_palindrome_counter(check_str) print(F"""{check_str} can {'' if status else 'not '}be rearranged as a palindrome""")
70
0
'''simple docstring''' from typing import Callable, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ = { "microsoft/xprophetnet-large-wiki100-cased": ( "https://huggingface.co/microsoft/xprophetnet-large-wiki100-cased/resolve/main/config.json" ), } class a ( __lowerCAmelCase ): """simple docstring""" __lowerCAmelCase = """xlm-prophetnet""" __lowerCAmelCase = ["""past_key_values"""] __lowerCAmelCase = { """num_attention_heads""": """num_encoder_attention_heads""", } def __init__( self , snake_case_ = 0.1 , snake_case_ = "gelu" , snake_case_ = 3_0522 , snake_case_ = 1024 , snake_case_ = 4096 , snake_case_ = 12 , snake_case_ = 16 , snake_case_ = 4096 , snake_case_ = 12 , snake_case_ = 16 , snake_case_ = 0.1 , snake_case_ = 0.1 , snake_case_ = 512 , snake_case_ = 0.0_2 , snake_case_ = True , snake_case_ = True , snake_case_ = 0 , snake_case_ = 2 , snake_case_ = 32 , snake_case_ = 128 , snake_case_ = False , snake_case_ = 0.0 , snake_case_ = True , snake_case_ = 0 , snake_case_ = 1 , snake_case_ = 2 , **snake_case_ , ): '''simple docstring''' __UpperCAmelCase: Optional[Any] = vocab_size __UpperCAmelCase: Any = hidden_size __UpperCAmelCase: List[Any] = encoder_ffn_dim __UpperCAmelCase: Any = num_encoder_layers __UpperCAmelCase: List[str] = num_encoder_attention_heads __UpperCAmelCase: Union[str, Any] = decoder_ffn_dim __UpperCAmelCase: Dict = num_decoder_layers __UpperCAmelCase: List[Any] = num_decoder_attention_heads __UpperCAmelCase: Union[str, Any] = max_position_embeddings __UpperCAmelCase: Tuple = init_std # Normal(0, this parameter) __UpperCAmelCase: int = activation_function # parameters for xlmprophetnet __UpperCAmelCase: List[Any] = ngram __UpperCAmelCase: List[Any] = num_buckets __UpperCAmelCase: Union[str, Any] = relative_max_distance __UpperCAmelCase: Optional[int] = disable_ngram_loss __UpperCAmelCase: int = eps # 3 Types of Dropout __UpperCAmelCase: str = attention_dropout __UpperCAmelCase: str = activation_dropout __UpperCAmelCase: str = dropout __UpperCAmelCase: str = use_cache super().__init__( pad_token_id=A_ , bos_token_id=A_ , eos_token_id=A_ , is_encoder_decoder=A_ , add_cross_attention=A_ , decoder_start_token_id=A_ , **A_ , ) @property def lowercase_ ( self ): '''simple docstring''' return self.num_encoder_layers + self.num_decoder_layers @num_hidden_layers.setter def lowercase_ ( self , snake_case_ ): '''simple docstring''' raise NotImplementedError( """This model does not support the setting of `num_hidden_layers`. Please set `num_encoder_layers` and""" """ `num_decoder_layers`.""" )
523
from __future__ import annotations from decimal import Decimal from math import * # noqa: F403 from sympy import diff def _SCREAMING_SNAKE_CASE ( lowercase : str , lowercase : float | Decimal , lowercase : float = 10**-10 ): '''simple docstring''' lowerCamelCase_ = a while True: lowerCamelCase_ = Decimal(lowercase ) - ( Decimal(eval(lowercase ) ) / Decimal(eval(str(diff(lowercase ) ) ) ) # noqa: S307 ) # This number dictates the accuracy of the answer if abs(eval(lowercase ) ) < precision: # noqa: S307 return float(lowercase ) # Let's Execute if __name__ == "__main__": # Find root of trigonometric function # Find value of pi print(F"""The root of sin(x) = 0 is {newton_raphson('sin(x)', 2)}""") # Find root of polynomial print(F"""The root of x**2 - 5*x + 2 = 0 is {newton_raphson('x**2 - 5*x + 2', 0.4)}""") # Find Square Root of 5 print(F"""The root of log(x) - 1 = 0 is {newton_raphson('log(x) - 1', 2)}""") # Exponential Roots print(F"""The root of exp(x) - 1 = 0 is {newton_raphson('exp(x) - 1', 0)}""")
70
0
import json import os import unittest from transformers import AutoTokenizer, GPTaTokenizer, GPTaTokenizerFast from transformers.models.gpta.tokenization_gpta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _SCREAMING_SNAKE_CASE ( snake_case, unittest.TestCase ): lowerCamelCase_ = GPTaTokenizer lowerCamelCase_ = GPTaTokenizerFast lowerCamelCase_ = True lowerCamelCase_ = {'add_prefix_space': True} lowerCamelCase_ = False def _UpperCAmelCase ( self : int ): """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt A : List[Any] = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', '''<|endoftext|>''', ] A : Any = dict(zip(A_ , range(len(A_ ) ) ) ) A : Tuple = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] A : Dict = {'''unk_token''': '''<unk>'''} A : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) A : List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A_ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(A_ ) ) def _UpperCAmelCase ( self : Union[str, Any] , **snake_case_ : Tuple ): """simple docstring""" kwargs.update(self.special_tokens_map ) return GPTaTokenizer.from_pretrained(self.tmpdirname , **A_ ) def _UpperCAmelCase ( self : int , **snake_case_ : List[Any] ): """simple docstring""" kwargs.update(self.special_tokens_map ) return GPTaTokenizerFast.from_pretrained(self.tmpdirname , **A_ ) def _UpperCAmelCase ( self : Union[str, Any] , snake_case_ : Dict ): """simple docstring""" A : Dict = '''lower newer''' A : Optional[Any] = '''lower newer''' return input_text, output_text def _UpperCAmelCase ( self : Dict ): """simple docstring""" A : int = GPTaTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) A : Optional[int] = '''lower newer''' A : Dict = ['''\u0120low''', '''er''', '''\u0120''', '''n''', '''e''', '''w''', '''er'''] A : Dict = tokenizer.tokenize(A_ , add_prefix_space=A_ ) self.assertListEqual(A_ , A_ ) A : Tuple = tokens + [tokenizer.unk_token] A : List[Any] = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(A_ ) , A_ ) def _UpperCAmelCase ( self : Dict ): """simple docstring""" if not self.test_rust_tokenizer: return A : List[str] = self.get_tokenizer() A : int = self.get_rust_tokenizer(add_prefix_space=A_ ) A : Optional[int] = '''lower newer''' # Testing tokenization A : Tuple = tokenizer.tokenize(A_ , add_prefix_space=A_ ) A : Dict = rust_tokenizer.tokenize(A_ ) self.assertListEqual(A_ , A_ ) # Testing conversion to ids without special tokens A : str = tokenizer.encode(A_ , add_special_tokens=A_ , add_prefix_space=A_ ) A : Any = rust_tokenizer.encode(A_ , add_special_tokens=A_ ) self.assertListEqual(A_ , A_ ) # Testing conversion to ids with special tokens A : Union[str, Any] = self.get_rust_tokenizer(add_prefix_space=A_ ) A : Optional[Any] = tokenizer.encode(A_ , add_prefix_space=A_ ) A : Optional[Any] = rust_tokenizer.encode(A_ ) self.assertListEqual(A_ , A_ ) # Testing the unknown token A : List[Any] = tokens + [rust_tokenizer.unk_token] A : Optional[Any] = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(A_ ) , A_ ) def _UpperCAmelCase ( self : int , *snake_case_ : List[Any] , **snake_case_ : List[Any] ): """simple docstring""" pass def _UpperCAmelCase ( self : Dict , snake_case_ : Tuple=15 ): """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): A : Any = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) # Simple input A : List[str] = '''This is a simple input''' A : str = ['''This is a simple input 1''', '''This is a simple input 2'''] A : List[str] = ('''This is a simple input''', '''This is a pair''') A : List[str] = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(A_ , tokenizer_r.encode , A_ , max_length=A_ , padding='''max_length''' ) # Simple input self.assertRaises(A_ , tokenizer_r.encode_plus , A_ , max_length=A_ , padding='''max_length''' ) # Simple input self.assertRaises( A_ , tokenizer_r.batch_encode_plus , A_ , max_length=A_ , padding='''max_length''' , ) # Pair input self.assertRaises(A_ , tokenizer_r.encode , A_ , max_length=A_ , padding='''max_length''' ) # Pair input self.assertRaises(A_ , tokenizer_r.encode_plus , A_ , max_length=A_ , padding='''max_length''' ) # Pair input self.assertRaises( A_ , tokenizer_r.batch_encode_plus , A_ , max_length=A_ , padding='''max_length''' , ) def _UpperCAmelCase ( self : Optional[Any] ): """simple docstring""" A : str = GPTaTokenizer.from_pretrained(self.tmpdirname , pad_token='''<pad>''' ) # Simple input A : str = '''This is a simple input''' A : Optional[Any] = ['''This is a simple input looooooooong''', '''This is a simple input'''] A : List[Any] = ('''This is a simple input''', '''This is a pair''') A : Optional[Any] = [ ('''This is a simple input loooooong''', '''This is a simple input'''), ('''This is a simple pair loooooong''', '''This is a simple pair'''), ] A : List[str] = tokenizer.pad_token_id A : Optional[Any] = tokenizer(A_ , padding='''max_length''' , max_length=30 , return_tensors='''np''' ) A : Tuple = tokenizer(A_ , padding=A_ , truncate=A_ , return_tensors='''np''' ) A : Any = tokenizer(*A_ , padding='''max_length''' , max_length=60 , return_tensors='''np''' ) A : Union[str, Any] = tokenizer(A_ , padding=A_ , truncate=A_ , return_tensors='''np''' ) # s # test single string max_length padding self.assertEqual(out_s['''input_ids'''].shape[-1] , 30 ) self.assertTrue(pad_token_id in out_s['''input_ids'''] ) self.assertTrue(0 in out_s['''attention_mask'''] ) # s2 # test automatic padding self.assertEqual(out_sa['''input_ids'''].shape[-1] , 33 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa['''input_ids'''][0] ) self.assertFalse(0 in out_sa['''attention_mask'''][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa['''input_ids'''][1] ) self.assertTrue(0 in out_sa['''attention_mask'''][1] ) # p # test single pair max_length padding self.assertEqual(out_p['''input_ids'''].shape[-1] , 60 ) self.assertTrue(pad_token_id in out_p['''input_ids'''] ) self.assertTrue(0 in out_p['''attention_mask'''] ) # p2 # test automatic padding pair self.assertEqual(out_pa['''input_ids'''].shape[-1] , 52 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa['''input_ids'''][0] ) self.assertFalse(0 in out_pa['''attention_mask'''][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa['''input_ids'''][1] ) self.assertTrue(0 in out_pa['''attention_mask'''][1] ) def _UpperCAmelCase ( self : Optional[Any] ): """simple docstring""" A : List[str] = '''$$$''' A : int = GPTaTokenizer.from_pretrained(self.tmpdirname , bos_token=A_ , add_bos_token=A_ ) A : Dict = '''This is a simple input''' A : Union[str, Any] = ['''This is a simple input 1''', '''This is a simple input 2'''] A : int = tokenizer.bos_token_id A : str = tokenizer(A_ ) A : int = tokenizer(A_ ) self.assertEqual(out_s.input_ids[0] , A_ ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) A : str = tokenizer.decode(out_s.input_ids ) A : Optional[Any] = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , A_ ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) def _UpperCAmelCase ( self : Union[str, Any] ): """simple docstring""" pass def _UpperCAmelCase ( self : Optional[int] ): """simple docstring""" A : int = [self.get_tokenizer(do_lower_case=A_ , add_bos_token=A_ )] for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): A : Optional[Any] = '''Encode this.''' A : Optional[Any] = '''This one too please.''' A : str = tokenizer.encode(A_ , add_special_tokens=A_ ) encoded_sequence += tokenizer.encode(A_ , add_special_tokens=A_ ) A : int = tokenizer.encode_plus( A_ , A_ , add_special_tokens=A_ , return_special_tokens_mask=A_ , ) A : Optional[int] = encoded_sequence_dict['''input_ids'''] A : Dict = encoded_sequence_dict['''special_tokens_mask'''] self.assertEqual(len(A_ ) , len(A_ ) ) A : int = [ (x if not special_tokens_mask[i] else None) for i, x in enumerate(A_ ) ] A : List[Any] = [x for x in filtered_sequence if x is not None] self.assertEqual(A_ , A_ ) @require_tokenizers class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): def _UpperCAmelCase ( self : int ): """simple docstring""" A : Any = AutoTokenizer.from_pretrained('''facebook/opt-350m''' , from_slow=A_ ) A : List[Any] = '''A photo of a cat''' A : Optional[int] = tokenizer.encode( A_ , ) self.assertEqual(A_ , [2, 250, 1345, 9, 10, 4758] ) tokenizer.save_pretrained('''test_opt''' ) A : Optional[int] = AutoTokenizer.from_pretrained('''./test_opt''' ) A : str = tokenizer.encode( A_ , ) self.assertEqual(A_ , [2, 250, 1345, 9, 10, 4758] ) def _UpperCAmelCase ( self : int ): """simple docstring""" A : Optional[int] = AutoTokenizer.from_pretrained('''facebook/opt-350m''' , use_slow=A_ ) A : List[Any] = '''A photo of a cat''' A : Union[str, Any] = tokenizer.encode( A_ , ) # Same as above self.assertEqual(A_ , [2, 250, 1345, 9, 10, 4758] ) @unittest.skip('''This test is failing because of a bug in the fast tokenizer''' ) def _UpperCAmelCase ( self : Optional[int] ): """simple docstring""" A : int = AutoTokenizer.from_pretrained('''facebook/opt-350m''' , from_slow=A_ ) A : List[str] = '''bos''' A : Tuple = tokenizer.get_vocab()['''bos'''] A : List[str] = '''A photo of a cat''' A : Dict = tokenizer.encode( A_ , ) # We changed the bos token self.assertEqual(A_ , [3_1957, 250, 1345, 9, 10, 4758] ) tokenizer.save_pretrained('''./tok''' ) A : Optional[int] = AutoTokenizer.from_pretrained('''./tok''' ) self.assertTrue(tokenizer.is_fast ) A : str = tokenizer.encode( A_ , ) self.assertEqual(A_ , [3_1957, 250, 1345, 9, 10, 4758] )
256
from __future__ import annotations from typing import Any class A( UpperCamelCase ): '''simple docstring''' pass class A: '''simple docstring''' def __init__( self : List[str] , A_ : Any ) -> None: """simple docstring""" lowerCamelCase_ = data lowerCamelCase_ = None def __iter__( self : int ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self lowerCamelCase_ = [] while node: if node in visited: raise ContainsLoopError visited.append(A_ ) yield node.data lowerCamelCase_ = node.next_node @property def a__ ( self : List[str] ) -> bool: """simple docstring""" try: list(self ) return False except ContainsLoopError: return True if __name__ == "__main__": lowerCamelCase : int = Node(1) lowerCamelCase : Optional[int] = Node(2) lowerCamelCase : Union[str, Any] = Node(3) lowerCamelCase : List[Any] = Node(4) print(root_node.has_loop) # False lowerCamelCase : int = root_node.next_node print(root_node.has_loop) # True lowerCamelCase : Dict = Node(5) lowerCamelCase : Optional[int] = Node(6) lowerCamelCase : str = Node(5) lowerCamelCase : Union[str, Any] = Node(6) print(root_node.has_loop) # False lowerCamelCase : List[str] = Node(1) print(root_node.has_loop) # False
70
0
"""simple docstring""" import importlib.metadata import operator import re import sys from typing import Optional from packaging import version _lowerCamelCase = { "<": operator.lt, "<=": operator.le, "==": operator.eq, "!=": operator.ne, ">=": operator.ge, ">": operator.gt, } def lowerCAmelCase_ ( lowercase_ : int , lowercase_ : Tuple , lowercase_ : List[Any] , lowercase_ : Tuple , lowercase_ : Any , lowercase_ : Optional[int] ): '''simple docstring''' if got_ver is None or want_ver is None: raise ValueError( F'''Unable to compare versions for {requirement}: need={want_ver} found={got_ver}. This is unusual. Consider''' F''' reinstalling {pkg}.''' ) if not ops[op](version.parse(lowercase_ ) , version.parse(lowercase_ ) ): raise ImportError( F'''{requirement} is required for a normal functioning of this module, but found {pkg}=={got_ver}.{hint}''' ) def lowerCAmelCase_ ( lowercase_ : str , lowercase_ : Optional[str] = None ): '''simple docstring''' __SCREAMING_SNAKE_CASE : Any = F'''\n{hint}''' if hint is not None else '''''' # non-versioned check if re.match(r'''^[\w_\-\d]+$''' , lowercase_ ): __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Optional[Any] = requirement, None, None else: __SCREAMING_SNAKE_CASE : Dict = re.findall(r'''^([^!=<>\s]+)([\s!=<>]{1,2}.+)''' , lowercase_ ) if not match: raise ValueError( '''requirement needs to be in the pip package format, .e.g., package_a==1.23, or package_b>=1.23, but''' F''' got {requirement}''' ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Union[str, Any] = match[0] __SCREAMING_SNAKE_CASE : Any = want_full.split(''',''' ) # there could be multiple requirements __SCREAMING_SNAKE_CASE : List[str] = {} for w in want_range: __SCREAMING_SNAKE_CASE : List[str] = re.findall(r'''^([\s!=<>]{1,2})(.+)''' , lowercase_ ) if not match: raise ValueError( '''requirement needs to be in the pip package format, .e.g., package_a==1.23, or package_b>=1.23,''' F''' but got {requirement}''' ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Union[str, Any] = match[0] __SCREAMING_SNAKE_CASE : List[str] = want_ver if op not in ops: raise ValueError(F'''{requirement}: need one of {list(ops.keys() )}, but got {op}''' ) # special case if pkg == "python": __SCREAMING_SNAKE_CASE : str = '''.'''.join([str(lowercase_ ) for x in sys.version_info[:3]] ) for op, want_ver in wanted.items(): _compare_versions(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) return # check if any version is installed try: __SCREAMING_SNAKE_CASE : List[Any] = importlib.metadata.version(lowercase_ ) except importlib.metadata.PackageNotFoundError: raise importlib.metadata.PackageNotFoundError( F'''The \'{requirement}\' distribution was not found and is required by this application. {hint}''' ) # check that the right version is installed if version number or a range was provided if want_ver is not None: for op, want_ver in wanted.items(): _compare_versions(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) def lowerCAmelCase_ ( lowercase_ : List[str] ): '''simple docstring''' __SCREAMING_SNAKE_CASE : Union[str, Any] = '''Try: pip install transformers -U or pip install -e \'.[dev]\' if you\'re working with git main''' return require_version(lowercase_ , lowercase_ )
674
import unittest import torch from diffusers import DDIMScheduler, DDPMScheduler, UNetaDModel from diffusers.training_utils import set_seed from diffusers.utils.testing_utils import slow lowerCamelCase : int = False class A( unittest.TestCase ): '''simple docstring''' def a__ ( self : int , A_ : Dict=32 ) -> Any: """simple docstring""" set_seed(0 ) lowerCamelCase_ = UNetaDModel(sample_size=A_ , in_channels=3 , out_channels=3 ) lowerCamelCase_ = torch.optim.SGD(model.parameters() , lr=0.0001 ) return model, optimizer @slow def a__ ( self : int ) -> str: """simple docstring""" lowerCamelCase_ = 'cpu' # ensure full determinism without setting the CUBLAS_WORKSPACE_CONFIG env variable lowerCamelCase_ = DDPMScheduler( num_train_timesteps=1000 , beta_start=0.0001 , beta_end=0.02 , beta_schedule='linear' , clip_sample=A_ , ) lowerCamelCase_ = DDIMScheduler( num_train_timesteps=1000 , beta_start=0.0001 , beta_end=0.02 , beta_schedule='linear' , clip_sample=A_ , ) assert ddpm_scheduler.config.num_train_timesteps == ddim_scheduler.config.num_train_timesteps # shared batches for DDPM and DDIM set_seed(0 ) lowerCamelCase_ = [torch.randn((4, 3, 32, 32) ).clip(-1 , 1 ).to(A_ ) for _ in range(4 )] lowerCamelCase_ = [torch.randn((4, 3, 32, 32) ).to(A_ ) for _ in range(4 )] lowerCamelCase_ = [torch.randint(0 , 1000 , (4,) ).long().to(A_ ) for _ in range(4 )] # train with a DDPM scheduler lowerCamelCase_ , lowerCamelCase_ = self.get_model_optimizer(resolution=32 ) model.train().to(A_ ) for i in range(4 ): optimizer.zero_grad() lowerCamelCase_ = ddpm_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) lowerCamelCase_ = model(A_ , timesteps[i] ).sample lowerCamelCase_ = torch.nn.functional.mse_loss(A_ , noise[i] ) loss.backward() optimizer.step() del model, optimizer # recreate the model and optimizer, and retry with DDIM lowerCamelCase_ , lowerCamelCase_ = self.get_model_optimizer(resolution=32 ) model.train().to(A_ ) for i in range(4 ): optimizer.zero_grad() lowerCamelCase_ = ddim_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) lowerCamelCase_ = model(A_ , timesteps[i] ).sample lowerCamelCase_ = torch.nn.functional.mse_loss(A_ , noise[i] ) loss.backward() optimizer.step() del model, optimizer self.assertTrue(torch.allclose(A_ , A_ , atol=1E-5 ) ) self.assertTrue(torch.allclose(A_ , A_ , atol=1E-5 ) )
70
0
'''simple docstring''' import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class lowercase_ ( A , A , unittest.TestCase ): """simple docstring""" lowerCamelCase_ = IFInpaintingSuperResolutionPipeline lowerCamelCase_ = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'''width''', '''height'''} lowerCamelCase_ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({'''original_image'''} ) lowerCamelCase_ = PipelineTesterMixin.required_optional_params - {'''latents'''} def lowerCAmelCase_ ( self : Dict ): """simple docstring""" return self._get_superresolution_dummy_components() def lowerCAmelCase_ ( self : str , __lowerCamelCase : List[str] , __lowerCamelCase : Tuple=0 ): """simple docstring""" if str(A_ ).startswith("mps" ): _SCREAMING_SNAKE_CASE = torch.manual_seed(A_ ) else: _SCREAMING_SNAKE_CASE = torch.Generator(device=A_ ).manual_seed(A_ ) _SCREAMING_SNAKE_CASE = floats_tensor((1, 3, 1_6, 1_6) , rng=random.Random(A_ ) ).to(A_ ) _SCREAMING_SNAKE_CASE = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(A_ ) ).to(A_ ) _SCREAMING_SNAKE_CASE = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(A_ ) ).to(A_ ) _SCREAMING_SNAKE_CASE = { "prompt": "A painting of a squirrel eating a burger", "image": image, "original_image": original_image, "mask_image": mask_image, "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def lowerCAmelCase_ ( self : Optional[Any] ): """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) def lowerCAmelCase_ ( self : List[str] ): """simple docstring""" self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def lowerCAmelCase_ ( self : Optional[int] ): """simple docstring""" super().test_save_load_floataa(expected_max_diff=1e-1 ) def lowerCAmelCase_ ( self : int ): """simple docstring""" self._test_attention_slicing_forward_pass(expected_max_diff=1e-2 ) def lowerCAmelCase_ ( self : int ): """simple docstring""" self._test_save_load_local() def lowerCAmelCase_ ( self : Union[str, Any] ): """simple docstring""" self._test_inference_batch_single_identical( expected_max_diff=1e-2 , )
418
def _SCREAMING_SNAKE_CASE ( lowercase : str , lowercase : str ): '''simple docstring''' if len(lowercase ) != len(lowercase ): raise ValueError('String lengths must match!' ) lowerCamelCase_ = 0 for chara, chara in zip(lowercase , lowercase ): if chara != chara: count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
70
0
'''simple docstring''' def _A ( UpperCAmelCase ,UpperCAmelCase ): '''simple docstring''' A__ = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def _A ( UpperCAmelCase ,UpperCAmelCase ,UpperCAmelCase ): '''simple docstring''' A__ = 0 while b > 0: if b & 1: A__ = ((res % c) + (a % c)) % c a += a b >>= 1 return res
531
def _SCREAMING_SNAKE_CASE ( lowercase : int = 10 ): '''simple docstring''' if not isinstance(lowercase , lowercase ) or n < 0: raise ValueError('Invalid input' ) lowerCamelCase_ = 10**n lowerCamelCase_ = 2_84_33 * (pow(2 , 7_83_04_57 , lowercase )) + 1 return str(number % modulus ) if __name__ == "__main__": from doctest import testmod testmod() print(F"""{solution(10) = }""")
70
0
'''simple docstring''' import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device lowerCamelCase : Optional[int] = False class A__ ( unittest.TestCase ): pass @nightly @require_torch_gpu class A__ ( unittest.TestCase ): def A ( self : str ) -> Tuple: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def A ( self : str ) -> str: '''simple docstring''' _SCREAMING_SNAKE_CASE =VersatileDiffusionPipeline.from_pretrained('shi-labs/versatile-diffusion' , torch_dtype=torch.floataa ) pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) _SCREAMING_SNAKE_CASE =load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) _SCREAMING_SNAKE_CASE =torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE =pipe.dual_guided( prompt='first prompt' , image=A_ , text_to_image_strength=0.75 , generator=A_ , guidance_scale=7.5 , num_inference_steps=2 , output_type='numpy' , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(A_ ) _SCREAMING_SNAKE_CASE =VersatileDiffusionPipeline.from_pretrained(A_ , torch_dtype=torch.floataa ) pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) _SCREAMING_SNAKE_CASE =generator.manual_seed(0 ) _SCREAMING_SNAKE_CASE =pipe.dual_guided( prompt='first prompt' , image=A_ , text_to_image_strength=0.75 , generator=A_ , guidance_scale=7.5 , num_inference_steps=2 , output_type='numpy' , ).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def A ( self : str ) -> Optional[int]: '''simple docstring''' _SCREAMING_SNAKE_CASE =VersatileDiffusionPipeline.from_pretrained('shi-labs/versatile-diffusion' , torch_dtype=torch.floataa ) pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) _SCREAMING_SNAKE_CASE ='cyberpunk 2077' _SCREAMING_SNAKE_CASE =load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) _SCREAMING_SNAKE_CASE =torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE =pipe.dual_guided( prompt=A_ , image=A_ , text_to_image_strength=0.75 , generator=A_ , guidance_scale=7.5 , num_inference_steps=50 , output_type='numpy' , ).images _SCREAMING_SNAKE_CASE =image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _SCREAMING_SNAKE_CASE =np.array([0.14_48, 0.16_19, 0.17_41, 0.10_86, 0.11_47, 0.11_28, 0.11_99, 0.11_65, 0.10_01] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 _SCREAMING_SNAKE_CASE ='A painting of a squirrel eating a burger ' _SCREAMING_SNAKE_CASE =torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE =pipe.text_to_image( prompt=A_ , generator=A_ , guidance_scale=7.5 , num_inference_steps=50 , output_type='numpy' ).images _SCREAMING_SNAKE_CASE =image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _SCREAMING_SNAKE_CASE =np.array([0.33_67, 0.31_69, 0.26_56, 0.38_70, 0.47_90, 0.37_96, 0.40_09, 0.48_78, 0.47_78] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 _SCREAMING_SNAKE_CASE =pipe.image_variation(A_ , generator=A_ , output_type='numpy' ).images _SCREAMING_SNAKE_CASE =image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _SCREAMING_SNAKE_CASE =np.array([0.30_76, 0.31_23, 0.32_84, 0.37_82, 0.37_70, 0.38_94, 0.42_97, 0.43_31, 0.44_56] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1
405
from maths.prime_check import is_prime def _SCREAMING_SNAKE_CASE ( lowercase : int ): '''simple docstring''' if not isinstance(lowercase , lowercase ): lowerCamelCase_ = f"""Input value of [number={number}] must be an integer""" raise TypeError(lowercase ) if is_prime(lowercase ) and is_prime(number + 2 ): return number + 2 else: return -1 if __name__ == "__main__": import doctest doctest.testmod()
70
0
'''simple docstring''' import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartTokenizer, MBartTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin _UpperCamelCase : Optional[int] = get_tests_dir("fixtures/test_sentencepiece.model") if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right _UpperCamelCase : Union[str, Any] = 25_0004 _UpperCamelCase : Tuple = 25_0020 @require_sentencepiece @require_tokenizers class _snake_case ( a_ , unittest.TestCase ): SCREAMING_SNAKE_CASE : Tuple = MBartTokenizer SCREAMING_SNAKE_CASE : List[str] = MBartTokenizerFast SCREAMING_SNAKE_CASE : Union[str, Any] = True SCREAMING_SNAKE_CASE : Any = True def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing lowerCAmelCase = MBartTokenizer(A_ , keep_accents=A_ ) tokenizer.save_pretrained(self.tmpdirname ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowerCAmelCase = MBartTokenizer(A_ , keep_accents=A_ ) lowerCAmelCase = tokenizer.tokenize('This is a test' ) self.assertListEqual(A_ , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(A_ ) , [value + tokenizer.fairseq_offset for value in [2_85, 46, 10, 1_70, 3_82]] , ) lowerCAmelCase = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( A_ , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) lowerCAmelCase = tokenizer.convert_tokens_to_ids(A_ ) self.assertListEqual( A_ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) lowerCAmelCase = tokenizer.convert_ids_to_tokens(A_ ) self.assertListEqual( A_ , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return lowerCAmelCase = (self.rust_tokenizer_class, 'hf-internal-testing/tiny-random-mbart', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})' ): lowerCAmelCase = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) lowerCAmelCase = self.tokenizer_class.from_pretrained(A_ , **A_ ) lowerCAmelCase = tempfile.mkdtemp() lowerCAmelCase = tokenizer_r.save_pretrained(A_ ) lowerCAmelCase = tokenizer_p.save_pretrained(A_ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('tokenizer.json' in f for f in tokenizer_r_files ) ) lowerCAmelCase = tuple(f for f in tokenizer_r_files if 'tokenizer.json' not in f ) self.assertSequenceEqual(A_ , A_ ) # Checks everything loads correctly in the same way lowerCAmelCase = tokenizer_r.from_pretrained(A_ ) lowerCAmelCase = tokenizer_p.from_pretrained(A_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(A_ , A_ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(A_ ) # Save tokenizer rust, legacy_format=True lowerCAmelCase = tempfile.mkdtemp() lowerCAmelCase = tokenizer_r.save_pretrained(A_ , legacy_format=A_ ) lowerCAmelCase = tokenizer_p.save_pretrained(A_ ) # Checks it save with the same files self.assertSequenceEqual(A_ , A_ ) # Checks everything loads correctly in the same way lowerCAmelCase = tokenizer_r.from_pretrained(A_ ) lowerCAmelCase = tokenizer_p.from_pretrained(A_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(A_ , A_ ) ) shutil.rmtree(A_ ) # Save tokenizer rust, legacy_format=False lowerCAmelCase = tempfile.mkdtemp() lowerCAmelCase = tokenizer_r.save_pretrained(A_ , legacy_format=A_ ) lowerCAmelCase = tokenizer_p.save_pretrained(A_ ) # Checks it saved the tokenizer.json file self.assertTrue(any('tokenizer.json' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way lowerCAmelCase = tokenizer_r.from_pretrained(A_ ) lowerCAmelCase = tokenizer_p.from_pretrained(A_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(A_ , A_ ) ) shutil.rmtree(A_ ) @require_torch @require_sentencepiece @require_tokenizers class _snake_case ( unittest.TestCase ): SCREAMING_SNAKE_CASE : List[Any] = '''facebook/mbart-large-en-ro''' SCREAMING_SNAKE_CASE : List[Any] = [ ''' UN Chief Says There Is No Military Solution in Syria''', ''' Secretary-General Ban Ki-moon says his response to Russia\'s stepped up military support for Syria is that "there is no military solution" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.''', ] SCREAMING_SNAKE_CASE : Union[str, Any] = [ '''Şeful ONU declară că nu există o soluţie militară în Siria''', '''Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei''' ''' pentru Siria este că "nu există o soluţie militară" la conflictul de aproape cinci ani şi că noi arme nu vor''' ''' face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.''', ] SCREAMING_SNAKE_CASE : Union[str, Any] = [8274, 127873, 25916, 7, 8622, 2071, 438, 67485, 53, 187895, 23, 51712, 2, EN_CODE] @classmethod def _SCREAMING_SNAKE_CASE ( cls ): '''simple docstring''' lowerCAmelCase = MBartTokenizer.from_pretrained( cls.checkpoint_name , src_lang='en_XX' , tgt_lang='ro_RO' ) lowerCAmelCase = 1 return cls def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['ar_AR'] , 25_00_01 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['en_EN'] , 25_00_04 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['ro_RO'] , 25_00_20 ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowerCAmelCase = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , A_ ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' self.assertIn(A_ , self.tokenizer.all_special_ids ) lowerCAmelCase = [RO_CODE, 8_84, 90_19, 96, 9, 9_16, 8_67_92, 36, 1_87_43, 1_55_96, 5, 2] lowerCAmelCase = self.tokenizer.decode(A_ , skip_special_tokens=A_ ) lowerCAmelCase = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=A_ ) self.assertEqual(A_ , A_ ) self.assertNotIn(self.tokenizer.eos_token , A_ ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowerCAmelCase = ['this is gunna be a long sentence ' * 20] assert isinstance(src_text[0] , A_ ) lowerCAmelCase = 10 lowerCAmelCase = self.tokenizer(A_ , max_length=A_ , truncation=A_ ).input_ids[0] self.assertEqual(ids[-2] , 2 ) self.assertEqual(ids[-1] , A_ ) self.assertEqual(len(A_ ) , A_ ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' self.assertListEqual(self.tokenizer.convert_tokens_to_ids(['<mask>', 'ar_AR'] ) , [25_00_26, 25_00_01] ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowerCAmelCase = tempfile.mkdtemp() lowerCAmelCase = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(A_ ) lowerCAmelCase = MBartTokenizer.from_pretrained(A_ ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , A_ ) @require_torch def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowerCAmelCase = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=A_ , return_tensors='pt' ) lowerCAmelCase = shift_tokens_right(batch['labels'] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][-2:].tolist() == [2, EN_CODE] assert batch.decoder_input_ids[1][0].tolist() == RO_CODE assert batch.decoder_input_ids[1][-1] == 2 assert batch.labels[1][-2:].tolist() == [2, RO_CODE] @require_torch def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowerCAmelCase = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=A_ , truncation=A_ , max_length=len(self.expected_src_tokens ) , return_tensors='pt' , ) lowerCAmelCase = shift_tokens_right(batch['labels'] , self.tokenizer.pad_token_id ) self.assertIsInstance(A_ , A_ ) self.assertEqual((2, 14) , batch.input_ids.shape ) self.assertEqual((2, 14) , batch.attention_mask.shape ) lowerCAmelCase = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , A_ ) self.assertEqual(2 , batch.decoder_input_ids[0, -1] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, EN_CODE] ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowerCAmelCase = self.tokenizer(self.src_text , padding=A_ , truncation=A_ , max_length=3 , return_tensors='pt' ) lowerCAmelCase = self.tokenizer( text_target=self.tgt_text , padding=A_ , truncation=A_ , max_length=10 , return_tensors='pt' ) lowerCAmelCase = targets['input_ids'] lowerCAmelCase = shift_tokens_right(A_ , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowerCAmelCase = self.tokenizer._build_translation_inputs( 'A test' , return_tensors='pt' , src_lang='en_XX' , tgt_lang='ar_AR' ) self.assertEqual( nested_simplify(A_ ) , { # A, test, EOS, en_XX 'input_ids': [[62, 30_34, 2, 25_00_04]], 'attention_mask': [[1, 1, 1, 1]], # ar_AR 'forced_bos_token_id': 25_00_01, } , )
284
# Algorithm for the pigeonhole sorting def _SCREAMING_SNAKE_CASE ( lowercase : str ): '''simple docstring''' lowerCamelCase_ = min(lowercase ) # min() finds the minimum value lowerCamelCase_ = max(lowercase ) # max() finds the maximum value lowerCamelCase_ = max_val - min_val + 1 # size is difference of max and min values plus one # list of pigeonholes of size equal to the variable size lowerCamelCase_ = [0] * size # Populate the pigeonholes. for x in a: assert isinstance(lowercase , lowercase ), "integers only please" holes[x - min_val] += 1 # Putting the elements back into the array in an order. lowerCamelCase_ = 0 for count in range(lowercase ): while holes[count] > 0: holes[count] -= 1 lowerCamelCase_ = count + min_val i += 1 def _SCREAMING_SNAKE_CASE ( ): '''simple docstring''' lowerCamelCase_ = [8, 3, 2, 7, 4, 6, 8] pigeonhole_sort(lowercase ) print('Sorted order is:' , ' '.join(lowercase ) ) if __name__ == "__main__": main()
70
0
'''simple docstring''' import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { "asapp/sew-tiny-100k": "https://huggingface.co/asapp/sew-tiny-100k/resolve/main/config.json", # See all SEW models at https://huggingface.co/models?filter=sew } class __SCREAMING_SNAKE_CASE (__A ): """simple docstring""" _a : Dict = '''sew''' def __init__( self , UpperCamelCase__=32 , UpperCamelCase__=768 , UpperCamelCase__=12 , UpperCamelCase__=12 , UpperCamelCase__=3_072 , UpperCamelCase__=2 , UpperCamelCase__="gelu" , UpperCamelCase__=0.1 , UpperCamelCase__=0.1 , UpperCamelCase__=0.1 , UpperCamelCase__=0.0 , UpperCamelCase__=0.1 , UpperCamelCase__=0.1 , UpperCamelCase__=0.02 , UpperCamelCase__=1e-5 , UpperCamelCase__="group" , UpperCamelCase__="gelu" , UpperCamelCase__=(64, 128, 128, 128, 128, 256, 256, 256, 256, 512, 512, 512, 512) , UpperCamelCase__=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , UpperCamelCase__=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , UpperCamelCase__=False , UpperCamelCase__=128 , UpperCamelCase__=16 , UpperCamelCase__=True , UpperCamelCase__=0.05 , UpperCamelCase__=10 , UpperCamelCase__=2 , UpperCamelCase__=0.0 , UpperCamelCase__=10 , UpperCamelCase__=0 , UpperCamelCase__="mean" , UpperCamelCase__=False , UpperCamelCase__=False , UpperCamelCase__=256 , UpperCamelCase__=0 , UpperCamelCase__=1 , UpperCamelCase__=2 , **UpperCamelCase__ , ): """simple docstring""" super().__init__(**A_ , pad_token_id=A_ , bos_token_id=A_ , eos_token_id=A_ ) a_ = hidden_size a_ = feat_extract_norm a_ = feat_extract_activation a_ = list(A_ ) a_ = list(A_ ) a_ = list(A_ ) a_ = conv_bias a_ = num_conv_pos_embeddings a_ = num_conv_pos_embedding_groups a_ = len(self.conv_dim ) a_ = num_hidden_layers a_ = intermediate_size a_ = squeeze_factor a_ = hidden_act a_ = num_attention_heads a_ = hidden_dropout a_ = attention_dropout a_ = activation_dropout a_ = feat_proj_dropout a_ = final_dropout a_ = layerdrop a_ = layer_norm_eps a_ = initializer_range a_ = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( 'Configuration for convolutional layers is incorrect.' 'It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,' f'but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)' f'= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`.' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 a_ = apply_spec_augment a_ = mask_time_prob a_ = mask_time_length a_ = mask_time_min_masks a_ = mask_feature_prob a_ = mask_feature_length a_ = mask_feature_min_masks # ctc loss a_ = ctc_loss_reduction a_ = ctc_zero_infinity # sequence classification a_ = use_weighted_layer_sum a_ = classifier_proj_size @property def _a ( self ): """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
536
import os import unittest from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, BertTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class A( UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = BertTokenizer UpperCamelCase = BertTokenizerFast UpperCamelCase = True UpperCamelCase = True UpperCamelCase = filter_non_english def a__ ( self : List[Any] ) -> int: """simple docstring""" super().setUp() lowerCamelCase_ = [ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] lowerCamelCase_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def a__ ( self : Tuple , A_ : List[str] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = 'UNwant\u00E9d,running' lowerCamelCase_ = 'unwanted, running' return input_text, output_text def a__ ( self : Any ) -> Tuple: """simple docstring""" lowerCamelCase_ = self.tokenizer_class(self.vocab_file ) lowerCamelCase_ = tokenizer.tokenize('UNwant\u00E9d,running' ) self.assertListEqual(A_ , ['un', '##want', '##ed', ',', 'runn', '##ing'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A_ ) , [9, 6, 7, 12, 10, 11] ) def a__ ( self : Tuple ) -> List[str]: """simple docstring""" if not self.test_rust_tokenizer: return lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_rust_tokenizer() lowerCamelCase_ = 'UNwant\u00E9d,running' lowerCamelCase_ = tokenizer.tokenize(A_ ) lowerCamelCase_ = rust_tokenizer.tokenize(A_ ) self.assertListEqual(A_ , A_ ) lowerCamelCase_ = tokenizer.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = rust_tokenizer.encode(A_ , add_special_tokens=A_ ) self.assertListEqual(A_ , A_ ) lowerCamelCase_ = self.get_rust_tokenizer() lowerCamelCase_ = tokenizer.encode(A_ ) lowerCamelCase_ = rust_tokenizer.encode(A_ ) self.assertListEqual(A_ , A_ ) # With lower casing lowerCamelCase_ = self.get_tokenizer(do_lower_case=A_ ) lowerCamelCase_ = self.get_rust_tokenizer(do_lower_case=A_ ) lowerCamelCase_ = 'UNwant\u00E9d,running' lowerCamelCase_ = tokenizer.tokenize(A_ ) lowerCamelCase_ = rust_tokenizer.tokenize(A_ ) self.assertListEqual(A_ , A_ ) lowerCamelCase_ = tokenizer.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = rust_tokenizer.encode(A_ , add_special_tokens=A_ ) self.assertListEqual(A_ , A_ ) lowerCamelCase_ = self.get_rust_tokenizer() lowerCamelCase_ = tokenizer.encode(A_ ) lowerCamelCase_ = rust_tokenizer.encode(A_ ) self.assertListEqual(A_ , A_ ) def a__ ( self : Any ) -> Dict: """simple docstring""" lowerCamelCase_ = BasicTokenizer() self.assertListEqual(tokenizer.tokenize('ah\u535A\u63A8zz' ) , ['ah', '\u535A', '\u63A8', 'zz'] ) def a__ ( self : Dict ) -> int: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? ' ) , ['hello', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['hello'] ) def a__ ( self : str ) -> int: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ , strip_accents=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['hällo', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['h\u00E9llo'] ) def a__ ( self : Optional[int] ) -> Any: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ , strip_accents=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['hallo', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['hello'] ) def a__ ( self : Tuple ) -> str: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['hallo', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['hello'] ) def a__ ( self : int ) -> List[Any]: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? ' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?'] ) def a__ ( self : str ) -> List[str]: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ , strip_accents=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['HäLLo', '!', 'how', 'Are', 'yoU', '?'] ) def a__ ( self : Dict ) -> Any: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ , strip_accents=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['HaLLo', '!', 'how', 'Are', 'yoU', '?'] ) def a__ ( self : int ) -> Any: """simple docstring""" lowerCamelCase_ = BasicTokenizer(do_lower_case=A_ , never_split=['[UNK]'] ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? [UNK]' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?', '[UNK]'] ) def a__ ( self : List[Any] ) -> int: """simple docstring""" lowerCamelCase_ = BasicTokenizer() lowerCamelCase_ = 'a\n\'ll !!to?\'d of, can\'t.' lowerCamelCase_ = ['a', '\'', 'll', '!', '!', 'to', '?', '\'', 'd', 'of', ',', 'can', '\'', 't', '.'] self.assertListEqual(tokenizer.tokenize(A_ ) , A_ ) def a__ ( self : Optional[int] ) -> Dict: """simple docstring""" lowerCamelCase_ = ['[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing'] lowerCamelCase_ = {} for i, token in enumerate(A_ ): lowerCamelCase_ = i lowerCamelCase_ = WordpieceTokenizer(vocab=A_ , unk_token='[UNK]' ) self.assertListEqual(tokenizer.tokenize('' ) , [] ) self.assertListEqual(tokenizer.tokenize('unwanted running' ) , ['un', '##want', '##ed', 'runn', '##ing'] ) self.assertListEqual(tokenizer.tokenize('unwantedX running' ) , ['[UNK]', 'runn', '##ing'] ) def a__ ( self : Optional[Any] ) -> str: """simple docstring""" self.assertTrue(_is_whitespace(' ' ) ) self.assertTrue(_is_whitespace('\t' ) ) self.assertTrue(_is_whitespace('\r' ) ) self.assertTrue(_is_whitespace('\n' ) ) self.assertTrue(_is_whitespace('\u00A0' ) ) self.assertFalse(_is_whitespace('A' ) ) self.assertFalse(_is_whitespace('-' ) ) def a__ ( self : List[Any] ) -> int: """simple docstring""" self.assertTrue(_is_control('\u0005' ) ) self.assertFalse(_is_control('A' ) ) self.assertFalse(_is_control(' ' ) ) self.assertFalse(_is_control('\t' ) ) self.assertFalse(_is_control('\r' ) ) def a__ ( self : List[str] ) -> Optional[Any]: """simple docstring""" self.assertTrue(_is_punctuation('-' ) ) self.assertTrue(_is_punctuation('$' ) ) self.assertTrue(_is_punctuation('`' ) ) self.assertTrue(_is_punctuation('.' ) ) self.assertFalse(_is_punctuation('A' ) ) self.assertFalse(_is_punctuation(' ' ) ) def a__ ( self : int ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_rust_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(A_ ) for t in ['Test', '\xad', 'test']] , [['[UNK]'], [], ['[UNK]']] ) self.assertListEqual( [rust_tokenizer.tokenize(A_ ) for t in ['Test', '\xad', 'test']] , [['[UNK]'], [], ['[UNK]']] ) @slow def a__ ( self : Any ) -> int: """simple docstring""" lowerCamelCase_ = self.tokenizer_class.from_pretrained('bert-base-uncased' ) lowerCamelCase_ = tokenizer.encode('sequence builders' , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer.encode('multi-sequence build' , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer.build_inputs_with_special_tokens(A_ ) lowerCamelCase_ = tokenizer.build_inputs_with_special_tokens(A_ , A_ ) assert encoded_sentence == [101] + text + [102] assert encoded_pair == [101] + text + [102] + text_a + [102] def a__ ( self : str ) -> str: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCamelCase_ = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) lowerCamelCase_ = f"""A, naïve {tokenizer_r.mask_token} AllenNLP sentence.""" lowerCamelCase_ = tokenizer_r.encode_plus( A_ , return_attention_mask=A_ , return_token_type_ids=A_ , return_offsets_mapping=A_ , add_special_tokens=A_ , ) lowerCamelCase_ = tokenizer_r.do_lower_case if hasattr(A_ , 'do_lower_case' ) else False lowerCamelCase_ = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), 'A'), ((1, 2), ','), ((3, 5), 'na'), ((5, 6), '##ï'), ((6, 8), '##ve'), ((9, 15), tokenizer_r.mask_token), ((16, 21), 'Allen'), ((21, 23), '##NL'), ((23, 24), '##P'), ((25, 33), 'sentence'), ((33, 34), '.'), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), 'a'), ((1, 2), ','), ((3, 8), 'naive'), ((9, 15), tokenizer_r.mask_token), ((16, 21), 'allen'), ((21, 23), '##nl'), ((23, 24), '##p'), ((25, 33), 'sentence'), ((33, 34), '.'), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens['input_ids'] ) ) self.assertEqual([e[0] for e in expected_results] , tokens['offset_mapping'] ) def a__ ( self : List[Any] ) -> Dict: """simple docstring""" lowerCamelCase_ = ['的', '人', '有'] lowerCamelCase_ = ''.join(A_ ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCamelCase_ = True lowerCamelCase_ = self.tokenizer_class.from_pretrained(A_ , **A_ ) lowerCamelCase_ = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) lowerCamelCase_ = tokenizer_p.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer_r.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer_r.convert_ids_to_tokens(A_ ) lowerCamelCase_ = tokenizer_p.convert_ids_to_tokens(A_ ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(A_ , A_ ) self.assertListEqual(A_ , A_ ) lowerCamelCase_ = False lowerCamelCase_ = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) lowerCamelCase_ = self.tokenizer_class.from_pretrained(A_ , **A_ ) lowerCamelCase_ = tokenizer_r.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer_p.encode(A_ , add_special_tokens=A_ ) lowerCamelCase_ = tokenizer_r.convert_ids_to_tokens(A_ ) lowerCamelCase_ = tokenizer_p.convert_ids_to_tokens(A_ ) # it is expected that only the first Chinese character is not preceded by "##". lowerCamelCase_ = [ f"""##{token}""" if idx != 0 else token for idx, token in enumerate(A_ ) ] self.assertListEqual(A_ , A_ ) self.assertListEqual(A_ , A_ )
70
0
from math import sqrt def A ( __UpperCamelCase = 1_000_000 ) -> List[str]: A__ = 0 A__ = 0 A__ = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(__UpperCamelCase , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(f'{solution() = }')
9
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) lowerCamelCase : Dict = logging.get_logger(__name__) # pylint: disable=invalid-name lowerCamelCase : List[str] = "\n Examples:\n ```py\n >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline\n >>> import torch\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-prior\")\n >>> pipe_prior.to(\"cuda\")\n >>> prompt = \"red cat, 4k photo\"\n >>> out = pipe_prior(prompt)\n >>> image_emb = out.image_embeds\n >>> zero_image_emb = out.negative_image_embeds\n >>> pipe = KandinskyV22Pipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-decoder\")\n >>> pipe.to(\"cuda\")\n >>> image = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=50,\n ... ).images\n >>> image[0].save(\"cat.png\")\n ```\n" def _SCREAMING_SNAKE_CASE ( lowercase : Any , lowercase : str , lowercase : Any=8 ): '''simple docstring''' lowerCamelCase_ = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 lowerCamelCase_ = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class A( UpperCamelCase ): '''simple docstring''' def __init__( self : str , A_ : UNetaDConditionModel , A_ : DDPMScheduler , A_ : VQModel , ) -> List[str]: """simple docstring""" super().__init__() self.register_modules( unet=A_ , scheduler=A_ , movq=A_ , ) lowerCamelCase_ = 2 ** (len(self.movq.config.block_out_channels ) - 1) def a__ ( self : List[Any] , A_ : Tuple , A_ : Dict , A_ : List[Any] , A_ : int , A_ : Any , A_ : Tuple ) -> Any: """simple docstring""" if latents is None: lowerCamelCase_ = randn_tensor(A_ , generator=A_ , device=A_ , dtype=A_ ) else: if latents.shape != shape: raise ValueError(f"""Unexpected latents shape, got {latents.shape}, expected {shape}""" ) lowerCamelCase_ = latents.to(A_ ) lowerCamelCase_ = latents * scheduler.init_noise_sigma return latents def a__ ( self : int , A_ : str=0 ) -> Optional[int]: """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) lowerCamelCase_ = torch.device(f"""cuda:{gpu_id}""" ) lowerCamelCase_ = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(A_ , A_ ) def a__ ( self : Tuple , A_ : Union[str, Any]=0 ) -> Dict: """simple docstring""" if is_accelerate_available() and is_accelerate_version('>=' , '0.17.0.dev0' ): from accelerate import cpu_offload_with_hook else: raise ImportError('`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.' ) lowerCamelCase_ = torch.device(f"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to('cpu' , silence_dtype_warnings=A_ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) lowerCamelCase_ = None for cpu_offloaded_model in [self.unet, self.movq]: lowerCamelCase_ , lowerCamelCase_ = cpu_offload_with_hook(A_ , A_ , prev_module_hook=A_ ) # We'll offload the last model manually. lowerCamelCase_ = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def a__ ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" if not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(A_ , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(A_ ) def __call__( self : List[Any] , A_ : Union[torch.FloatTensor, List[torch.FloatTensor]] , A_ : Union[torch.FloatTensor, List[torch.FloatTensor]] , A_ : int = 512 , A_ : int = 512 , A_ : int = 100 , A_ : float = 4.0 , A_ : int = 1 , A_ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , A_ : Optional[torch.FloatTensor] = None , A_ : Optional[str] = "pil" , A_ : bool = True , ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self._execution_device lowerCamelCase_ = guidance_scale > 1.0 if isinstance(A_ , A_ ): lowerCamelCase_ = torch.cat(A_ , dim=0 ) lowerCamelCase_ = image_embeds.shape[0] * num_images_per_prompt if isinstance(A_ , A_ ): lowerCamelCase_ = torch.cat(A_ , dim=0 ) if do_classifier_free_guidance: lowerCamelCase_ = image_embeds.repeat_interleave(A_ , dim=0 ) lowerCamelCase_ = negative_image_embeds.repeat_interleave(A_ , dim=0 ) lowerCamelCase_ = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=A_ ) self.scheduler.set_timesteps(A_ , device=A_ ) lowerCamelCase_ = self.scheduler.timesteps lowerCamelCase_ = self.unet.config.in_channels lowerCamelCase_ , lowerCamelCase_ = downscale_height_and_width(A_ , A_ , self.movq_scale_factor ) # create initial latent lowerCamelCase_ = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , A_ , A_ , A_ , self.scheduler , ) for i, t in enumerate(self.progress_bar(A_ ) ): # expand the latents if we are doing classifier free guidance lowerCamelCase_ = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowerCamelCase_ = {'image_embeds': image_embeds} lowerCamelCase_ = self.unet( sample=A_ , timestep=A_ , encoder_hidden_states=A_ , added_cond_kwargs=A_ , return_dict=A_ , )[0] if do_classifier_free_guidance: lowerCamelCase_ , lowerCamelCase_ = noise_pred.split(latents.shape[1] , dim=1 ) lowerCamelCase_ , lowerCamelCase_ = noise_pred.chunk(2 ) lowerCamelCase_ , lowerCamelCase_ = variance_pred.chunk(2 ) lowerCamelCase_ = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) lowerCamelCase_ = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , 'variance_type' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): lowerCamelCase_ , lowerCamelCase_ = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 lowerCamelCase_ = self.scheduler.step( A_ , A_ , A_ , generator=A_ , )[0] # post-processing lowerCamelCase_ = self.movq.decode(A_ , force_not_quantize=A_ )['sample'] if output_type not in ["pt", "np", "pil"]: raise ValueError(f"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: lowerCamelCase_ = image * 0.5 + 0.5 lowerCamelCase_ = image.clamp(0 , 1 ) lowerCamelCase_ = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowerCamelCase_ = self.numpy_to_pil(A_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=A_ )
70
0
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class _snake_case ( snake_case , unittest.TestCase ): """simple docstring""" _UpperCamelCase = ShapEPipeline _UpperCamelCase = ["prompt"] _UpperCamelCase = ["prompt"] _UpperCamelCase = [ "num_images_per_prompt", "num_inference_steps", "generator", "latents", "guidance_scale", "frame_size", "output_type", "return_dict", ] _UpperCamelCase = False @property def __SCREAMING_SNAKE_CASE ( self ) -> str: return 32 @property def __SCREAMING_SNAKE_CASE ( self ) -> Tuple: return 32 @property def __SCREAMING_SNAKE_CASE ( self ) -> int: return self.time_input_dim * 4 @property def __SCREAMING_SNAKE_CASE ( self ) -> Tuple: return 8 @property def __SCREAMING_SNAKE_CASE ( self ) -> Union[str, Any]: a_ = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) return tokenizer @property def __SCREAMING_SNAKE_CASE ( self ) -> Dict: torch.manual_seed(0 ) a_ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModelWithProjection(A_ ) @property def __SCREAMING_SNAKE_CASE ( self ) -> Dict: torch.manual_seed(0 ) a_ = { 'num_attention_heads': 2, 'attention_head_dim': 16, 'embedding_dim': self.time_input_dim, 'num_embeddings': 32, 'embedding_proj_dim': self.text_embedder_hidden_size, 'time_embed_dim': self.time_embed_dim, 'num_layers': 1, 'clip_embed_dim': self.time_input_dim * 2, 'additional_embeddings': 0, 'time_embed_act_fn': 'gelu', 'norm_in_type': 'layer', 'encoder_hid_proj_type': None, 'added_emb_type': None, } a_ = PriorTransformer(**A_ ) return model @property def __SCREAMING_SNAKE_CASE ( self ) -> Any: torch.manual_seed(0 ) a_ = { 'param_shapes': ( (self.renderer_dim, 93), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), 'd_latent': self.time_input_dim, 'd_hidden': self.renderer_dim, 'n_output': 12, 'background': ( 0.1, 0.1, 0.1, ), } a_ = ShapERenderer(**A_ ) return model def __SCREAMING_SNAKE_CASE ( self ) -> str: a_ = self.dummy_prior a_ = self.dummy_text_encoder a_ = self.dummy_tokenizer a_ = self.dummy_renderer a_ = HeunDiscreteScheduler( beta_schedule='exp' , num_train_timesteps=1024 , prediction_type='sample' , use_karras_sigmas=A_ , clip_sample=A_ , clip_sample_range=1.0 , ) a_ = { 'prior': prior, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'renderer': renderer, 'scheduler': scheduler, } return components def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase__ , UpperCAmelCase__=0 ) -> Optional[int]: if str(A_ ).startswith('mps' ): a_ = torch.manual_seed(A_ ) else: a_ = torch.Generator(device=A_ ).manual_seed(A_ ) a_ = { 'prompt': 'horse', 'generator': generator, 'num_inference_steps': 1, 'frame_size': 32, 'output_type': 'np', } return inputs def __SCREAMING_SNAKE_CASE ( self ) -> Union[str, Any]: a_ = 'cpu' a_ = self.get_dummy_components() a_ = self.pipeline_class(**A_ ) a_ = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) a_ = pipe(**self.get_dummy_inputs(A_ ) ) a_ = output.images[0] a_ = image[0, -3:, -3:, -1] assert image.shape == (20, 32, 32, 3) a_ = np.array( [ 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def __SCREAMING_SNAKE_CASE ( self ) -> Union[str, Any]: self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def __SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: a_ = torch_device == 'cpu' a_ = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=A_ , relax_max_difference=A_ , ) def __SCREAMING_SNAKE_CASE ( self ) -> Union[str, Any]: a_ = self.get_dummy_components() a_ = self.pipeline_class(**A_ ) a_ = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) a_ = 1 a_ = 2 a_ = self.get_dummy_inputs(A_ ) for key in inputs.keys(): if key in self.batch_params: a_ = batch_size * [inputs[key]] a_ = pipe(**A_ , num_images_per_prompt=A_ )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class _snake_case ( unittest.TestCase ): """simple docstring""" def __SCREAMING_SNAKE_CASE ( self ) -> int: super().tearDown() gc.collect() torch.cuda.empty_cache() def __SCREAMING_SNAKE_CASE ( self ) -> Tuple: a_ = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/shap_e/test_shap_e_np_out.npy' ) a_ = ShapEPipeline.from_pretrained('openai/shap-e' ) a_ = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) a_ = torch.Generator(device=A_ ).manual_seed(0 ) a_ = pipe( 'a shark' , generator=A_ , guidance_scale=1_5.0 , num_inference_steps=64 , frame_size=64 , output_type='np' , ).images[0] assert images.shape == (20, 64, 64, 3) assert_mean_pixel_difference(A_ , A_ )
697
from PIL import Image def _SCREAMING_SNAKE_CASE ( lowercase : Image ): '''simple docstring''' lowerCamelCase_ , lowerCamelCase_ = image.size lowerCamelCase_ = 0 lowerCamelCase_ = image.load() for i in range(lowercase ): for j in range(lowercase ): lowerCamelCase_ = pixels[j, i] mean += pixel mean //= width * height for j in range(lowercase ): for i in range(lowercase ): lowerCamelCase_ = 2_55 if pixels[i, j] > mean else 0 return image if __name__ == "__main__": lowerCamelCase : Optional[Any] = mean_threshold(Image.open("path_to_image").convert("L")) image.save("output_image_path")
70
0
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTImageProcessor, ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() A : List[str] = logging.get_logger(__name__) def __lowerCAmelCase ( a__ , a__=False ) -> Union[str, Any]: __a = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"""blocks.{i}.norm1.weight""", F"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""blocks.{i}.norm1.bias""", F"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((F"""blocks.{i}.attn.proj.weight""", F"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.attn.proj.bias""", F"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((F"""blocks.{i}.norm2.weight""", F"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""blocks.{i}.norm2.bias""", F"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.weight""", F"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.bias""", F"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.weight""", F"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.bias""", F"""vit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ('''cls_token''', '''vit.embeddings.cls_token'''), ('''patch_embed.proj.weight''', '''vit.embeddings.patch_embeddings.projection.weight'''), ('''patch_embed.proj.bias''', '''vit.embeddings.patch_embeddings.projection.bias'''), ('''pos_embed''', '''vit.embeddings.position_embeddings'''), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ('''pre_logits.fc.weight''', '''pooler.dense.weight'''), ('''pre_logits.fc.bias''', '''pooler.dense.bias'''), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" __a = [(pair[0], pair[1][4:]) if pair[1].startswith('''vit''' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('''norm.weight''', '''vit.layernorm.weight'''), ('''norm.bias''', '''vit.layernorm.bias'''), ('''head.weight''', '''classifier.weight'''), ('''head.bias''', '''classifier.bias'''), ] ) return rename_keys def __lowerCAmelCase ( a__ , a__ , a__=False ) -> Optional[int]: for i in range(config.num_hidden_layers ): if base_model: __a = '''''' else: __a = '''vit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) __a = state_dict.pop(F"""blocks.{i}.attn.qkv.weight""" ) __a = state_dict.pop(F"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict __a = in_proj_weight[ : config.hidden_size, : ] __a = in_proj_bias[: config.hidden_size] __a = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] __a = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] __a = in_proj_weight[ -config.hidden_size :, : ] __a = in_proj_bias[-config.hidden_size :] def __lowerCAmelCase ( a__ ) -> Any: __a = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(a__ , a__ ) def __lowerCAmelCase ( a__ , a__ , a__ ) -> Dict: __a = dct.pop(a__ ) __a = val def __lowerCAmelCase ( ) -> Dict: __a = '''http://images.cocodataset.org/val2017/000000039769.jpg''' __a = Image.open(requests.get(a__ , stream=a__ ).raw ) return im @torch.no_grad() def __lowerCAmelCase ( a__ , a__ ) -> List[str]: __a = ViTConfig() __a = False # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size if vit_name[-5:] == "in21k": __a = True __a = int(vit_name[-12:-10] ) __a = int(vit_name[-9:-6] ) else: __a = 1000 __a = '''huggingface/label-files''' __a = '''imagenet-1k-id2label.json''' __a = json.load(open(hf_hub_download(a__ , a__ , repo_type='''dataset''' ) , '''r''' ) ) __a = {int(a__ ): v for k, v in idalabel.items()} __a = idalabel __a = {v: k for k, v in idalabel.items()} __a = int(vit_name[-6:-4] ) __a = int(vit_name[-3:] ) # size of the architecture if "deit" in vit_name: if vit_name[9:].startswith('''tiny''' ): __a = 192 __a = 768 __a = 12 __a = 3 elif vit_name[9:].startswith('''small''' ): __a = 384 __a = 1536 __a = 12 __a = 6 else: pass else: if vit_name[4:].startswith('''small''' ): __a = 768 __a = 2304 __a = 8 __a = 8 elif vit_name[4:].startswith('''base''' ): pass elif vit_name[4:].startswith('''large''' ): __a = 1024 __a = 4096 __a = 24 __a = 16 elif vit_name[4:].startswith('''huge''' ): __a = 1280 __a = 5120 __a = 32 __a = 16 # load original model from timm __a = timm.create_model(a__ , pretrained=a__ ) timm_model.eval() # load state_dict of original model, remove and rename some keys __a = timm_model.state_dict() if base_model: remove_classification_head_(a__ ) __a = create_rename_keys(a__ , a__ ) for src, dest in rename_keys: rename_key(a__ , a__ , a__ ) read_in_q_k_v(a__ , a__ , a__ ) # load HuggingFace model if vit_name[-5:] == "in21k": __a = ViTModel(a__ ).eval() else: __a = ViTForImageClassification(a__ ).eval() model.load_state_dict(a__ ) # Check outputs on an image, prepared by ViTImageProcessor/DeiTImageProcessor if "deit" in vit_name: __a = DeiTImageProcessor(size=config.image_size ) else: __a = ViTImageProcessor(size=config.image_size ) __a = image_processor(images=prepare_img() , return_tensors='''pt''' ) __a = encoding['''pixel_values'''] __a = model(a__ ) if base_model: __a = timm_model.forward_features(a__ ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(a__ , outputs.pooler_output , atol=1e-3 ) else: __a = timm_model(a__ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(a__ , outputs.logits , atol=1e-3 ) Path(a__ ).mkdir(exist_ok=a__ ) print(F"""Saving model {vit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(a__ ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(a__ ) if __name__ == "__main__": A : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--vit_name', default='vit_base_patch16_224', type=str, help='Name of the ViT timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) A : Optional[Any] = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path)
219
# this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under git won't be considered # since the output of this script is fed into Makefile commands it doesn't print a newline after the results import re import subprocess import sys lowerCamelCase : List[Any] = subprocess.check_output("git merge-base main HEAD".split()).decode("utf-8") lowerCamelCase : Tuple = ( subprocess.check_output(F"""git diff --diff-filter=d --name-only {fork_point_sha}""".split()).decode("utf-8").split() ) lowerCamelCase : Tuple = "|".join(sys.argv[1:]) lowerCamelCase : Any = re.compile(rF"""^({joined_dirs}).*?\.py$""") lowerCamelCase : List[str] = [x for x in modified_files if regex.match(x)] print(" ".join(relevant_modified_files), end="")
70
0
'''simple docstring''' import numpy as np from cva import COLOR_BGR2GRAY, cvtColor, imread from numpy import array, uinta from PIL import Image from digital_image_processing import change_contrast as cc from digital_image_processing import convert_to_negative as cn from digital_image_processing import sepia as sp from digital_image_processing.dithering import burkes as bs from digital_image_processing.edge_detection import canny from digital_image_processing.filters import convolve as conv from digital_image_processing.filters import gaussian_filter as gg from digital_image_processing.filters import local_binary_pattern as lbp from digital_image_processing.filters import median_filter as med from digital_image_processing.filters import sobel_filter as sob from digital_image_processing.resize import resize as rs SCREAMING_SNAKE_CASE_ = imread(R'digital_image_processing/image_data/lena_small.jpg') SCREAMING_SNAKE_CASE_ = cvtColor(img, COLOR_BGR2GRAY) def UpperCamelCase__ ( ) -> List[str]: __UpperCAmelCase: List[str] = cn.convert_to_negative(_lowercase ) # assert negative_img array for at least one True assert negative_img.any() def UpperCamelCase__ ( ) -> str: with Image.open("""digital_image_processing/image_data/lena_small.jpg""" ) as img: # Work around assertion for response assert str(cc.change_contrast(_lowercase , 1_1_0 ) ).startswith( """<PIL.Image.Image image mode=RGB size=100x100 at""" ) def UpperCamelCase__ ( ) -> Any: __UpperCAmelCase: Optional[Any] = canny.gen_gaussian_kernel(9 , sigma=1.4 ) # Assert ambiguous array assert resp.all() def UpperCamelCase__ ( ) -> Dict: __UpperCAmelCase: Optional[Any] = imread("""digital_image_processing/image_data/lena_small.jpg""" , 0 ) # assert ambiguous array for all == True assert canny_img.all() __UpperCAmelCase: str = canny.canny(_lowercase ) # assert canny array for at least one True assert canny_array.any() def UpperCamelCase__ ( ) -> List[str]: assert gg.gaussian_filter(_lowercase , 5 , sigma=0.9 ).all() def UpperCamelCase__ ( ) -> Optional[int]: __UpperCAmelCase: str = array([[0.25, 0.5, 0.25], [0.5, -3, 0.5], [0.25, 0.5, 0.25]] ) __UpperCAmelCase: Tuple = conv.img_convolve(_lowercase , _lowercase ).astype(_lowercase ) assert res.any() def UpperCamelCase__ ( ) -> List[str]: assert med.median_filter(_lowercase , 3 ).any() def UpperCamelCase__ ( ) -> Union[str, Any]: __UpperCAmelCase, __UpperCAmelCase: Union[str, Any] = sob.sobel_filter(_lowercase ) assert grad.any() and theta.any() def UpperCamelCase__ ( ) -> str: __UpperCAmelCase: int = sp.make_sepia(_lowercase , 2_0 ) assert sepia.all() def UpperCamelCase__ ( _lowercase : str = "digital_image_processing/image_data/lena_small.jpg" ) -> List[str]: __UpperCAmelCase: str = bs.Burkes(imread(_lowercase , 1 ) , 1_2_0 ) burkes.process() assert burkes.output_img.any() def UpperCamelCase__ ( _lowercase : str = "digital_image_processing/image_data/lena_small.jpg" , ) -> str: __UpperCAmelCase: Tuple = rs.NearestNeighbour(imread(_lowercase , 1 ) , 4_0_0 , 2_0_0 ) nn.process() assert nn.output.any() def UpperCamelCase__ ( ) -> Optional[Any]: __UpperCAmelCase: Tuple = """digital_image_processing/image_data/lena.jpg""" # Reading the image and converting it to grayscale. __UpperCAmelCase: List[Any] = imread(_lowercase , 0 ) # Test for get_neighbors_pixel function() return not None __UpperCAmelCase: Optional[Any] = 0 __UpperCAmelCase: Dict = 0 __UpperCAmelCase: List[str] = image[x_coordinate][y_coordinate] __UpperCAmelCase: Optional[Any] = lbp.get_neighbors_pixel( _lowercase , _lowercase , _lowercase , _lowercase ) assert neighbors_pixels is not None # Test for local_binary_pattern function() # Create a numpy array as the same height and width of read image __UpperCAmelCase: List[Any] = np.zeros((image.shape[0], image.shape[1]) ) # Iterating through the image and calculating the local binary pattern value # for each pixel. for i in range(0 , image.shape[0] ): for j in range(0 , image.shape[1] ): __UpperCAmelCase: int = lbp.local_binary_value(_lowercase , _lowercase , _lowercase ) assert lbp_image.any()
523
import argparse import json import subprocess def _SCREAMING_SNAKE_CASE ( lowercase : Dict , lowercase : List[str] ): '''simple docstring''' lowerCamelCase_ = [] lowerCamelCase_ = ( f"""curl -H \"Accept: application/vnd.github+json\" -H \"Authorization: Bearer {token}\"""" ' https://api.github.com/repos/huggingface/transformers/actions/runners' ) lowerCamelCase_ = subprocess.run(lowercase , shell=lowercase , stdout=subprocess.PIPE ) lowerCamelCase_ = output.stdout.decode('utf-8' ) lowerCamelCase_ = json.loads(lowercase ) lowerCamelCase_ = status['runners'] for runner in runners: if runner["name"] in target_runners: if runner["status"] == "offline": offline_runners.append(lowercase ) # save the result so we can report them on Slack with open('offline_runners.txt' , 'w' ) as fp: fp.write(json.dumps(lowercase ) ) if len(lowercase ) > 0: lowerCamelCase_ = '\n'.join([x['name'] for x in offline_runners] ) raise ValueError(f"""The following runners are offline:\n{failed}""" ) if __name__ == "__main__": def _SCREAMING_SNAKE_CASE ( lowercase : List[str] ): '''simple docstring''' return values.split(',' ) lowerCamelCase : str = argparse.ArgumentParser() # Required parameters parser.add_argument( "--target_runners", default=None, type=list_str, required=True, help="Comma-separated list of runners to check status.", ) parser.add_argument( "--token", default=None, type=str, required=True, help="A token that has actions:read permission." ) lowerCamelCase : Optional[int] = parser.parse_args() get_runner_status(args.target_runners, args.token)
70
0
from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_herbert import HerbertTokenizer UpperCamelCase_ = logging.get_logger(__name__) UpperCamelCase_ = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} UpperCamelCase_ = { "vocab_file": { "allegro/herbert-base-cased": "https://huggingface.co/allegro/herbert-base-cased/resolve/main/vocab.json" }, "merges_file": { "allegro/herbert-base-cased": "https://huggingface.co/allegro/herbert-base-cased/resolve/main/merges.txt" }, } UpperCamelCase_ = {"allegro/herbert-base-cased": 5_14} UpperCamelCase_ = {} class _SCREAMING_SNAKE_CASE ( snake_case ): lowerCamelCase_ = VOCAB_FILES_NAMES lowerCamelCase_ = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase_ = PRETRAINED_INIT_CONFIGURATION lowerCamelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase_ = HerbertTokenizer def __init__( self : Tuple , snake_case_ : Optional[int]=None , snake_case_ : str=None , snake_case_ : Any=None , snake_case_ : Optional[int]="<s>" , snake_case_ : Union[str, Any]="<unk>" , snake_case_ : Any="<pad>" , snake_case_ : Tuple="<mask>" , snake_case_ : Optional[Any]="</s>" , **snake_case_ : int , ): """simple docstring""" super().__init__( A_ , A_ , tokenizer_file=A_ , cls_token=A_ , unk_token=A_ , pad_token=A_ , mask_token=A_ , sep_token=A_ , **A_ , ) def _UpperCAmelCase ( self : str , snake_case_ : List[int] , snake_case_ : Optional[List[int]] = None ): """simple docstring""" A : List[str] = [self.cls_token_id] A : List[Any] = [self.sep_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def _UpperCAmelCase ( self : Union[str, Any] , snake_case_ : List[int] , snake_case_ : Optional[List[int]] = None , snake_case_ : bool = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=A_ , token_ids_a=A_ , already_has_special_tokens=A_ ) if token_ids_a is None: return [1] + ([0] * len(A_ )) + [1] return [1] + ([0] * len(A_ )) + [1] + ([0] * len(A_ )) + [1] def _UpperCAmelCase ( self : Optional[Any] , snake_case_ : List[int] , snake_case_ : Optional[List[int]] = None ): """simple docstring""" A : Optional[Any] = [self.sep_token_id] A : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _UpperCAmelCase ( self : Any , snake_case_ : str , snake_case_ : Optional[str] = None ): """simple docstring""" A : Tuple = self._tokenizer.model.save(A_ , name=A_ ) return tuple(A_ )
256
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from timm import create_model from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import BitConfig, BitForImageClassification, BitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase : Optional[Any] = logging.get_logger(__name__) def _SCREAMING_SNAKE_CASE ( lowercase : List[Any] ): '''simple docstring''' lowerCamelCase_ = 'huggingface/label-files' lowerCamelCase_ = 'imagenet-1k-id2label.json' lowerCamelCase_ = json.load(open(hf_hub_download(lowercase , lowercase , repo_type='dataset' ) , 'r' ) ) lowerCamelCase_ = {int(lowercase ): v for k, v in idalabel.items()} lowerCamelCase_ = {v: k for k, v in idalabel.items()} lowerCamelCase_ = 'std_conv' if 'bit' in model_name else False # note that when using BiT as backbone for ViT-hybrid checkpoints, # one needs to additionally set config.layer_type = "bottleneck", config.stem_type = "same", # config.conv_layer = "std_conv_same" lowerCamelCase_ = BitConfig( conv_layer=lowercase , num_labels=10_00 , idalabel=lowercase , labelaid=lowercase , ) return config def _SCREAMING_SNAKE_CASE ( lowercase : Any ): '''simple docstring''' if "stem.conv" in name: lowerCamelCase_ = name.replace('stem.conv' , 'bit.embedder.convolution' ) if "blocks" in name: lowerCamelCase_ = name.replace('blocks' , 'layers' ) if "head.fc" in name: lowerCamelCase_ = name.replace('head.fc' , 'classifier.1' ) if name.startswith('norm' ): lowerCamelCase_ = 'bit.' + name if "bit" not in name and "classifier" not in name: lowerCamelCase_ = 'bit.encoder.' + name return name def _SCREAMING_SNAKE_CASE ( ): '''simple docstring''' lowerCamelCase_ = 'http://images.cocodataset.org/val2017/000000039769.jpg' lowerCamelCase_ = Image.open(requests.get(lowercase , stream=lowercase ).raw ) return im @torch.no_grad() def _SCREAMING_SNAKE_CASE ( lowercase : str , lowercase : int , lowercase : Union[str, Any]=False ): '''simple docstring''' lowerCamelCase_ = get_config(lowercase ) # load original model from timm lowerCamelCase_ = create_model(lowercase , pretrained=lowercase ) timm_model.eval() # load state_dict of original model lowerCamelCase_ = timm_model.state_dict() for key in state_dict.copy().keys(): lowerCamelCase_ = state_dict.pop(lowercase ) lowerCamelCase_ = val.squeeze() if 'head' in key else val # load HuggingFace model lowerCamelCase_ = BitForImageClassification(lowercase ) model.eval() model.load_state_dict(lowercase ) # create image processor lowerCamelCase_ = create_transform(**resolve_data_config({} , model=lowercase ) ) lowerCamelCase_ = transform.transforms lowerCamelCase_ = { 'bilinear': PILImageResampling.BILINEAR, 'bicubic': PILImageResampling.BICUBIC, 'nearest': PILImageResampling.NEAREST, } lowerCamelCase_ = BitImageProcessor( do_resize=lowercase , size={'shortest_edge': timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=lowercase , crop_size={'height': timm_transforms[1].size[0], 'width': timm_transforms[1].size[1]} , do_normalize=lowercase , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) lowerCamelCase_ = prepare_img() lowerCamelCase_ = transform(lowercase ).unsqueeze(0 ) lowerCamelCase_ = processor(lowercase , return_tensors='pt' ).pixel_values # verify pixel values assert torch.allclose(lowercase , lowercase ) # verify logits with torch.no_grad(): lowerCamelCase_ = model(lowercase ) lowerCamelCase_ = outputs.logits print('Logits:' , logits[0, :3] ) print('Predicted class:' , model.config.idalabel[logits.argmax(-1 ).item()] ) lowerCamelCase_ = timm_model(lowercase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(lowercase , outputs.logits , atol=1e-3 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: Path(lowercase ).mkdir(exist_ok=lowercase ) print(f"""Saving model {model_name} and processor to {pytorch_dump_folder_path}""" ) model.save_pretrained(lowercase ) processor.save_pretrained(lowercase ) if push_to_hub: print(f"""Pushing model {model_name} and processor to the hub""" ) model.push_to_hub(f"""ybelkada/{model_name}""" ) processor.push_to_hub(f"""ybelkada/{model_name}""" ) if __name__ == "__main__": lowerCamelCase : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="resnetv2_50x1_bitm", type=str, help="Name of the BiT timm model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to push the model to the hub.", ) lowerCamelCase : Optional[int] = parser.parse_args() convert_bit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
70
0
"""simple docstring""" def lowerCAmelCase_ ( lowercase_ : str ): '''simple docstring''' __SCREAMING_SNAKE_CASE : List[str] = hex_num.strip() if not hex_num: raise ValueError('''No value was passed to the function''' ) __SCREAMING_SNAKE_CASE : Tuple = hex_num[0] == '''-''' if is_negative: __SCREAMING_SNAKE_CASE : Tuple = hex_num[1:] try: __SCREAMING_SNAKE_CASE : int = int(lowercase_ , 16 ) except ValueError: raise ValueError('''Invalid value was passed to the function''' ) __SCREAMING_SNAKE_CASE : List[str] = '''''' while int_num > 0: __SCREAMING_SNAKE_CASE : List[str] = str(int_num % 2 ) + bin_str int_num >>= 1 return int(('''-''' + bin_str) if is_negative else bin_str ) if __name__ == "__main__": import doctest doctest.testmod()
674
from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class A: '''simple docstring''' def __init__( self : List[Any] , A_ : Dict , A_ : Union[str, Any]=13 , A_ : List[Any]=30 , A_ : Optional[Any]=2 , A_ : List[str]=3 , A_ : List[str]=True , A_ : Dict=True , A_ : List[Any]=32 , A_ : Any=2 , A_ : Any=4 , A_ : Optional[int]=37 , A_ : Dict="gelu" , A_ : List[Any]=0.1 , A_ : Optional[int]=0.1 , A_ : Union[str, Any]=10 , A_ : Optional[Any]=0.02 , A_ : List[Any]=3 , A_ : str=None , ) -> str: """simple docstring""" lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = image_size lowerCamelCase_ = patch_size lowerCamelCase_ = num_channels lowerCamelCase_ = is_training lowerCamelCase_ = use_labels lowerCamelCase_ = hidden_size lowerCamelCase_ = num_hidden_layers lowerCamelCase_ = num_attention_heads lowerCamelCase_ = intermediate_size lowerCamelCase_ = hidden_act lowerCamelCase_ = hidden_dropout_prob lowerCamelCase_ = attention_probs_dropout_prob lowerCamelCase_ = type_sequence_label_size lowerCamelCase_ = initializer_range lowerCamelCase_ = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) lowerCamelCase_ = (image_size // patch_size) ** 2 lowerCamelCase_ = num_patches + 1 def a__ ( self : List[str] ) -> Dict: """simple docstring""" lowerCamelCase_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase_ = None if self.use_labels: lowerCamelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase_ = self.get_config() return config, pixel_values, labels def a__ ( self : List[Any] ) -> Any: """simple docstring""" return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=A_ , initializer_range=self.initializer_range , ) def a__ ( self : Any , A_ : int , A_ : int , A_ : int ) -> List[Any]: """simple docstring""" lowerCamelCase_ = TFViTModel(config=A_ ) lowerCamelCase_ = model(A_ , training=A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # Test with an image with different size than the one specified in config. lowerCamelCase_ = self.image_size // 2 lowerCamelCase_ = pixel_values[:, :, :image_size, :image_size] lowerCamelCase_ = model(A_ , interpolate_pos_encoding=A_ , training=A_ ) lowerCamelCase_ = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, seq_length, self.hidden_size) ) def a__ ( self : List[Any] , A_ : List[Any] , A_ : Any , A_ : Any ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.type_sequence_label_size lowerCamelCase_ = TFViTForImageClassification(A_ ) lowerCamelCase_ = model(A_ , labels=A_ , training=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # Test with an image with different size than the one specified in config. lowerCamelCase_ = self.image_size // 2 lowerCamelCase_ = pixel_values[:, :, :image_size, :image_size] lowerCamelCase_ = model(A_ , interpolate_pos_encoding=A_ , training=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowerCamelCase_ = 1 lowerCamelCase_ = TFViTForImageClassification(A_ ) lowerCamelCase_ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCamelCase_ = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def a__ ( self : List[Any] ) -> Any: """simple docstring""" lowerCamelCase_ = self.prepare_config_and_inputs() lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = config_and_inputs lowerCamelCase_ = {'pixel_values': pixel_values} return config, inputs_dict @require_tf class A( UpperCamelCase , UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () UpperCamelCase = ( {'''feature-extraction''': TFViTModel, '''image-classification''': TFViTForImageClassification} if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def a__ ( self : int ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = TFViTModelTester(self ) lowerCamelCase_ = ConfigTester(self , config_class=A_ , has_text_modality=A_ , hidden_size=37 ) def a__ ( self : int ) -> List[str]: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='ViT does not use inputs_embeds' ) def a__ ( self : List[str] ) -> Tuple: """simple docstring""" pass @unittest.skip(reason='ViT does not use inputs_embeds' ) def a__ ( self : int ) -> Optional[Any]: """simple docstring""" pass def a__ ( self : str ) -> List[str]: """simple docstring""" lowerCamelCase_ , lowerCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase_ = model_class(A_ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) lowerCamelCase_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(A_ , tf.keras.layers.Layer ) ) def a__ ( self : List[Any] ) -> Dict: """simple docstring""" lowerCamelCase_ , lowerCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase_ = model_class(A_ ) lowerCamelCase_ = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase_ = [*signature.parameters.keys()] lowerCamelCase_ = ['pixel_values'] self.assertListEqual(arg_names[:1] , A_ ) def a__ ( self : Tuple ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def a__ ( self : Tuple ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A_ ) @slow def a__ ( self : Any ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = TFViTModel.from_pretrained('google/vit-base-patch16-224' ) self.assertIsNotNone(A_ ) def _SCREAMING_SNAKE_CASE ( ): '''simple docstring''' lowerCamelCase_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class A( unittest.TestCase ): '''simple docstring''' @cached_property def a__ ( self : int ) -> Tuple: """simple docstring""" return ViTImageProcessor.from_pretrained('google/vit-base-patch16-224' ) if is_vision_available() else None @slow def a__ ( self : List[str] ) -> str: """simple docstring""" lowerCamelCase_ = TFViTForImageClassification.from_pretrained('google/vit-base-patch16-224' ) lowerCamelCase_ = self.default_image_processor lowerCamelCase_ = prepare_img() lowerCamelCase_ = image_processor(images=A_ , return_tensors='tf' ) # forward pass lowerCamelCase_ = model(**A_ ) # verify the logits lowerCamelCase_ = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , A_ ) lowerCamelCase_ = tf.constant([-0.2744, 0.8215, -0.0836] ) tf.debugging.assert_near(outputs.logits[0, :3] , A_ , atol=1E-4 )
70
0
'''simple docstring''' import os import sys import tempfile import torch from .state import AcceleratorState from .utils import PrecisionType, PrepareForLaunch, is_mps_available, patch_environment def SCREAMING_SNAKE_CASE_ ( __A : List[str] , __A : List[Any]=() , __A : Any=None , __A : str="no" , __A : int="29500" ) -> int: _SCREAMING_SNAKE_CASE = False _SCREAMING_SNAKE_CASE = False if any(key.startswith("KAGGLE" ) for key in os.environ.keys() ): _SCREAMING_SNAKE_CASE = True elif "IPython" in sys.modules: _SCREAMING_SNAKE_CASE = "google.colab" in str(sys.modules["IPython"].get_ipython() ) try: _SCREAMING_SNAKE_CASE = PrecisionType(mixed_precision.lower() ) except ValueError: raise ValueError( f"""Unknown mixed_precision mode: {args.mixed_precision.lower()}. Choose between {PrecisionType.list()}.""" ) if (in_colab or in_kaggle) and (os.environ.get("TPU_NAME" , __A ) is not None): # TPU launch import torch_xla.distributed.xla_multiprocessing as xmp if len(AcceleratorState._shared_state ) > 0: raise ValueError( "To train on TPU in Colab or Kaggle Kernel, the `Accelerator` should only be initialized inside " "your training function. Restart your notebook and make sure no cells initializes an " "`Accelerator`." ) if num_processes is None: _SCREAMING_SNAKE_CASE = 8 _SCREAMING_SNAKE_CASE = PrepareForLaunch(__A , distributed_type="TPU" ) print(f"""Launching a training on {num_processes} TPU cores.""" ) xmp.spawn(__A , args=__A , nprocs=__A , start_method="fork" ) elif in_colab: # No need for a distributed launch otherwise as it's either CPU or one GPU. if torch.cuda.is_available(): print("Launching training on one GPU." ) else: print("Launching training on one CPU." ) function(*__A ) else: if num_processes is None: raise ValueError( "You have to specify the number of GPUs you would like to use, add `num_processes=...` to your call." ) if num_processes > 1: # Multi-GPU launch from torch.multiprocessing import start_processes from torch.multiprocessing.spawn import ProcessRaisedException if len(AcceleratorState._shared_state ) > 0: raise ValueError( "To launch a multi-GPU training from your notebook, the `Accelerator` should only be initialized " "inside your training function. Restart your notebook and make sure no cells initializes an " "`Accelerator`." ) if torch.cuda.is_initialized(): raise ValueError( "To launch a multi-GPU training from your notebook, you need to avoid running any instruction " "using `torch.cuda` in any cell. Restart your notebook and make sure no cells use any CUDA " "function." ) # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=__A , master_addr="127.0.01" , master_port=__A , mixed_precision=__A ): _SCREAMING_SNAKE_CASE = PrepareForLaunch(__A , distributed_type="MULTI_GPU" ) print(f"""Launching training on {num_processes} GPUs.""" ) try: start_processes(__A , args=__A , nprocs=__A , start_method="fork" ) except ProcessRaisedException as e: if "Cannot re-initialize CUDA in forked subprocess" in e.args[0]: raise RuntimeError( "CUDA has been initialized before the `notebook_launcher` could create a forked subprocess. " "This likely stems from an outside import causing issues once the `notebook_launcher()` is called. " "Please review your imports and test them when running the `notebook_launcher()` to identify " "which one is problematic." ) from e else: # No need for a distributed launch otherwise as it's either CPU, GPU or MPS. if is_mps_available(): _SCREAMING_SNAKE_CASE = "1" print("Launching training on MPS." ) elif torch.cuda.is_available(): print("Launching training on one GPU." ) else: print("Launching training on CPU." ) function(*__A ) def SCREAMING_SNAKE_CASE_ ( __A : List[str] , __A : Tuple=() , __A : Any=2 ) -> Optional[Any]: from torch.multiprocessing import start_processes with tempfile.NamedTemporaryFile() as tmp_file: # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=__A , master_addr="127.0.01" , master_port="29500" , accelerate_mixed_precision="no" , accelerate_debug_rdv_file=tmp_file.name , accelerate_use_cpu="yes" , ): _SCREAMING_SNAKE_CASE = PrepareForLaunch(__A , debug=__A ) start_processes(__A , args=__A , nprocs=__A , start_method="fork" )
418
import itertools import random import unittest import numpy as np from transformers import is_speech_available from transformers.testing_utils import require_torch, require_torchaudio from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import SpeechaTextFeatureExtractor lowerCamelCase : Any = random.Random() def _SCREAMING_SNAKE_CASE ( lowercase : Union[str, Any] , lowercase : int=1.0 , lowercase : List[str]=None , lowercase : str=None ): '''simple docstring''' if rng is None: lowerCamelCase_ = global_rng lowerCamelCase_ = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class A( unittest.TestCase ): '''simple docstring''' def __init__( self : List[Any] , A_ : Dict , A_ : int=7 , A_ : str=400 , A_ : Dict=2000 , A_ : List[Any]=24 , A_ : List[Any]=24 , A_ : int=0.0 , A_ : Dict=16000 , A_ : List[Any]=True , A_ : str=True , ) -> Dict: """simple docstring""" lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = min_seq_length lowerCamelCase_ = max_seq_length lowerCamelCase_ = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) lowerCamelCase_ = feature_size lowerCamelCase_ = num_mel_bins lowerCamelCase_ = padding_value lowerCamelCase_ = sampling_rate lowerCamelCase_ = return_attention_mask lowerCamelCase_ = do_normalize def a__ ( self : List[str] ) -> Optional[int]: """simple docstring""" return { "feature_size": self.feature_size, "num_mel_bins": self.num_mel_bins, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def a__ ( self : List[Any] , A_ : str=False , A_ : Union[str, Any]=False ) -> str: """simple docstring""" def _flatten(A_ : List[Any] ): return list(itertools.chain(*A_ ) ) if equal_length: lowerCamelCase_ = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size lowerCamelCase_ = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: lowerCamelCase_ = [np.asarray(A_ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class A( UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = SpeechaTextFeatureExtractor if is_speech_available() else None def a__ ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = SpeechaTextFeatureExtractionTester(self ) def a__ ( self : str , A_ : Dict ) -> Dict: """simple docstring""" self.assertTrue(np.all(np.mean(A_ , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(A_ , axis=0 ) - 1 ) < 1E-3 ) ) def a__ ( self : int ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = [np.asarray(A_ ) for speech_input in speech_inputs] # Test feature size lowerCamelCase_ = feature_extractor(A_ , padding=A_ , return_tensors='np' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.feature_size ) # Test not batched input lowerCamelCase_ = feature_extractor(speech_inputs[0] , return_tensors='np' ).input_features lowerCamelCase_ = feature_extractor(np_speech_inputs[0] , return_tensors='np' ).input_features self.assertTrue(np.allclose(A_ , A_ , atol=1E-3 ) ) # Test batched lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. lowerCamelCase_ = [floats_list((1, x) )[0] for x in (800, 800, 800)] lowerCamelCase_ = np.asarray(A_ ) lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1E-3 ) ) def a__ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = ['longest', 'max_length', 'do_not_pad'] lowerCamelCase_ = [None, 16, None] for max_length, padding in zip(A_ , A_ ): lowerCamelCase_ = feature_extractor( A_ , padding=A_ , max_length=A_ , return_attention_mask=A_ ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = [np.sum(A_ ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def a__ ( self : List[Any] ) -> Any: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = ['longest', 'max_length', 'do_not_pad'] lowerCamelCase_ = [None, 16, None] for max_length, padding in zip(A_ , A_ ): lowerCamelCase_ = feature_extractor( A_ , max_length=A_ , padding=A_ , return_tensors='np' , return_attention_mask=A_ ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = [np.sum(A_ ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self.assertTrue(input_features[0][fbank_feat_lengths[0] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self.assertTrue(input_features[0][fbank_feat_lengths[1] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def a__ ( self : List[str] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = feature_extractor( A_ , padding='max_length' , max_length=4 , truncation=A_ , return_tensors='np' , return_attention_mask=A_ , ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1] ) self._check_zero_mean_unit_variance(input_features[2] ) def a__ ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = feature_extractor( A_ , padding='longest' , max_length=4 , truncation=A_ , return_tensors='np' , return_attention_mask=A_ , ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 4, 24) ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = feature_extractor( A_ , padding='longest' , max_length=16 , truncation=A_ , return_tensors='np' , return_attention_mask=A_ , ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 6, 24) ) def a__ ( self : List[Any] ) -> List[str]: """simple docstring""" import torch lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = np.random.rand(100 , 32 ).astype(np.floataa ) lowerCamelCase_ = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: lowerCamelCase_ = feature_extractor.pad([{'input_features': inputs}] , return_tensors='np' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) lowerCamelCase_ = feature_extractor.pad([{'input_features': inputs}] , return_tensors='pt' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def a__ ( self : List[str] , A_ : Union[str, Any] ) -> List[Any]: """simple docstring""" from datasets import load_dataset lowerCamelCase_ = load_dataset('hf-internal-testing/librispeech_asr_dummy' , 'clean' , split='validation' ) # automatic decoding with librispeech lowerCamelCase_ = ds.sort('id' ).select(range(A_ ) )[:num_samples]['audio'] return [x["array"] for x in speech_samples] def a__ ( self : str ) -> Tuple: """simple docstring""" lowerCamelCase_ = np.array([ -1.5745, -1.7713, -1.7020, -1.6069, -1.2250, -1.1105, -0.9072, -0.8241, -1.2310, -0.8098, -0.3320, -0.4101, -0.7985, -0.4996, -0.8213, -0.9128, -1.0420, -1.1286, -1.0440, -0.7999, -0.8405, -1.2275, -1.5443, -1.4625, ] ) # fmt: on lowerCamelCase_ = self._load_datasamples(1 ) lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = feature_extractor(A_ , return_tensors='pt' ).input_features self.assertEquals(input_features.shape , (1, 584, 24) ) self.assertTrue(np.allclose(input_features[0, 0, :30] , A_ , atol=1E-4 ) )
70
0
'''simple docstring''' import argparse from copy import deepcopy import numpy as np from datasets import ClassLabel, DatasetDict, load_dataset from evaluate import load from transformers import ( AutoModelForSequenceClassification, AutoTokenizer, DataCollatorWithPadding, Trainer, TrainerCallback, TrainingArguments, set_seed, ) def _A ( ): '''simple docstring''' A__ = argparse.ArgumentParser() parser.add_argument('--model_ckpt' ,type=UpperCAmelCase ,default='microsoft/unixcoder-base-nine' ) parser.add_argument('--num_epochs' ,type=UpperCAmelCase ,default=5 ) parser.add_argument('--batch_size' ,type=UpperCAmelCase ,default=6 ) parser.add_argument('--gradient_accumulation_steps' ,type=UpperCAmelCase ,default=1 ) parser.add_argument('--freeze' ,type=UpperCAmelCase ,default=UpperCAmelCase ) parser.add_argument('--learning_rate' ,type=UpperCAmelCase ,default=5e-4 ) parser.add_argument('--seed' ,type=UpperCAmelCase ,default=0 ) parser.add_argument('--lr_scheduler_type' ,type=UpperCAmelCase ,default='cosine' ) parser.add_argument('--num_warmup_steps' ,type=UpperCAmelCase ,default=10 ) parser.add_argument('--weight_decay' ,type=UpperCAmelCase ,default=0.01 ) parser.add_argument('--output_dir' ,type=UpperCAmelCase ,default='./results' ) return parser.parse_args() lowerCAmelCase_ = load('''accuracy''') def _A ( UpperCAmelCase ): '''simple docstring''' A__ , A__ = eval_pred A__ = np.argmax(UpperCAmelCase ,axis=1 ) return metric.compute(predictions=UpperCAmelCase ,references=UpperCAmelCase ) class _snake_case( UpperCAmelCase ): def __init__(self : Union[str, Any] , a : Optional[Any] ) -> None: """simple docstring""" super().__init__() A__ = trainer def _UpperCamelCase (self : Any , a : Optional[Any] , a : Tuple , a : Union[str, Any] , **a : Tuple ) -> List[Any]: """simple docstring""" if control.should_evaluate: A__ = deepcopy(A_ ) self._trainer.evaluate(eval_dataset=self._trainer.train_dataset , metric_key_prefix='train' ) return control_copy def _A ( ): '''simple docstring''' A__ = get_args() set_seed(args.seed ) A__ = load_dataset('codeparrot/codecomplex' ,split='train' ) A__ = dataset.train_test_split(test_size=0.2 ) A__ = train_test['test'].train_test_split(test_size=0.5 ) A__ = DatasetDict( { 'train': train_test['train'], 'test': test_validation['train'], 'valid': test_validation['test'], } ) print('Loading tokenizer and model' ) A__ = AutoTokenizer.from_pretrained(args.model_ckpt ) A__ = tokenizer.eos_token A__ = AutoModelForSequenceClassification.from_pretrained(args.model_ckpt ,num_labels=7 ) A__ = model.config.eos_token_id if args.freeze: for param in model.roberta.parameters(): A__ = False A__ = ClassLabel(num_classes=7 ,names=list(set(train_test_validation['train']['complexity'] ) ) ) def tokenize(UpperCAmelCase ): A__ = tokenizer(example['src'] ,truncation=UpperCAmelCase ,max_length=1024 ) A__ = labels.straint(example['complexity'] ) return { "input_ids": inputs["input_ids"], "attention_mask": inputs["attention_mask"], "label": label, } A__ = train_test_validation.map( UpperCAmelCase ,batched=UpperCAmelCase ,remove_columns=train_test_validation['train'].column_names ,) A__ = DataCollatorWithPadding(tokenizer=UpperCAmelCase ) A__ = TrainingArguments( output_dir=args.output_dir ,learning_rate=args.learning_rate ,lr_scheduler_type=args.lr_scheduler_type ,evaluation_strategy='epoch' ,save_strategy='epoch' ,logging_strategy='epoch' ,per_device_train_batch_size=args.batch_size ,per_device_eval_batch_size=args.batch_size ,num_train_epochs=args.num_epochs ,gradient_accumulation_steps=args.gradient_accumulation_steps ,weight_decay=0.01 ,metric_for_best_model='accuracy' ,run_name='complexity-java' ,report_to='wandb' ,) A__ = Trainer( model=UpperCAmelCase ,args=UpperCAmelCase ,train_dataset=tokenized_datasets['train'] ,eval_dataset=tokenized_datasets['valid'] ,tokenizer=UpperCAmelCase ,data_collator=UpperCAmelCase ,compute_metrics=UpperCAmelCase ,) print('Training...' ) trainer.add_callback(CustomCallback(UpperCAmelCase ) ) trainer.train() if __name__ == "__main__": main()
531
import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class A( UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = TransfoXLTokenizer UpperCamelCase = False UpperCamelCase = False def a__ ( self : Optional[Any] ) -> int: """simple docstring""" super().setUp() lowerCamelCase_ = [ '<unk>', '[CLS]', '[SEP]', 'want', 'unwanted', 'wa', 'un', 'running', ',', 'low', 'l', ] lowerCamelCase_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def a__ ( self : Optional[Any] , **A_ : Tuple ) -> Any: """simple docstring""" lowerCamelCase_ = True return TransfoXLTokenizer.from_pretrained(self.tmpdirname , **A_ ) def a__ ( self : List[str] , A_ : Dict ) -> Any: """simple docstring""" lowerCamelCase_ = '<unk> UNwanted , running' lowerCamelCase_ = '<unk> unwanted, running' return input_text, output_text def a__ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = TransfoXLTokenizer(vocab_file=self.vocab_file , lower_case=A_ ) lowerCamelCase_ = tokenizer.tokenize('<unk> UNwanted , running' ) self.assertListEqual(A_ , ['<unk>', 'unwanted', ',', 'running'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A_ ) , [0, 4, 8, 7] ) def a__ ( self : Any ) -> str: """simple docstring""" lowerCamelCase_ = TransfoXLTokenizer(lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['hello', '!', 'how', 'are', 'you', '?'] ) def a__ ( self : int ) -> Dict: """simple docstring""" lowerCamelCase_ = TransfoXLTokenizer(lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?'] ) def a__ ( self : List[Any] ) -> Tuple: """simple docstring""" lowerCamelCase_ = TransfoXLTokenizer(lower_case=A_ ) lowerCamelCase_ = 'Hello (bracket) and side-scrolled [and] Henry\'s $5,000 with 3.34 m. What\'s up!?' lowerCamelCase_ = [ 'Hello', '(', 'bracket', ')', 'and', 'side', '@-@', 'scrolled', '[', 'and', ']', 'Henry', '\'s', '$', '5', '@,@', '000', 'with', '3', '@.@', '34', 'm', '.', 'What', '\'s', 'up', '!', '?', ] self.assertListEqual(tokenizer.tokenize(A_ ) , A_ ) self.assertEqual(tokenizer.convert_tokens_to_string(A_ ) , A_ ) def a__ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = len(A_ ) tokenizer.add_tokens(['new1', 'new2'] ) tokenizer.move_added_token('new1' , 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(A_ ) , original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode('new1' ) , [1] ) self.assertEqual(tokenizer.decode([1] ) , 'new1' )
70
0
'''simple docstring''' from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy import tensorflow as tf from transformers import ( TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, TF_DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, TF_DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, BertConfig, DPRConfig, TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, ) class A__ : def __init__( self : Dict , _a : Optional[Any] , _a : Optional[int]=13 , _a : List[Any]=7 , _a : Dict=True , _a : Any=True , _a : int=True , _a : int=True , _a : Any=99 , _a : str=32 , _a : str=2 , _a : Optional[Any]=4 , _a : Optional[int]=37 , _a : Union[str, Any]="gelu" , _a : Dict=0.1 , _a : Optional[Any]=0.1 , _a : Optional[Any]=512 , _a : List[str]=16 , _a : Optional[int]=2 , _a : Optional[Any]=0.02 , _a : List[Any]=3 , _a : List[Any]=4 , _a : Optional[Any]=None , _a : Any=0 , ) -> Dict: '''simple docstring''' _SCREAMING_SNAKE_CASE =parent _SCREAMING_SNAKE_CASE =batch_size _SCREAMING_SNAKE_CASE =seq_length _SCREAMING_SNAKE_CASE =is_training _SCREAMING_SNAKE_CASE =use_input_mask _SCREAMING_SNAKE_CASE =use_token_type_ids _SCREAMING_SNAKE_CASE =use_labels _SCREAMING_SNAKE_CASE =vocab_size _SCREAMING_SNAKE_CASE =hidden_size _SCREAMING_SNAKE_CASE =num_hidden_layers _SCREAMING_SNAKE_CASE =num_attention_heads _SCREAMING_SNAKE_CASE =intermediate_size _SCREAMING_SNAKE_CASE =hidden_act _SCREAMING_SNAKE_CASE =hidden_dropout_prob _SCREAMING_SNAKE_CASE =attention_probs_dropout_prob _SCREAMING_SNAKE_CASE =max_position_embeddings _SCREAMING_SNAKE_CASE =type_vocab_size _SCREAMING_SNAKE_CASE =type_sequence_label_size _SCREAMING_SNAKE_CASE =initializer_range _SCREAMING_SNAKE_CASE =num_labels _SCREAMING_SNAKE_CASE =num_choices _SCREAMING_SNAKE_CASE =scope _SCREAMING_SNAKE_CASE =projection_dim def A ( self : Any ) -> int: '''simple docstring''' _SCREAMING_SNAKE_CASE =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _SCREAMING_SNAKE_CASE =None if self.use_input_mask: # follow test_modeling_tf_ctrl.py _SCREAMING_SNAKE_CASE =random_attention_mask([self.batch_size, self.seq_length] ) _SCREAMING_SNAKE_CASE =None if self.use_token_type_ids: _SCREAMING_SNAKE_CASE =ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _SCREAMING_SNAKE_CASE =None _SCREAMING_SNAKE_CASE =None _SCREAMING_SNAKE_CASE =None if self.use_labels: _SCREAMING_SNAKE_CASE =ids_tensor([self.batch_size] , self.type_sequence_label_size ) _SCREAMING_SNAKE_CASE =ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _SCREAMING_SNAKE_CASE =ids_tensor([self.batch_size] , self.num_choices ) _SCREAMING_SNAKE_CASE =BertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=A_ , initializer_range=self.initializer_range , ) _SCREAMING_SNAKE_CASE =DPRConfig(projection_dim=self.projection_dim , **config.to_dict() ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : Dict , _a : Optional[int] , _a : Optional[Any] , _a : Optional[int] , _a : Optional[Any] , _a : Any , _a : Any , _a : Optional[int] ) -> str: '''simple docstring''' _SCREAMING_SNAKE_CASE =TFDPRContextEncoder(config=A_ ) _SCREAMING_SNAKE_CASE =model(A_ , attention_mask=A_ , token_type_ids=A_ ) _SCREAMING_SNAKE_CASE =model(A_ , token_type_ids=A_ ) _SCREAMING_SNAKE_CASE =model(A_ ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.projection_dim or self.hidden_size) ) def A ( self : Any , _a : List[str] , _a : int , _a : List[str] , _a : List[Any] , _a : str , _a : Tuple , _a : List[Any] ) -> Dict: '''simple docstring''' _SCREAMING_SNAKE_CASE =TFDPRQuestionEncoder(config=A_ ) _SCREAMING_SNAKE_CASE =model(A_ , attention_mask=A_ , token_type_ids=A_ ) _SCREAMING_SNAKE_CASE =model(A_ , token_type_ids=A_ ) _SCREAMING_SNAKE_CASE =model(A_ ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.projection_dim or self.hidden_size) ) def A ( self : Optional[Any] , _a : List[str] , _a : List[Any] , _a : Optional[int] , _a : Optional[int] , _a : str , _a : Union[str, Any] , _a : int ) -> Tuple: '''simple docstring''' _SCREAMING_SNAKE_CASE =TFDPRReader(config=A_ ) _SCREAMING_SNAKE_CASE =model(A_ , attention_mask=A_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.relevance_logits.shape , (self.batch_size,) ) def A ( self : str ) -> Tuple: '''simple docstring''' _SCREAMING_SNAKE_CASE =self.prepare_config_and_inputs() ( ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ) =config_and_inputs _SCREAMING_SNAKE_CASE ={'input_ids': input_ids} return config, inputs_dict @require_tf class A__ ( A__ , A__ , unittest.TestCase ): A__ = ( ( TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, ) if is_tf_available() else () ) A__ = {'feature-extraction': TFDPRQuestionEncoder} if is_tf_available() else {} A__ = False A__ = False A__ = False A__ = False A__ = False def A ( self : Dict ) -> List[Any]: '''simple docstring''' _SCREAMING_SNAKE_CASE =TFDPRModelTester(self ) _SCREAMING_SNAKE_CASE =ConfigTester(self , config_class=A_ , hidden_size=37 ) def A ( self : Union[str, Any] ) -> Tuple: '''simple docstring''' self.config_tester.run_common_tests() def A ( self : int ) -> List[str]: '''simple docstring''' _SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_context_encoder(*A_ ) def A ( self : Optional[int] ) -> Dict: '''simple docstring''' _SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_question_encoder(*A_ ) def A ( self : Union[str, Any] ) -> str: '''simple docstring''' _SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_reader(*A_ ) @slow def A ( self : str ) -> Union[str, Any]: '''simple docstring''' for model_name in TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _SCREAMING_SNAKE_CASE =TFDPRContextEncoder.from_pretrained(A_ ) self.assertIsNotNone(A_ ) for model_name in TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _SCREAMING_SNAKE_CASE =TFDPRContextEncoder.from_pretrained(A_ ) self.assertIsNotNone(A_ ) for model_name in TF_DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _SCREAMING_SNAKE_CASE =TFDPRQuestionEncoder.from_pretrained(A_ ) self.assertIsNotNone(A_ ) for model_name in TF_DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _SCREAMING_SNAKE_CASE =TFDPRReader.from_pretrained(A_ ) self.assertIsNotNone(A_ ) @require_tf class A__ ( unittest.TestCase ): @slow def A ( self : int ) -> int: '''simple docstring''' _SCREAMING_SNAKE_CASE =TFDPRQuestionEncoder.from_pretrained('facebook/dpr-question_encoder-single-nq-base' ) _SCREAMING_SNAKE_CASE =tf.constant( [[101, 7592, 1010, 2003, 2026, 3899, 1_0140, 1029, 102]] ) # [CLS] hello, is my dog cute? [SEP] _SCREAMING_SNAKE_CASE =model(A_ )[0] # embedding shape = (1, 768) # compare the actual values for a slice. _SCREAMING_SNAKE_CASE =tf.constant( [ [ 0.03_23_62_53, 0.12_75_33_35, 0.16_81_85_09, 0.00_27_97_86, 0.3_89_69_33, 0.24_26_49_45, 0.2_17_89_71, -0.02_33_52_27, -0.08_48_19_59, -0.14_32_41_17, ] ] ) self.assertTrue(numpy.allclose(output[:, :10].numpy() , expected_slice.numpy() , atol=1e-4 ) )
405
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetrImageProcessor class A( unittest.TestCase ): '''simple docstring''' def __init__( self : Optional[Any] , A_ : Dict , A_ : int=7 , A_ : Any=3 , A_ : List[str]=30 , A_ : Union[str, Any]=400 , A_ : List[str]=True , A_ : int=None , A_ : Any=True , A_ : str=1 / 255 , A_ : int=True , A_ : List[Any]=[0.5, 0.5, 0.5] , A_ : Union[str, Any]=[0.5, 0.5, 0.5] , A_ : Union[str, Any]=True , ) -> List[str]: """simple docstring""" lowerCamelCase_ = size if size is not None else {'shortest_edge': 18, 'longest_edge': 1333} lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = num_channels lowerCamelCase_ = min_resolution lowerCamelCase_ = max_resolution lowerCamelCase_ = do_resize lowerCamelCase_ = size lowerCamelCase_ = do_rescale lowerCamelCase_ = rescale_factor lowerCamelCase_ = do_normalize lowerCamelCase_ = image_mean lowerCamelCase_ = image_std lowerCamelCase_ = do_pad def a__ ( self : Tuple ) -> Dict: """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_pad": self.do_pad, } def a__ ( self : Union[str, Any] , A_ : Dict , A_ : Any=False ) -> Union[str, Any]: """simple docstring""" if not batched: lowerCamelCase_ = image_inputs[0] if isinstance(A_ , Image.Image ): lowerCamelCase_ , lowerCamelCase_ = image.size else: lowerCamelCase_ , lowerCamelCase_ = image.shape[1], image.shape[2] if w < h: lowerCamelCase_ = int(self.size['shortest_edge'] * h / w ) lowerCamelCase_ = self.size['shortest_edge'] elif w > h: lowerCamelCase_ = self.size['shortest_edge'] lowerCamelCase_ = int(self.size['shortest_edge'] * w / h ) else: lowerCamelCase_ = self.size['shortest_edge'] lowerCamelCase_ = self.size['shortest_edge'] else: lowerCamelCase_ = [] for image in image_inputs: lowerCamelCase_ , lowerCamelCase_ = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) lowerCamelCase_ = max(A_ , key=lambda A_ : item[0] )[0] lowerCamelCase_ = max(A_ , key=lambda A_ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class A( UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = DetrImageProcessor if is_vision_available() else None def a__ ( self : List[Any] ) -> str: """simple docstring""" lowerCamelCase_ = DetrImageProcessingTester(self ) @property def a__ ( self : List[str] ) -> List[Any]: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def a__ ( self : Dict ) -> int: """simple docstring""" lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , 'image_mean' ) ) self.assertTrue(hasattr(A_ , 'image_std' ) ) self.assertTrue(hasattr(A_ , 'do_normalize' ) ) self.assertTrue(hasattr(A_ , 'do_rescale' ) ) self.assertTrue(hasattr(A_ , 'rescale_factor' ) ) self.assertTrue(hasattr(A_ , 'do_resize' ) ) self.assertTrue(hasattr(A_ , 'size' ) ) self.assertTrue(hasattr(A_ , 'do_pad' ) ) def a__ ( self : List[str] ) -> str: """simple docstring""" lowerCamelCase_ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 18, 'longest_edge': 1333} ) self.assertEqual(image_processor.do_pad , A_ ) lowerCamelCase_ = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A_ ) self.assertEqual(image_processor.size , {'shortest_edge': 42, 'longest_edge': 84} ) self.assertEqual(image_processor.do_pad , A_ ) def a__ ( self : Dict ) -> Any: """simple docstring""" pass def a__ ( self : Tuple ) -> Tuple: """simple docstring""" lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCamelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ ) for image in image_inputs: self.assertIsInstance(A_ , Image.Image ) # Test not batched input lowerCamelCase_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) lowerCamelCase_ = image_processing(A_ , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def a__ ( self : List[str] ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCamelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , np.ndarray ) # Test not batched input lowerCamelCase_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase_ = image_processing(A_ , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def a__ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCamelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test not batched input lowerCamelCase_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase_ = image_processing(A_ , return_tensors='pt' ).pixel_values lowerCamelCase_ , lowerCamelCase_ = self.image_processor_tester.get_expected_values(A_ , batched=A_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def a__ ( self : Tuple ) -> List[Any]: """simple docstring""" lowerCamelCase_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_annotations.txt' , 'r' ) as f: lowerCamelCase_ = json.loads(f.read() ) lowerCamelCase_ = {'image_id': 39769, 'annotations': target} # encode them lowerCamelCase_ = DetrImageProcessor.from_pretrained('facebook/detr-resnet-50' ) lowerCamelCase_ = image_processing(images=A_ , annotations=A_ , return_tensors='pt' ) # verify pixel values lowerCamelCase_ = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , A_ ) lowerCamelCase_ = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area lowerCamelCase_ = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , A_ ) ) # verify boxes lowerCamelCase_ = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , A_ ) lowerCamelCase_ = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , A_ , atol=1E-3 ) ) # verify image_id lowerCamelCase_ = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , A_ ) ) # verify is_crowd lowerCamelCase_ = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , A_ ) ) # verify class_labels lowerCamelCase_ = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , A_ ) ) # verify orig_size lowerCamelCase_ = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , A_ ) ) # verify size lowerCamelCase_ = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , A_ ) ) @slow def a__ ( self : str ) -> Any: """simple docstring""" lowerCamelCase_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt' , 'r' ) as f: lowerCamelCase_ = json.loads(f.read() ) lowerCamelCase_ = {'file_name': '000000039769.png', 'image_id': 39769, 'segments_info': target} lowerCamelCase_ = pathlib.Path('./tests/fixtures/tests_samples/COCO/coco_panoptic' ) # encode them lowerCamelCase_ = DetrImageProcessor.from_pretrained('facebook/detr-resnet-50-panoptic' ) lowerCamelCase_ = image_processing(images=A_ , annotations=A_ , masks_path=A_ , return_tensors='pt' ) # verify pixel values lowerCamelCase_ = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , A_ ) lowerCamelCase_ = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , A_ , atol=1E-4 ) ) # verify area lowerCamelCase_ = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , A_ ) ) # verify boxes lowerCamelCase_ = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , A_ ) lowerCamelCase_ = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , A_ , atol=1E-3 ) ) # verify image_id lowerCamelCase_ = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , A_ ) ) # verify is_crowd lowerCamelCase_ = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , A_ ) ) # verify class_labels lowerCamelCase_ = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , A_ ) ) # verify masks lowerCamelCase_ = 822873 self.assertEqual(encoding['labels'][0]['masks'].sum().item() , A_ ) # verify orig_size lowerCamelCase_ = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , A_ ) ) # verify size lowerCamelCase_ = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , A_ ) )
70
0
'''simple docstring''' def snake_case ( snake_case : str ) -> Union[str, Any]: """simple docstring""" return "".join(chr(ord(snake_case ) - 32 ) if 'a' <= char <= 'z' else char for char in word ) if __name__ == "__main__": from doctest import testmod testmod()
284
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase : Dict = logging.get_logger(__name__) lowerCamelCase : int = { "microsoft/swinv2-tiny-patch4-window8-256": ( "https://huggingface.co/microsoft/swinv2-tiny-patch4-window8-256/resolve/main/config.json" ), } class A( UpperCamelCase ): '''simple docstring''' UpperCamelCase = '''swinv2''' UpperCamelCase = { '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers''', } def __init__( self : Dict , A_ : List[Any]=224 , A_ : Optional[Any]=4 , A_ : int=3 , A_ : Dict=96 , A_ : Any=[2, 2, 6, 2] , A_ : Optional[Any]=[3, 6, 12, 24] , A_ : Tuple=7 , A_ : Tuple=4.0 , A_ : str=True , A_ : str=0.0 , A_ : Union[str, Any]=0.0 , A_ : Optional[Any]=0.1 , A_ : str="gelu" , A_ : int=False , A_ : str=0.02 , A_ : List[Any]=1E-5 , A_ : Any=32 , **A_ : Tuple , ) -> Any: """simple docstring""" super().__init__(**A_ ) lowerCamelCase_ = image_size lowerCamelCase_ = patch_size lowerCamelCase_ = num_channels lowerCamelCase_ = embed_dim lowerCamelCase_ = depths lowerCamelCase_ = len(A_ ) lowerCamelCase_ = num_heads lowerCamelCase_ = window_size lowerCamelCase_ = mlp_ratio lowerCamelCase_ = qkv_bias lowerCamelCase_ = hidden_dropout_prob lowerCamelCase_ = attention_probs_dropout_prob lowerCamelCase_ = drop_path_rate lowerCamelCase_ = hidden_act lowerCamelCase_ = use_absolute_embeddings lowerCamelCase_ = layer_norm_eps lowerCamelCase_ = initializer_range lowerCamelCase_ = encoder_stride # we set the hidden_size attribute in order to make Swinv2 work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowerCamelCase_ = int(embed_dim * 2 ** (len(A_ ) - 1) ) lowerCamelCase_ = (0, 0, 0, 0)
70
0
'''simple docstring''' import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class __SCREAMING_SNAKE_CASE (__A , unittest.TestCase ): """simple docstring""" _a : Any = TransfoXLTokenizer _a : int = False _a : List[Any] = False def _a ( self ): """simple docstring""" super().setUp() a_ = [ '<unk>', '[CLS]', '[SEP]', 'want', 'unwanted', 'wa', 'un', 'running', ',', 'low', 'l', ] a_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def _a ( self , **UpperCamelCase__ ): """simple docstring""" a_ = True return TransfoXLTokenizer.from_pretrained(self.tmpdirname , **A_ ) def _a ( self , UpperCamelCase__ ): """simple docstring""" a_ = '<unk> UNwanted , running' a_ = '<unk> unwanted, running' return input_text, output_text def _a ( self ): """simple docstring""" a_ = TransfoXLTokenizer(vocab_file=self.vocab_file , lower_case=A_ ) a_ = tokenizer.tokenize('<unk> UNwanted , running' ) self.assertListEqual(A_ , ['<unk>', 'unwanted', ',', 'running'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(A_ ) , [0, 4, 8, 7] ) def _a ( self ): """simple docstring""" a_ = TransfoXLTokenizer(lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['hello', '!', 'how', 'are', 'you', '?'] ) def _a ( self ): """simple docstring""" a_ = TransfoXLTokenizer(lower_case=A_ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo ! how \n Are yoU ? ' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?'] ) def _a ( self ): """simple docstring""" a_ = TransfoXLTokenizer(lower_case=A_ ) a_ = 'Hello (bracket) and side-scrolled [and] Henry\'s $5,000 with 3.34 m. What\'s up!?' a_ = [ 'Hello', '(', 'bracket', ')', 'and', 'side', '@-@', 'scrolled', '[', 'and', ']', 'Henry', '\'s', '$', '5', '@,@', '000', 'with', '3', '@.@', '34', 'm', '.', 'What', '\'s', 'up', '!', '?', ] self.assertListEqual(tokenizer.tokenize(A_ ) , A_ ) self.assertEqual(tokenizer.convert_tokens_to_string(A_ ) , A_ ) def _a ( self ): """simple docstring""" a_ = self.get_tokenizer() a_ = len(A_ ) tokenizer.add_tokens(['new1', 'new2'] ) tokenizer.move_added_token('new1' , 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(A_ ) , original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode('new1' ) , [1] ) self.assertEqual(tokenizer.decode([1] ) , 'new1' )
536
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import ( AutoProcessor, BertTokenizerFast, BlipImageProcessor, GPTaTokenizer, InstructBlipProcessor, PreTrainedTokenizerFast, ) @require_vision class A( unittest.TestCase ): '''simple docstring''' def a__ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = tempfile.mkdtemp() lowerCamelCase_ = BlipImageProcessor() lowerCamelCase_ = GPTaTokenizer.from_pretrained('hf-internal-testing/tiny-random-GPT2Model' ) lowerCamelCase_ = BertTokenizerFast.from_pretrained('hf-internal-testing/tiny-random-bert' ) lowerCamelCase_ = InstructBlipProcessor(A_ , A_ , A_ ) processor.save_pretrained(self.tmpdirname ) def a__ ( self : Optional[int] , **A_ : Optional[int] ) -> Dict: """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **A_ ).tokenizer def a__ ( self : List[str] , **A_ : str ) -> Optional[Any]: """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **A_ ).image_processor def a__ ( self : Tuple , **A_ : Any ) -> Optional[int]: """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **A_ ).qformer_tokenizer def a__ ( self : str ) -> str: """simple docstring""" shutil.rmtree(self.tmpdirname ) def a__ ( self : Dict ) -> Dict: """simple docstring""" lowerCamelCase_ = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] lowerCamelCase_ = [Image.fromarray(np.moveaxis(A_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def a__ ( self : Tuple ) -> Tuple: """simple docstring""" lowerCamelCase_ = InstructBlipProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() , qformer_tokenizer=self.get_qformer_tokenizer() , ) processor.save_pretrained(self.tmpdirname ) lowerCamelCase_ = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) lowerCamelCase_ = self.get_image_processor(do_normalize=A_ , padding_value=1.0 ) lowerCamelCase_ = InstructBlipProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=A_ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , A_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , A_ ) self.assertIsInstance(processor.qformer_tokenizer , A_ ) def a__ ( self : List[str] ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = self.prepare_image_inputs() lowerCamelCase_ = image_processor(A_ , return_tensors='np' ) lowerCamelCase_ = processor(images=A_ , return_tensors='np' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def a__ ( self : Union[str, Any] ) -> Tuple: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = 'lower newer' lowerCamelCase_ = processor(text=A_ ) lowerCamelCase_ = tokenizer(A_ , return_token_type_ids=A_ ) lowerCamelCase_ = qformer_tokenizer(A_ , return_token_type_ids=A_ ) for key in encoded_tokens.keys(): self.assertListEqual(encoded_tokens[key] , encoded_processor[key] ) for key in encoded_tokens_qformer.keys(): self.assertListEqual(encoded_tokens_qformer[key] , encoded_processor['qformer_' + key] ) def a__ ( self : Tuple ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = 'lower newer' lowerCamelCase_ = self.prepare_image_inputs() lowerCamelCase_ = processor(text=A_ , images=A_ ) self.assertListEqual( list(inputs.keys() ) , ['input_ids', 'attention_mask', 'qformer_input_ids', 'qformer_attention_mask', 'pixel_values'] , ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def a__ ( self : Optional[int] ) -> int: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowerCamelCase_ = processor.batch_decode(A_ ) lowerCamelCase_ = tokenizer.batch_decode(A_ ) self.assertListEqual(A_ , A_ ) def a__ ( self : Tuple ) -> List[str]: """simple docstring""" lowerCamelCase_ = self.get_image_processor() lowerCamelCase_ = self.get_tokenizer() lowerCamelCase_ = self.get_qformer_tokenizer() lowerCamelCase_ = InstructBlipProcessor( tokenizer=A_ , image_processor=A_ , qformer_tokenizer=A_ ) lowerCamelCase_ = 'lower newer' lowerCamelCase_ = self.prepare_image_inputs() lowerCamelCase_ = processor(text=A_ , images=A_ ) self.assertListEqual( list(inputs.keys() ) , ['input_ids', 'attention_mask', 'qformer_input_ids', 'qformer_attention_mask', 'pixel_values'] , )
70
0
import argparse import json from collections import OrderedDict from functools import partial from pathlib import Path import timm import torch from huggingface_hub import hf_hub_download from transformers import LevitConfig, LevitForImageClassificationWithTeacher, LevitImageProcessor from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE__ = logging.get_logger() def A ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = True ) -> Any: print(f'''Converting {name}...''' ) with torch.no_grad(): if hidden_sizes == 128: if name[-1] == "S": A__ = timm.create_model('levit_128s' , pretrained=__UpperCamelCase ) else: A__ = timm.create_model('levit_128' , pretrained=__UpperCamelCase ) if hidden_sizes == 192: A__ = timm.create_model('levit_192' , pretrained=__UpperCamelCase ) if hidden_sizes == 256: A__ = timm.create_model('levit_256' , pretrained=__UpperCamelCase ) if hidden_sizes == 384: A__ = timm.create_model('levit_384' , pretrained=__UpperCamelCase ) from_model.eval() A__ = LevitForImageClassificationWithTeacher(__UpperCamelCase ).eval() A__ = OrderedDict() A__ = from_model.state_dict() A__ = list(from_model.state_dict().keys() ) A__ = list(our_model.state_dict().keys() ) print(len(__UpperCamelCase ) , len(__UpperCamelCase ) ) for i in range(len(__UpperCamelCase ) ): A__ = weights[og_keys[i]] our_model.load_state_dict(__UpperCamelCase ) A__ = torch.randn((2, 3, 224, 224) ) A__ = from_model(__UpperCamelCase ) A__ = our_model(__UpperCamelCase ).logits assert torch.allclose(__UpperCamelCase , __UpperCamelCase ), "The model logits don't match the original one." A__ = name print(__UpperCamelCase ) if push_to_hub: our_model.save_pretrained(save_directory / checkpoint_name ) A__ = LevitImageProcessor() image_processor.save_pretrained(save_directory / checkpoint_name ) print(f'''Pushed {checkpoint_name}''' ) def A ( __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = True ) -> Optional[int]: A__ = 'imagenet-1k-id2label.json' A__ = 1_000 A__ = (1, num_labels) A__ = 'huggingface/label-files' A__ = num_labels A__ = json.load(open(hf_hub_download(__UpperCamelCase , __UpperCamelCase , repo_type='dataset' ) , 'r' ) ) A__ = {int(__UpperCamelCase ): v for k, v in idalabel.items()} A__ = idalabel A__ = {v: k for k, v in idalabel.items()} A__ = partial(__UpperCamelCase , num_labels=__UpperCamelCase , idalabel=__UpperCamelCase , labelaid=__UpperCamelCase ) A__ = { 'levit-128S': 128, 'levit-128': 128, 'levit-192': 192, 'levit-256': 256, 'levit-384': 384, } A__ = { 'levit-128S': ImageNetPreTrainedConfig( hidden_sizes=[128, 256, 384] , num_attention_heads=[4, 6, 8] , depths=[2, 3, 4] , key_dim=[16, 16, 16] , drop_path_rate=0 , ), 'levit-128': ImageNetPreTrainedConfig( hidden_sizes=[128, 256, 384] , num_attention_heads=[4, 8, 12] , depths=[4, 4, 4] , key_dim=[16, 16, 16] , drop_path_rate=0 , ), 'levit-192': ImageNetPreTrainedConfig( hidden_sizes=[192, 288, 384] , num_attention_heads=[3, 5, 6] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0 , ), 'levit-256': ImageNetPreTrainedConfig( hidden_sizes=[256, 384, 512] , num_attention_heads=[4, 6, 8] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0 , ), 'levit-384': ImageNetPreTrainedConfig( hidden_sizes=[384, 512, 768] , num_attention_heads=[6, 9, 12] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0.1 , ), } if model_name: convert_weight_and_push( names_to_hidden_sizes[model_name] , __UpperCamelCase , names_to_config[model_name] , __UpperCamelCase , __UpperCamelCase ) else: for model_name, config in names_to_config.items(): convert_weight_and_push(names_to_hidden_sizes[model_name] , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) return config, expected_shape if __name__ == "__main__": SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default=None, type=str, help='''The name of the model you wish to convert, it must be one of the supported Levit* architecture,''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default='''levit-dump-folder/''', type=Path, required=False, help='''Path to the output PyTorch model directory.''', ) parser.add_argument('''--push_to_hub''', action='''store_true''', help='''Push model and image processor to the hub''') parser.add_argument( '''--no-push_to_hub''', dest='''push_to_hub''', action='''store_false''', help='''Do not push model and image processor to the hub''', ) SCREAMING_SNAKE_CASE__ = parser.parse_args() SCREAMING_SNAKE_CASE__ = args.pytorch_dump_folder_path pytorch_dump_folder_path.mkdir(exist_ok=True, parents=True) convert_weights_and_push(pytorch_dump_folder_path, args.model_name, args.push_to_hub)
9
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_OBJECT_DETECTION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING lowerCamelCase : Tuple = logging.get_logger(__name__) lowerCamelCase : List[Any] = Dict[str, Any] lowerCamelCase : Dict = List[Prediction] @add_end_docstrings(UpperCamelCase ) class A( UpperCamelCase ): '''simple docstring''' def __init__( self : Tuple , *A_ : int , **A_ : int ) -> Optional[int]: """simple docstring""" super().__init__(*A_ , **A_ ) if self.framework == "tf": raise ValueError(f"""The {self.__class__} is only available in PyTorch.""" ) requires_backends(self , 'vision' ) self.check_model_type( dict(MODEL_FOR_OBJECT_DETECTION_MAPPING.items() + MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING.items() ) ) def a__ ( self : Union[str, Any] , **A_ : Union[str, Any] ) -> List[str]: """simple docstring""" lowerCamelCase_ = {} if "threshold" in kwargs: lowerCamelCase_ = kwargs['threshold'] return {}, {}, postprocess_kwargs def __call__( self : str , *A_ : Optional[int] , **A_ : Tuple ) -> Union[Predictions, List[Prediction]]: """simple docstring""" return super().__call__(*A_ , **A_ ) def a__ ( self : Union[str, Any] , A_ : Tuple ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = load_image(A_ ) lowerCamelCase_ = torch.IntTensor([[image.height, image.width]] ) lowerCamelCase_ = self.image_processor(images=[image] , return_tensors='pt' ) if self.tokenizer is not None: lowerCamelCase_ = self.tokenizer(text=inputs['words'] , boxes=inputs['boxes'] , return_tensors='pt' ) lowerCamelCase_ = target_size return inputs def a__ ( self : Union[str, Any] , A_ : List[str] ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = model_inputs.pop('target_size' ) lowerCamelCase_ = self.model(**A_ ) lowerCamelCase_ = outputs.__class__({'target_size': target_size, **outputs} ) if self.tokenizer is not None: lowerCamelCase_ = model_inputs['bbox'] return model_outputs def a__ ( self : str , A_ : Any , A_ : Tuple=0.9 ) -> str: """simple docstring""" lowerCamelCase_ = model_outputs['target_size'] if self.tokenizer is not None: # This is a LayoutLMForTokenClassification variant. # The OCR got the boxes and the model classified the words. lowerCamelCase_ , lowerCamelCase_ = target_size[0].tolist() def unnormalize(A_ : Dict ): return self._get_bounding_box( torch.Tensor( [ (width * bbox[0] / 1000), (height * bbox[1] / 1000), (width * bbox[2] / 1000), (height * bbox[3] / 1000), ] ) ) lowerCamelCase_ , lowerCamelCase_ = model_outputs['logits'].squeeze(0 ).softmax(dim=-1 ).max(dim=-1 ) lowerCamelCase_ = [self.model.config.idalabel[prediction] for prediction in classes.tolist()] lowerCamelCase_ = [unnormalize(A_ ) for bbox in model_outputs['bbox'].squeeze(0 )] lowerCamelCase_ = ['score', 'label', 'box'] lowerCamelCase_ = [dict(zip(A_ , A_ ) ) for vals in zip(scores.tolist() , A_ , A_ ) if vals[0] > threshold] else: # This is a regular ForObjectDetectionModel lowerCamelCase_ = self.image_processor.post_process_object_detection(A_ , A_ , A_ ) lowerCamelCase_ = raw_annotations[0] lowerCamelCase_ = raw_annotation['scores'] lowerCamelCase_ = raw_annotation['labels'] lowerCamelCase_ = raw_annotation['boxes'] lowerCamelCase_ = scores.tolist() lowerCamelCase_ = [self.model.config.idalabel[label.item()] for label in labels] lowerCamelCase_ = [self._get_bounding_box(A_ ) for box in boxes] # {"scores": [...], ...} --> [{"score":x, ...}, ...] lowerCamelCase_ = ['score', 'label', 'box'] lowerCamelCase_ = [ dict(zip(A_ , A_ ) ) for vals in zip(raw_annotation['scores'] , raw_annotation['labels'] , raw_annotation['boxes'] ) ] return annotation def a__ ( self : Union[str, Any] , A_ : "torch.Tensor" ) -> Dict[str, int]: """simple docstring""" if self.framework != "pt": raise ValueError('The ObjectDetectionPipeline is only available in PyTorch.' ) lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = box.int().tolist() lowerCamelCase_ = { 'xmin': xmin, 'ymin': ymin, 'xmax': xmax, 'ymax': ymax, } return bbox
70
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) __lowerCAmelCase ={ "configuration_clip": [ "CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "CLIPConfig", "CLIPOnnxConfig", "CLIPTextConfig", "CLIPVisionConfig", ], "processing_clip": ["CLIPProcessor"], "tokenization_clip": ["CLIPTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase =["CLIPTokenizerFast"] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase =["CLIPFeatureExtractor"] __lowerCAmelCase =["CLIPImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase =[ "CLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "CLIPModel", "CLIPPreTrainedModel", "CLIPTextModel", "CLIPTextModelWithProjection", "CLIPVisionModel", "CLIPVisionModelWithProjection", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase =[ "TF_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "TFCLIPModel", "TFCLIPPreTrainedModel", "TFCLIPTextModel", "TFCLIPVisionModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase =[ "FlaxCLIPModel", "FlaxCLIPPreTrainedModel", "FlaxCLIPTextModel", "FlaxCLIPTextPreTrainedModel", "FlaxCLIPVisionModel", "FlaxCLIPVisionPreTrainedModel", ] if TYPE_CHECKING: from .configuration_clip import ( CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, CLIPConfig, CLIPOnnxConfig, CLIPTextConfig, CLIPVisionConfig, ) from .processing_clip import CLIPProcessor from .tokenization_clip import CLIPTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_clip_fast import CLIPTokenizerFast try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_clip import CLIPFeatureExtractor from .image_processing_clip import CLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_clip import ( CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, CLIPModel, CLIPPreTrainedModel, CLIPTextModel, CLIPTextModelWithProjection, CLIPVisionModel, CLIPVisionModelWithProjection, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_clip import ( TF_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFCLIPModel, TFCLIPPreTrainedModel, TFCLIPTextModel, TFCLIPVisionModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_clip import ( FlaxCLIPModel, FlaxCLIPPreTrainedModel, FlaxCLIPTextModel, FlaxCLIPTextPreTrainedModel, FlaxCLIPVisionModel, FlaxCLIPVisionPreTrainedModel, ) else: import sys __lowerCAmelCase =_LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
697
from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
70
0
A : List[Any] = range(2, 2_0 + 1) A : int = [1_0**k for k in range(ks[-1] + 1)] A : dict[int, dict[int, list[list[int]]]] = {} def __lowerCAmelCase ( a__ , a__ , a__ , a__ ) -> int: __a = sum(a_i[j] for j in range(a__ , len(a__ ) ) ) __a = sum(a_i[j] * base[j] for j in range(min(len(a__ ) , a__ ) ) ) __a , __a = 0, 0 __a = n - i __a = memo.get(a__ ) if sub_memo is not None: __a = sub_memo.get(a__ ) if jumps is not None and len(a__ ) > 0: # find and make the largest jump without going over __a = -1 for _k in range(len(a__ ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: __a = _k break if max_jump >= 0: __a , __a , __a = jumps[max_jump] # since the difference between jumps is cached, add c __a = diff + c for j in range(min(a__ , len(a__ ) ) ): __a , __a = divmod(a__ , 10 ) if new_c > 0: add(a__ , a__ , a__ ) else: __a = [] else: __a = {c: []} __a = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps __a , __a = next_term(a__ , k - 1 , i + dn , a__ ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead __a , __a = compute(a__ , a__ , i + dn , a__ ) diff += _diff dn += terms_jumped __a = sub_memo[c] # keep jumps sorted by # of terms skipped __a = 0 while j < len(a__ ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(a__ , (diff, dn, k) ) return (diff, dn) def __lowerCAmelCase ( a__ , a__ , a__ , a__ ) -> Any: if i >= n: return 0, i if k > len(a__ ): a_i.extend([0 for _ in range(k - len(a__ ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) __a = i __a , __a , __a = 0, 0, 0 for j in range(len(a__ ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 __a = ds_c + ds_b diff += addend __a = 0 for j in range(a__ ): __a = a_i[j] + addend __a , __a = divmod(a__ , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(a__ , a__ , a__ ) return diff, i - start_i def __lowerCAmelCase ( a__ , a__ , a__ ) -> Dict: for j in range(a__ , len(a__ ) ): __a = digits[j] + addend if s >= 10: __a , __a = divmod(a__ , 10 ) __a = addend // 10 + quotient else: __a = s __a = addend // 10 if addend == 0: break while addend > 0: __a , __a = divmod(a__ , 10 ) digits.append(a__ ) def __lowerCAmelCase ( a__ = 10**15 ) -> List[str]: __a = [1] __a = 1 __a = 0 while True: __a , __a = next_term(a__ , 20 , i + dn , a__ ) dn += terms_jumped if dn == n - i: break __a = 0 for j in range(len(a__ ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(F"{solution() = }")
219
from collections import Counter from timeit import timeit def _SCREAMING_SNAKE_CASE ( lowercase : str = "" , ): '''simple docstring''' return sum(c % 2 for c in Counter(input_str.replace(' ' , '' ).lower() ).values() ) < 2 def _SCREAMING_SNAKE_CASE ( lowercase : str = "" ): '''simple docstring''' if len(lowercase ) == 0: return True lowerCamelCase_ = input_str.replace(' ' , '' ).lower() # character_freq_dict: Stores the frequency of every character in the input string lowerCamelCase_ = {} for character in lower_case_input_str: lowerCamelCase_ = character_freq_dict.get(lowercase , 0 ) + 1 lowerCamelCase_ = 0 for character_count in character_freq_dict.values(): if character_count % 2: odd_char += 1 if odd_char > 1: return False return True def _SCREAMING_SNAKE_CASE ( lowercase : str = "" ): '''simple docstring''' print('\nFor string = ' , lowercase , ':' ) print( '> can_string_be_rearranged_as_palindrome_counter()' , '\tans =' , can_string_be_rearranged_as_palindrome_counter(lowercase ) , '\ttime =' , timeit( 'z.can_string_be_rearranged_as_palindrome_counter(z.check_str)' , setup='import __main__ as z' , ) , 'seconds' , ) print( '> can_string_be_rearranged_as_palindrome()' , '\tans =' , can_string_be_rearranged_as_palindrome(lowercase ) , '\ttime =' , timeit( 'z.can_string_be_rearranged_as_palindrome(z.check_str)' , setup='import __main__ as z' , ) , 'seconds' , ) if __name__ == "__main__": lowerCamelCase : Optional[Any] = input( "Enter string to determine if it can be rearranged as a palindrome or not: " ).strip() benchmark(check_str) lowerCamelCase : int = can_string_be_rearranged_as_palindrome_counter(check_str) print(F"""{check_str} can {'' if status else 'not '}be rearranged as a palindrome""")
70
0
'''simple docstring''' import os # Precomputes a list of the 100 first triangular numbers SCREAMING_SNAKE_CASE_ = [int(0.5 * n * (n + 1)) for n in range(1, 1_01)] def UpperCamelCase__ ( ) -> int: __UpperCAmelCase: Union[str, Any] = os.path.dirname(os.path.realpath(_lowercase ) ) __UpperCAmelCase: Tuple = os.path.join(_lowercase , """words.txt""" ) __UpperCAmelCase: List[Any] = """""" with open(_lowercase ) as f: __UpperCAmelCase: Optional[int] = f.readline() __UpperCAmelCase: Dict = [word.strip("""\"""" ) for word in words.strip("""\r\n""" ).split(""",""" )] __UpperCAmelCase: Any = [ word for word in [sum(ord(_lowercase ) - 6_4 for x in word ) for word in words] if word in TRIANGULAR_NUMBERS ] return len(_lowercase ) if __name__ == "__main__": print(solution())
523
from __future__ import annotations from decimal import Decimal from math import * # noqa: F403 from sympy import diff def _SCREAMING_SNAKE_CASE ( lowercase : str , lowercase : float | Decimal , lowercase : float = 10**-10 ): '''simple docstring''' lowerCamelCase_ = a while True: lowerCamelCase_ = Decimal(lowercase ) - ( Decimal(eval(lowercase ) ) / Decimal(eval(str(diff(lowercase ) ) ) ) # noqa: S307 ) # This number dictates the accuracy of the answer if abs(eval(lowercase ) ) < precision: # noqa: S307 return float(lowercase ) # Let's Execute if __name__ == "__main__": # Find root of trigonometric function # Find value of pi print(F"""The root of sin(x) = 0 is {newton_raphson('sin(x)', 2)}""") # Find root of polynomial print(F"""The root of x**2 - 5*x + 2 = 0 is {newton_raphson('x**2 - 5*x + 2', 0.4)}""") # Find Square Root of 5 print(F"""The root of log(x) - 1 = 0 is {newton_raphson('log(x) - 1', 2)}""") # Exponential Roots print(F"""The root of exp(x) - 1 = 0 is {newton_raphson('exp(x) - 1', 0)}""")
70
0
from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class _SCREAMING_SNAKE_CASE : lowerCamelCase_ = 4_2 # [batch_size x 3] lowerCamelCase_ = 4_2 # [batch_size x 3] lowerCamelCase_ = 4_2 # [batch_size x 3] lowerCamelCase_ = 4_2 # [batch_size x 3] lowerCamelCase_ = 4_2 lowerCamelCase_ = 4_2 lowerCamelCase_ = 4_2 lowerCamelCase_ = 4_2 lowerCamelCase_ = 4_2 def _UpperCAmelCase ( self : Optional[int] ): """simple docstring""" assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def _UpperCAmelCase ( self : Union[str, Any] ): """simple docstring""" return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def _UpperCAmelCase ( self : Dict ): """simple docstring""" return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def _UpperCAmelCase ( self : List[Any] ): """simple docstring""" A : Optional[int] = torch.arange(self.height * self.width ) A : List[str] = torch.stack( [ pixel_indices % self.width, torch.div(A_ , self.width , rounding_mode='''trunc''' ), ] , axis=1 , ) return coords @property def _UpperCAmelCase ( self : List[Any] ): """simple docstring""" A , *A : str = self.shape A : Optional[int] = int(np.prod(A_ ) ) A : Dict = self.get_image_coords() A : Any = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) A : List[Any] = self.get_camera_rays(A_ ) A : List[str] = rays.view(A_ , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def _UpperCAmelCase ( self : int , snake_case_ : torch.Tensor ): """simple docstring""" A , *A , A : Union[str, Any] = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] A : Any = coords.view(A_ , -1 , 2 ) A : Dict = self.resolution() A : Tuple = self.fov() A : Tuple = (flat.float() / (res - 1)) * 2 - 1 A : List[Any] = fracs * torch.tan(fov / 2 ) A : List[Any] = fracs.view(A_ , -1 , 2 ) A : str = ( self.z.view(A_ , 1 , 3 ) + self.x.view(A_ , 1 , 3 ) * fracs[:, :, :1] + self.y.view(A_ , 1 , 3 ) * fracs[:, :, 1:] ) A : List[Any] = directions / directions.norm(dim=-1 , keepdim=A_ ) A : Tuple = torch.stack( [ torch.broadcast_to(self.origin.view(A_ , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(A_ , *A_ , 2 , 3 ) def _UpperCAmelCase ( self : Any , snake_case_ : int , snake_case_ : int ): """simple docstring""" assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=A_ , height=A_ , x_fov=self.x_fov , y_fov=self.y_fov , ) def _lowerCamelCase ( lowerCamelCase_: int ): '''simple docstring''' A : List[str] = [] A : Tuple = [] A : Optional[Any] = [] A : List[str] = [] for theta in np.linspace(0 , 2 * np.pi , num=20 ): A : Optional[Any] = np.array([np.sin(lowerCamelCase_ ), np.cos(lowerCamelCase_ ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) A : Dict = -z * 4 A : Optional[Any] = np.array([np.cos(lowerCamelCase_ ), -np.sin(lowerCamelCase_ ), 0.0] ) A : str = np.cross(lowerCamelCase_ , lowerCamelCase_ ) origins.append(lowerCamelCase_ ) xs.append(lowerCamelCase_ ) ys.append(lowerCamelCase_ ) zs.append(lowerCamelCase_ ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(lowerCamelCase_ , axis=0 ) ).float() , x=torch.from_numpy(np.stack(lowerCamelCase_ , axis=0 ) ).float() , y=torch.from_numpy(np.stack(lowerCamelCase_ , axis=0 ) ).float() , z=torch.from_numpy(np.stack(lowerCamelCase_ , axis=0 ) ).float() , width=lowerCamelCase_ , height=lowerCamelCase_ , x_fov=0.7 , y_fov=0.7 , shape=(1, len(lowerCamelCase_ )) , )
256
from __future__ import annotations from typing import Any class A( UpperCamelCase ): '''simple docstring''' pass class A: '''simple docstring''' def __init__( self : List[str] , A_ : Any ) -> None: """simple docstring""" lowerCamelCase_ = data lowerCamelCase_ = None def __iter__( self : int ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self lowerCamelCase_ = [] while node: if node in visited: raise ContainsLoopError visited.append(A_ ) yield node.data lowerCamelCase_ = node.next_node @property def a__ ( self : List[str] ) -> bool: """simple docstring""" try: list(self ) return False except ContainsLoopError: return True if __name__ == "__main__": lowerCamelCase : int = Node(1) lowerCamelCase : Optional[int] = Node(2) lowerCamelCase : Union[str, Any] = Node(3) lowerCamelCase : List[Any] = Node(4) print(root_node.has_loop) # False lowerCamelCase : int = root_node.next_node print(root_node.has_loop) # True lowerCamelCase : Dict = Node(5) lowerCamelCase : Optional[int] = Node(6) lowerCamelCase : str = Node(5) lowerCamelCase : Union[str, Any] = Node(6) print(root_node.has_loop) # False lowerCamelCase : List[str] = Node(1) print(root_node.has_loop) # False
70
0
"""simple docstring""" from __future__ import annotations import inspect import unittest from typing import List, Tuple from transformers import RegNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFRegNetForImageClassification, TFRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class snake_case : def __init__( self :Dict , _lowerCamelCase :List[str] , _lowerCamelCase :List[str]=3 , _lowerCamelCase :Dict=3_2 , _lowerCamelCase :Union[str, Any]=3 , _lowerCamelCase :Any=1_0 , _lowerCamelCase :List[Any]=[1_0, 2_0, 3_0, 4_0] , _lowerCamelCase :int=[1, 1, 2, 1] , _lowerCamelCase :List[str]=True , _lowerCamelCase :Tuple=True , _lowerCamelCase :Tuple="relu" , _lowerCamelCase :str=3 , _lowerCamelCase :int=None , ): __SCREAMING_SNAKE_CASE : Tuple = parent __SCREAMING_SNAKE_CASE : int = batch_size __SCREAMING_SNAKE_CASE : int = image_size __SCREAMING_SNAKE_CASE : Optional[int] = num_channels __SCREAMING_SNAKE_CASE : List[str] = embeddings_size __SCREAMING_SNAKE_CASE : int = hidden_sizes __SCREAMING_SNAKE_CASE : Any = depths __SCREAMING_SNAKE_CASE : List[Any] = is_training __SCREAMING_SNAKE_CASE : str = use_labels __SCREAMING_SNAKE_CASE : List[str] = hidden_act __SCREAMING_SNAKE_CASE : List[Any] = num_labels __SCREAMING_SNAKE_CASE : List[str] = scope __SCREAMING_SNAKE_CASE : Any = len(A_ ) def SCREAMING_SNAKE_CASE_ ( self :Union[str, Any] ): __SCREAMING_SNAKE_CASE : List[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __SCREAMING_SNAKE_CASE : Dict = None if self.use_labels: __SCREAMING_SNAKE_CASE : int = ids_tensor([self.batch_size] , self.num_labels ) __SCREAMING_SNAKE_CASE : Any = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE_ ( self :str ): return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , ) def SCREAMING_SNAKE_CASE_ ( self :Optional[Any] , _lowerCamelCase :Optional[Any] , _lowerCamelCase :Union[str, Any] , _lowerCamelCase :List[Any] ): __SCREAMING_SNAKE_CASE : Dict = TFRegNetModel(config=A_ ) __SCREAMING_SNAKE_CASE : List[Any] = model(A_ , training=A_ ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 3_2, self.image_size // 3_2) , ) def SCREAMING_SNAKE_CASE_ ( self :int , _lowerCamelCase :Tuple , _lowerCamelCase :int , _lowerCamelCase :str ): __SCREAMING_SNAKE_CASE : List[str] = self.num_labels __SCREAMING_SNAKE_CASE : List[Any] = TFRegNetForImageClassification(A_ ) __SCREAMING_SNAKE_CASE : List[str] = model(A_ , labels=A_ , training=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE_ ( self :Tuple ): __SCREAMING_SNAKE_CASE : Dict = self.prepare_config_and_inputs() __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : List[str] = config_and_inputs __SCREAMING_SNAKE_CASE : Optional[Any] = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class snake_case ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): lowerCamelCase__ = (TFRegNetModel, TFRegNetForImageClassification) if is_tf_available() else () lowerCamelCase__ = ( {'''feature-extraction''': TFRegNetModel, '''image-classification''': TFRegNetForImageClassification} if is_tf_available() else {} ) lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False def SCREAMING_SNAKE_CASE_ ( self :str ): __SCREAMING_SNAKE_CASE : Any = TFRegNetModelTester(self ) __SCREAMING_SNAKE_CASE : int = ConfigTester(self , config_class=A_ , has_text_modality=A_ ) def SCREAMING_SNAKE_CASE_ ( self :Optional[Any] ): return @unittest.skip(reason='''RegNet does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE_ ( self :Optional[Any] ): pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices('''GPU''' ) ) == 0 , reason='''TF does not support backprop for grouped convolutions on CPU.''' , ) @slow def SCREAMING_SNAKE_CASE_ ( self :Optional[int] ): super().test_keras_fit() @unittest.skip(reason='''RegNet does not support input and output embeddings''' ) def SCREAMING_SNAKE_CASE_ ( self :Optional[int] ): pass def SCREAMING_SNAKE_CASE_ ( self :Union[str, Any] ): __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __SCREAMING_SNAKE_CASE : Dict = model_class(A_ ) __SCREAMING_SNAKE_CASE : List[str] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __SCREAMING_SNAKE_CASE : Tuple = [*signature.parameters.keys()] __SCREAMING_SNAKE_CASE : List[str] = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , A_ ) def SCREAMING_SNAKE_CASE_ ( self :Optional[Any] ): __SCREAMING_SNAKE_CASE : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def SCREAMING_SNAKE_CASE_ ( self :Optional[int] ): def check_hidden_states_output(_lowerCamelCase :List[str] , _lowerCamelCase :List[str] , _lowerCamelCase :List[Any] ): __SCREAMING_SNAKE_CASE : Union[str, Any] = model_class(A_ ) __SCREAMING_SNAKE_CASE : Optional[int] = model(**self._prepare_for_class(A_ , A_ ) , training=A_ ) __SCREAMING_SNAKE_CASE : List[str] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __SCREAMING_SNAKE_CASE : Any = self.model_tester.num_stages self.assertEqual(len(A_ ) , expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 2, self.model_tester.image_size // 2] , ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : List[str] = self.model_tester.prepare_config_and_inputs_for_common() __SCREAMING_SNAKE_CASE : Optional[Any] = ['''basic''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: __SCREAMING_SNAKE_CASE : Tuple = layer_type __SCREAMING_SNAKE_CASE : Any = True check_hidden_states_output(A_ , A_ , A_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __SCREAMING_SNAKE_CASE : Any = True check_hidden_states_output(A_ , A_ , A_ ) def SCREAMING_SNAKE_CASE_ ( self :Any ): __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : List[str] = self.model_tester.prepare_config_and_inputs_for_common() def check_equivalence(_lowerCamelCase :Union[str, Any] , _lowerCamelCase :Optional[Any] , _lowerCamelCase :Optional[Any] , _lowerCamelCase :Dict={} ): __SCREAMING_SNAKE_CASE : Tuple = model(A_ , return_dict=A_ , **A_ ) __SCREAMING_SNAKE_CASE : Any = model(A_ , return_dict=A_ , **A_ ).to_tuple() def recursive_check(_lowerCamelCase :Any , _lowerCamelCase :Optional[Any] ): if isinstance(A_ , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(A_ , A_ ): recursive_check(A_ , A_ ) elif tuple_object is None: return else: self.assertTrue( all(tf.equal(A_ , A_ ) ) , msg=( '''Tuple and dict output are not equal. Difference:''' f''' {tf.math.reduce_max(tf.abs(tuple_object - dict_object ) )}''' ) , ) recursive_check(A_ , A_ ) for model_class in self.all_model_classes: __SCREAMING_SNAKE_CASE : Tuple = model_class(A_ ) __SCREAMING_SNAKE_CASE : Optional[Any] = self._prepare_for_class(A_ , A_ ) __SCREAMING_SNAKE_CASE : Optional[Any] = self._prepare_for_class(A_ , A_ ) check_equivalence(A_ , A_ , A_ ) __SCREAMING_SNAKE_CASE : Optional[int] = self._prepare_for_class(A_ , A_ , return_labels=A_ ) __SCREAMING_SNAKE_CASE : Optional[int] = self._prepare_for_class(A_ , A_ , return_labels=A_ ) check_equivalence(A_ , A_ , A_ ) __SCREAMING_SNAKE_CASE : Optional[Any] = self._prepare_for_class(A_ , A_ ) __SCREAMING_SNAKE_CASE : List[str] = self._prepare_for_class(A_ , A_ ) check_equivalence(A_ , A_ , A_ , {'''output_hidden_states''': True} ) __SCREAMING_SNAKE_CASE : int = self._prepare_for_class(A_ , A_ , return_labels=A_ ) __SCREAMING_SNAKE_CASE : Tuple = self._prepare_for_class(A_ , A_ , return_labels=A_ ) check_equivalence(A_ , A_ , A_ , {'''output_hidden_states''': True} ) def SCREAMING_SNAKE_CASE_ ( self :str ): __SCREAMING_SNAKE_CASE : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A_ ) @slow def SCREAMING_SNAKE_CASE_ ( self :str ): for model_name in TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __SCREAMING_SNAKE_CASE : Any = TFRegNetModel.from_pretrained(A_ ) self.assertIsNotNone(A_ ) def lowerCAmelCase_ ( ): '''simple docstring''' __SCREAMING_SNAKE_CASE : str = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class snake_case ( unittest.TestCase ): @cached_property def SCREAMING_SNAKE_CASE_ ( self :List[str] ): return ( AutoImageProcessor.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def SCREAMING_SNAKE_CASE_ ( self :Any ): __SCREAMING_SNAKE_CASE : Any = TFRegNetForImageClassification.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) __SCREAMING_SNAKE_CASE : List[Any] = self.default_image_processor __SCREAMING_SNAKE_CASE : int = prepare_img() __SCREAMING_SNAKE_CASE : Optional[Any] = image_processor(images=A_ , return_tensors='''tf''' ) # forward pass __SCREAMING_SNAKE_CASE : List[str] = model(**A_ , training=A_ ) # verify the logits __SCREAMING_SNAKE_CASE : Any = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , A_ ) __SCREAMING_SNAKE_CASE : Any = tf.constant([-0.4_1_8_0, -1.5_0_5_1, -3.4_8_3_6] ) tf.debugging.assert_near(outputs.logits[0, :3] , A_ , atol=1e-4 )
674
import unittest import torch from diffusers import DDIMScheduler, DDPMScheduler, UNetaDModel from diffusers.training_utils import set_seed from diffusers.utils.testing_utils import slow lowerCamelCase : int = False class A( unittest.TestCase ): '''simple docstring''' def a__ ( self : int , A_ : Dict=32 ) -> Any: """simple docstring""" set_seed(0 ) lowerCamelCase_ = UNetaDModel(sample_size=A_ , in_channels=3 , out_channels=3 ) lowerCamelCase_ = torch.optim.SGD(model.parameters() , lr=0.0001 ) return model, optimizer @slow def a__ ( self : int ) -> str: """simple docstring""" lowerCamelCase_ = 'cpu' # ensure full determinism without setting the CUBLAS_WORKSPACE_CONFIG env variable lowerCamelCase_ = DDPMScheduler( num_train_timesteps=1000 , beta_start=0.0001 , beta_end=0.02 , beta_schedule='linear' , clip_sample=A_ , ) lowerCamelCase_ = DDIMScheduler( num_train_timesteps=1000 , beta_start=0.0001 , beta_end=0.02 , beta_schedule='linear' , clip_sample=A_ , ) assert ddpm_scheduler.config.num_train_timesteps == ddim_scheduler.config.num_train_timesteps # shared batches for DDPM and DDIM set_seed(0 ) lowerCamelCase_ = [torch.randn((4, 3, 32, 32) ).clip(-1 , 1 ).to(A_ ) for _ in range(4 )] lowerCamelCase_ = [torch.randn((4, 3, 32, 32) ).to(A_ ) for _ in range(4 )] lowerCamelCase_ = [torch.randint(0 , 1000 , (4,) ).long().to(A_ ) for _ in range(4 )] # train with a DDPM scheduler lowerCamelCase_ , lowerCamelCase_ = self.get_model_optimizer(resolution=32 ) model.train().to(A_ ) for i in range(4 ): optimizer.zero_grad() lowerCamelCase_ = ddpm_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) lowerCamelCase_ = model(A_ , timesteps[i] ).sample lowerCamelCase_ = torch.nn.functional.mse_loss(A_ , noise[i] ) loss.backward() optimizer.step() del model, optimizer # recreate the model and optimizer, and retry with DDIM lowerCamelCase_ , lowerCamelCase_ = self.get_model_optimizer(resolution=32 ) model.train().to(A_ ) for i in range(4 ): optimizer.zero_grad() lowerCamelCase_ = ddim_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) lowerCamelCase_ = model(A_ , timesteps[i] ).sample lowerCamelCase_ = torch.nn.functional.mse_loss(A_ , noise[i] ) loss.backward() optimizer.step() del model, optimizer self.assertTrue(torch.allclose(A_ , A_ , atol=1E-5 ) ) self.assertTrue(torch.allclose(A_ , A_ , atol=1E-5 ) )
70
0