code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
"""simple docstring""" import copy import os from collections import OrderedDict from typing import TYPE_CHECKING, Any, Dict, Mapping, Optional, Union if TYPE_CHECKING: from ...processing_utils import ProcessorMixin from ...utils import TensorType from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { "google/owlvit-base-patch32": "https://huggingface.co/google/owlvit-base-patch32/resolve/main/config.json", "google/owlvit-base-patch16": "https://huggingface.co/google/owlvit-base-patch16/resolve/main/config.json", "google/owlvit-large-patch14": "https://huggingface.co/google/owlvit-large-patch14/resolve/main/config.json", } class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = "owlvit_text_model" def __init__(self , _lowercase=49408 , _lowercase=512 , _lowercase=2048 , _lowercase=12 , _lowercase=8 , _lowercase=16 , _lowercase="quick_gelu" , _lowercase=1e-5 , _lowercase=0.0 , _lowercase=0.02 , _lowercase=1.0 , _lowercase=0 , _lowercase=49406 , _lowercase=49407 , **_lowercase , ): '''simple docstring''' super().__init__(pad_token_id=_lowercase , bos_token_id=_lowercase , eos_token_id=_lowercase , **_lowercase ) __a : Tuple = vocab_size __a : Union[str, Any] = hidden_size __a : Tuple = intermediate_size __a : Union[str, Any] = num_hidden_layers __a : Optional[Any] = num_attention_heads __a : Union[str, Any] = max_position_embeddings __a : List[Any] = hidden_act __a : Tuple = layer_norm_eps __a : Tuple = attention_dropout __a : List[Any] = initializer_range __a : List[Any] = initializer_factor @classmethod def lowerCAmelCase__(cls , _lowercase , **_lowercase ): '''simple docstring''' cls._set_token_in_kwargs(_lowercase ) __a : List[Any] = cls.get_config_dict(_lowercase , **_lowercase ) # get the text config dict if we are loading from OwlViTConfig if config_dict.get("""model_type""" ) == "owlvit": __a : Dict = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(_lowercase , **_lowercase ) class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = "owlvit_vision_model" def __init__(self , _lowercase=768 , _lowercase=3072 , _lowercase=12 , _lowercase=12 , _lowercase=3 , _lowercase=768 , _lowercase=32 , _lowercase="quick_gelu" , _lowercase=1e-5 , _lowercase=0.0 , _lowercase=0.02 , _lowercase=1.0 , **_lowercase , ): '''simple docstring''' super().__init__(**_lowercase ) __a : Any = hidden_size __a : Tuple = intermediate_size __a : str = num_hidden_layers __a : int = num_attention_heads __a : int = num_channels __a : List[Any] = image_size __a : int = patch_size __a : Any = hidden_act __a : int = layer_norm_eps __a : Union[str, Any] = attention_dropout __a : str = initializer_range __a : Union[str, Any] = initializer_factor @classmethod def lowerCAmelCase__(cls , _lowercase , **_lowercase ): '''simple docstring''' cls._set_token_in_kwargs(_lowercase ) __a : List[str] = cls.get_config_dict(_lowercase , **_lowercase ) # get the vision config dict if we are loading from OwlViTConfig if config_dict.get("""model_type""" ) == "owlvit": __a : List[str] = config_dict["""vision_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(_lowercase , **_lowercase ) class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = "owlvit" _lowerCAmelCase = True def __init__(self , _lowercase=None , _lowercase=None , _lowercase=512 , _lowercase=2.6592 , _lowercase=True , **_lowercase , ): '''simple docstring''' super().__init__(**_lowercase ) if text_config is None: __a : Union[str, Any] = {} logger.info("""text_config is None. Initializing the OwlViTTextConfig with default values.""" ) if vision_config is None: __a : Tuple = {} logger.info("""vision_config is None. initializing the OwlViTVisionConfig with default values.""" ) __a : Tuple = OwlViTTextConfig(**_lowercase ) __a : Dict = OwlViTVisionConfig(**_lowercase ) __a : int = projection_dim __a : List[str] = logit_scale_init_value __a : List[str] = return_dict __a : Any = 1.0 @classmethod def lowerCAmelCase__(cls , _lowercase , **_lowercase ): '''simple docstring''' cls._set_token_in_kwargs(_lowercase ) __a : List[Any] = cls.get_config_dict(_lowercase , **_lowercase ) if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(_lowercase , **_lowercase ) @classmethod def lowerCAmelCase__(cls , _lowercase , _lowercase , **_lowercase ): '''simple docstring''' __a : Any = {} __a : int = text_config __a : Any = vision_config return cls.from_dict(_lowercase , **_lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' __a : List[Any] = copy.deepcopy(self.__dict__ ) __a : List[str] = self.text_config.to_dict() __a : Union[str, Any] = self.vision_config.to_dict() __a : List[str] = self.__class__.model_type return output class SCREAMING_SNAKE_CASE__ ( __snake_case ): @property def lowerCAmelCase__(self ): '''simple docstring''' return OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """sequence"""}), ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""attention_mask""", {0: """batch""", 1: """sequence"""}), ] ) @property def lowerCAmelCase__(self ): '''simple docstring''' return OrderedDict( [ ("""logits_per_image""", {0: """batch"""}), ("""logits_per_text""", {0: """batch"""}), ("""text_embeds""", {0: """batch"""}), ("""image_embeds""", {0: """batch"""}), ] ) @property def lowerCAmelCase__(self ): '''simple docstring''' return 1e-4 def lowerCAmelCase__(self , _lowercase , _lowercase = -1 , _lowercase = -1 , _lowercase = None , ): '''simple docstring''' __a : Optional[int] = super().generate_dummy_inputs( processor.tokenizer , batch_size=_lowercase , seq_length=_lowercase , framework=_lowercase ) __a : Optional[int] = super().generate_dummy_inputs( processor.image_processor , batch_size=_lowercase , framework=_lowercase ) return {**text_input_dict, **image_input_dict} @property def lowerCAmelCase__(self ): '''simple docstring''' return 14
707
"""simple docstring""" import importlib import torch import yaml from omegaconf import OmegaConf from taming.models.vqgan import VQModel def __magic_name__ ( _lowerCamelCase : int , _lowerCamelCase : Optional[Any]=False ): __a : Dict = OmegaConf.load(_lowerCamelCase ) if display: print(yaml.dump(OmegaConf.to_container(_lowerCamelCase ) ) ) return config def __magic_name__ ( _lowerCamelCase : str , _lowerCamelCase : Union[str, Any]=None , _lowerCamelCase : int=None ): if conf_path is None: __a : str = """./model_checkpoints/vqgan_only.yaml""" __a : List[Any] = load_config(_lowerCamelCase , display=_lowerCamelCase ) __a : Dict = VQModel(**config.model.params ) if ckpt_path is None: __a : List[Any] = """./model_checkpoints/vqgan_only.pt""" __a : Tuple = torch.load(_lowerCamelCase , map_location=_lowerCamelCase ) if ".ckpt" in ckpt_path: __a : List[str] = sd["""state_dict"""] model.load_state_dict(_lowerCamelCase , strict=_lowerCamelCase ) model.to(_lowerCamelCase ) del sd return model def __magic_name__ ( _lowerCamelCase : Optional[int] , _lowerCamelCase : List[str] ): __a , __a , __a : Tuple = model.encode(_lowerCamelCase ) print(F'''VQGAN --- {model.__class__.__name__}: latent shape: {z.shape[2:]}''' ) __a : Union[str, Any] = model.decode(_lowerCamelCase ) return xrec def __magic_name__ ( _lowerCamelCase : Optional[int] , _lowerCamelCase : Union[str, Any]=False ): __a , __a : Optional[Any] = string.rsplit(""".""" , 1 ) if reload: __a : Optional[Any] = importlib.import_module(_lowerCamelCase ) importlib.reload(_lowerCamelCase ) return getattr(importlib.import_module(_lowerCamelCase , package=_lowerCamelCase ) , cls ) def __magic_name__ ( _lowerCamelCase : Any ): if "target" not in config: raise KeyError("""Expected key `target` to instantiate.""" ) return get_obj_from_str(config["""target"""] )(**config.get("""params""" , {} ) ) def __magic_name__ ( _lowerCamelCase : Any , _lowerCamelCase : Dict , _lowerCamelCase : int=True , _lowerCamelCase : int=True ): __a : Union[str, Any] = instantiate_from_config(_lowerCamelCase ) if sd is not None: model.load_state_dict(_lowerCamelCase ) if gpu: model.cuda() if eval_mode: model.eval() return {"model": model} def __magic_name__ ( _lowerCamelCase : Any , _lowerCamelCase : Union[str, Any] , _lowerCamelCase : str , _lowerCamelCase : int ): # load the specified checkpoint if ckpt: __a : List[str] = torch.load(_lowerCamelCase , map_location="""cpu""" ) __a : Any = pl_sd["""global_step"""] print(F'''loaded model from global step {global_step}.''' ) else: __a : List[Any] = {"""state_dict""": None} __a : Any = None __a : Union[str, Any] = load_model_from_config(config.model , pl_sd["""state_dict"""] , gpu=_lowerCamelCase , eval_mode=_lowerCamelCase )["""model"""] return model, global_step
63
0
"""simple docstring""" from math import pi, sqrt def __magic_name__ ( _lowerCamelCase : float ): if num <= 0: raise ValueError("""math domain error""" ) if num > 1_7_1.5: raise OverflowError("""math range error""" ) elif num - int(_lowerCamelCase ) not in (0, 0.5): raise NotImplementedError("""num must be an integer or a half-integer""" ) elif num == 0.5: return sqrt(_lowerCamelCase ) else: return 1.0 if num == 1 else (num - 1) * gamma(num - 1 ) def __magic_name__ ( ): assert gamma(0.5 ) == sqrt(_lowerCamelCase ) assert gamma(1 ) == 1.0 assert gamma(2 ) == 1.0 if __name__ == "__main__": from doctest import testmod testmod() lowercase__ = 1.0 while num: lowercase__ = float(input("Gamma of: ")) print(f'gamma({num}) = {gamma(num)}') print("\nEnter 0 to exit...")
708
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) lowercase__ = { "configuration_llama": ["LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP", "LlamaConfig"], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = ["LlamaTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = ["LlamaTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ "LlamaForCausalLM", "LlamaModel", "LlamaPreTrainedModel", "LlamaForSequenceClassification", ] if TYPE_CHECKING: from .configuration_llama import LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP, LlamaConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama import LlamaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama_fast import LlamaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_llama import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaPreTrainedModel else: import sys lowercase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
63
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) lowercase__ = {"configuration_reformer": ["REFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "ReformerConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = ["ReformerTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = ["ReformerTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ "REFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "ReformerAttention", "ReformerForMaskedLM", "ReformerForQuestionAnswering", "ReformerForSequenceClassification", "ReformerLayer", "ReformerModel", "ReformerModelWithLMHead", "ReformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_reformer import REFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, ReformerConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_reformer import ReformerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_reformer_fast import ReformerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_reformer import ( REFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ReformerAttention, ReformerForMaskedLM, ReformerForQuestionAnswering, ReformerForSequenceClassification, ReformerLayer, ReformerModel, ReformerModelWithLMHead, ReformerPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
709
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { "microsoft/unispeech-large-1500h-cv": ( "https://huggingface.co/microsoft/unispeech-large-1500h-cv/resolve/main/config.json" ), # See all UniSpeech models at https://huggingface.co/models?filter=unispeech } class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = "unispeech" def __init__(self , _lowercase=32 , _lowercase=768 , _lowercase=12 , _lowercase=12 , _lowercase=3072 , _lowercase="gelu" , _lowercase=0.1 , _lowercase=0.1 , _lowercase=0.1 , _lowercase=0.0 , _lowercase=0.0 , _lowercase=0.1 , _lowercase=0.1 , _lowercase=0.02 , _lowercase=1e-5 , _lowercase="group" , _lowercase="gelu" , _lowercase=(512, 512, 512, 512, 512, 512, 512) , _lowercase=(5, 2, 2, 2, 2, 2, 2) , _lowercase=(10, 3, 3, 3, 3, 2, 2) , _lowercase=False , _lowercase=128 , _lowercase=16 , _lowercase=False , _lowercase=True , _lowercase=0.05 , _lowercase=10 , _lowercase=2 , _lowercase=0.0 , _lowercase=10 , _lowercase=0 , _lowercase=320 , _lowercase=2 , _lowercase=0.1 , _lowercase=100 , _lowercase=256 , _lowercase=256 , _lowercase=0.1 , _lowercase="mean" , _lowercase=False , _lowercase=False , _lowercase=256 , _lowercase=80 , _lowercase=0 , _lowercase=1 , _lowercase=2 , _lowercase=0.5 , **_lowercase , ): '''simple docstring''' super().__init__(**_lowercase , pad_token_id=_lowercase , bos_token_id=_lowercase , eos_token_id=_lowercase ) __a : Union[str, Any] = hidden_size __a : Any = feat_extract_norm __a : Union[str, Any] = feat_extract_activation __a : Tuple = list(_lowercase ) __a : Dict = list(_lowercase ) __a : List[Any] = list(_lowercase ) __a : List[Any] = conv_bias __a : Optional[Any] = num_conv_pos_embeddings __a : Union[str, Any] = num_conv_pos_embedding_groups __a : Dict = len(self.conv_dim ) __a : Dict = num_hidden_layers __a : Union[str, Any] = intermediate_size __a : List[str] = hidden_act __a : int = num_attention_heads __a : int = hidden_dropout __a : Any = attention_dropout __a : List[Any] = activation_dropout __a : List[Any] = feat_proj_dropout __a : Union[str, Any] = final_dropout __a : str = layerdrop __a : Dict = layer_norm_eps __a : Dict = initializer_range __a : Union[str, Any] = num_ctc_classes __a : List[Any] = vocab_size __a : Any = do_stable_layer_norm __a : List[str] = use_weighted_layer_sum __a : List[str] = classifier_proj_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==""" """ `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =""" F''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' F''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __a : Dict = apply_spec_augment __a : Union[str, Any] = mask_time_prob __a : List[str] = mask_time_length __a : Dict = mask_time_min_masks __a : List[Any] = mask_feature_prob __a : Tuple = mask_feature_length __a : int = mask_feature_min_masks # parameters for pretraining with codevector quantized representations __a : List[Any] = num_codevectors_per_group __a : Union[str, Any] = num_codevector_groups __a : List[Any] = contrastive_logits_temperature __a : Any = feat_quantizer_dropout __a : Optional[int] = num_negatives __a : List[str] = codevector_dim __a : List[Any] = proj_codevector_dim __a : Tuple = diversity_loss_weight # ctc loss __a : Any = ctc_loss_reduction __a : List[str] = ctc_zero_infinity # pretraining loss __a : Tuple = replace_prob @property def lowerCAmelCase__(self ): '''simple docstring''' return functools.reduce(operator.mul , self.conv_stride , 1 )
63
0
"""simple docstring""" from __future__ import annotations from collections import deque from collections.abc import Iterator from dataclasses import dataclass @dataclass class SCREAMING_SNAKE_CASE__ : _lowerCAmelCase = 4_2 _lowerCAmelCase = 4_2 class SCREAMING_SNAKE_CASE__ : def __init__(self , _lowercase ): '''simple docstring''' __a : list[list[Edge]] = [[] for _ in range(_lowercase )] __a : Dict = size def __getitem__(self , _lowercase ): '''simple docstring''' return iter(self._graph[vertex] ) @property def lowerCAmelCase__(self ): '''simple docstring''' return self._size def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase ): '''simple docstring''' if weight not in (0, 1): raise ValueError("""Edge weight must be either 0 or 1.""" ) if to_vertex < 0 or to_vertex >= self.size: raise ValueError("""Vertex indexes must be in [0; size).""" ) self._graph[from_vertex].append(Edge(_lowercase , _lowercase ) ) def lowerCAmelCase__(self , _lowercase , _lowercase ): '''simple docstring''' __a : Optional[int] = deque([start_vertex] ) __a : list[int | None] = [None] * self.size __a : Tuple = 0 while queue: __a : Union[str, Any] = queue.popleft() __a : Any = distances[current_vertex] if current_distance is None: continue for edge in self[current_vertex]: __a : Union[str, Any] = current_distance + edge.weight __a : Dict = distances[edge.destination_vertex] if ( isinstance(_lowercase , _lowercase ) and new_distance >= dest_vertex_distance ): continue __a : Optional[Any] = new_distance if edge.weight == 0: queue.appendleft(edge.destination_vertex ) else: queue.append(edge.destination_vertex ) if distances[finish_vertex] is None: raise ValueError("""No path from start_vertex to finish_vertex.""" ) return distances[finish_vertex] if __name__ == "__main__": import doctest doctest.testmod()
710
"""simple docstring""" import inspect import unittest from typing import List import numpy as np from transformers import EfficientFormerConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, ) from transformers.models.efficientformer.modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_vision_available(): from PIL import Image from transformers import EfficientFormerImageProcessor class SCREAMING_SNAKE_CASE__ : def __init__(self , _lowercase , _lowercase = 13 , _lowercase = 64 , _lowercase = 2 , _lowercase = 3 , _lowercase = 3 , _lowercase = True , _lowercase = True , _lowercase = 128 , _lowercase=[16, 32, 64, 128] , _lowercase = 7 , _lowercase = 4 , _lowercase = 37 , _lowercase = "gelu" , _lowercase = 0.1 , _lowercase = 0.1 , _lowercase = 10 , _lowercase = 0.02 , _lowercase = 2 , _lowercase = 1 , _lowercase = 128 , _lowercase = [2, 2, 2, 2] , _lowercase = 2 , _lowercase = 2 , ): '''simple docstring''' __a : str = parent __a : List[Any] = batch_size __a : int = image_size __a : Tuple = patch_size __a : str = num_channels __a : Union[str, Any] = is_training __a : List[Any] = use_labels __a : int = hidden_size __a : Optional[Any] = num_hidden_layers __a : List[Any] = num_attention_heads __a : Dict = intermediate_size __a : str = hidden_act __a : Dict = hidden_dropout_prob __a : str = attention_probs_dropout_prob __a : Optional[int] = type_sequence_label_size __a : Dict = initializer_range __a : Dict = encoder_stride __a : int = num_attention_outputs __a : List[Any] = embed_dim __a : Optional[Any] = embed_dim + 1 __a : Optional[Any] = resolution __a : Optional[Any] = depths __a : Union[str, Any] = hidden_sizes __a : List[str] = dim __a : Any = mlp_expansion_ratio def lowerCAmelCase__(self ): '''simple docstring''' __a : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __a : str = None if self.use_labels: __a : List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __a : List[str] = self.get_config() return config, pixel_values, labels def lowerCAmelCase__(self ): '''simple docstring''' return EfficientFormerConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowercase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , resolution=self.resolution , depths=self.depths , hidden_sizes=self.hidden_sizes , dim=self.dim , mlp_expansion_ratio=self.mlp_expansion_ratio , ) def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase ): '''simple docstring''' __a : Optional[Any] = TFEfficientFormerModel(config=_lowercase ) __a : List[Any] = model(_lowercase , training=_lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase ): '''simple docstring''' __a : Optional[Any] = self.type_sequence_label_size __a : Any = TFEfficientFormerForImageClassification(_lowercase ) __a : Union[str, Any] = model(_lowercase , labels=_lowercase , training=_lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __a : Optional[Any] = 1 __a : int = TFEfficientFormerForImageClassification(_lowercase ) __a : List[str] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __a : str = model(_lowercase , labels=_lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Any = self.prepare_config_and_inputs() __a , __a , __a : Tuple = config_and_inputs __a : Tuple = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class SCREAMING_SNAKE_CASE__ ( __snake_case , __snake_case , unittest.TestCase ): _lowerCAmelCase = ( ( TFEfficientFormerModel, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerForImageClassification, ) if is_tf_available() else () ) _lowerCAmelCase = ( { "feature-extraction": TFEfficientFormerModel, "image-classification": ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, ), } if is_tf_available() else {} ) _lowerCAmelCase = False _lowerCAmelCase = False _lowerCAmelCase = False _lowerCAmelCase = False _lowerCAmelCase = False def lowerCAmelCase__(self ): '''simple docstring''' __a : Tuple = TFEfficientFormerModelTester(self ) __a : Any = ConfigTester( self , config_class=_lowercase , has_text_modality=_lowercase , hidden_size=37 ) def lowerCAmelCase__(self ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="""EfficientFormer does not use inputs_embeds""" ) def lowerCAmelCase__(self ): '''simple docstring''' pass @unittest.skip(reason="""EfficientFormer does not support input and output embeddings""" ) def lowerCAmelCase__(self ): '''simple docstring''' pass def lowerCAmelCase__(self ): '''simple docstring''' __a , __a : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a : Dict = model_class(_lowercase ) __a : Optional[Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __a : Optional[Any] = [*signature.parameters.keys()] __a : Union[str, Any] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' def check_hidden_states_output(_lowercase , _lowercase , _lowercase ): __a : Tuple = model_class(_lowercase ) __a : int = model(**self._prepare_for_class(_lowercase , _lowercase ) , training=_lowercase ) __a : Tuple = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __a : str = getattr( self.model_tester , """expected_num_hidden_layers""" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(_lowercase ) , _lowercase ) if hasattr(self.model_tester , """encoder_seq_length""" ): __a : Any = self.model_tester.encoder_seq_length if hasattr(self.model_tester , """chunk_length""" ) and self.model_tester.chunk_length > 1: __a : int = seq_length * self.model_tester.chunk_length else: __a : Any = self.model_tester.seq_length self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) if config.is_encoder_decoder: __a : Optional[int] = outputs.decoder_hidden_states self.asseretIsInstance(_lowercase , (list, tuple) ) self.assertEqual(len(_lowercase ) , _lowercase ) __a : Any = getattr(self.model_tester , """seq_length""" , _lowercase ) __a : List[Any] = getattr(self.model_tester , """decoder_seq_length""" , _lowercase ) self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [decoder_seq_length, self.model_tester.hidden_size] , ) __a , __a : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a : Dict = True check_hidden_states_output(_lowercase , _lowercase , _lowercase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __a : int = True check_hidden_states_output(_lowercase , _lowercase , _lowercase ) def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase=False ): '''simple docstring''' __a : Any = super()._prepare_for_class(_lowercase , _lowercase , return_labels=_lowercase ) if return_labels: if model_class.__name__ == "TFEfficientFormerForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowercase ) @unittest.skip(reason="""EfficientFormer does not implement masked image modeling yet""" ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowercase ) @slow def lowerCAmelCase__(self ): '''simple docstring''' for model_name in TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a : Union[str, Any] = TFEfficientFormerModel.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' __a , __a : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() __a : int = True __a : Optional[int] = getattr(self.model_tester , """seq_length""" , _lowercase ) __a : Dict = getattr(self.model_tester , """encoder_seq_length""" , _lowercase ) __a : Dict = getattr(self.model_tester , """key_length""" , _lowercase ) __a : int = getattr(self.model_tester , """chunk_length""" , _lowercase ) if chunk_length is not None and hasattr(self.model_tester , """num_hashes""" ): __a : List[str] = encoder_seq_length * self.model_tester.num_hashes for model_class in self.all_model_classes: __a : List[Any] = True __a : Tuple = False __a : List[Any] = True __a : int = model_class(_lowercase ) __a : List[Any] = model(**self._prepare_for_class(_lowercase , _lowercase ) , training=_lowercase ) __a : Dict = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(_lowercase ) , self.model_tester.num_attention_outputs ) # check that output_attentions also work using config del inputs_dict["output_attentions"] __a : Optional[Any] = True __a : List[str] = model_class(_lowercase ) __a : Dict = model(**self._prepare_for_class(_lowercase , _lowercase ) , training=_lowercase ) __a : int = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(_lowercase ) , self.model_tester.num_attention_outputs ) if chunk_length is not None: self.assertListEqual( list(attentions[0].shape[-4:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, chunk_length, encoder_key_length] , ) else: self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, encoder_key_length] , ) def lowerCAmelCase__(self ): '''simple docstring''' __a , __a : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # Prepare our model __a : Dict = model_class(_lowercase ) # These are maximally general inputs for the model, with multiple None dimensions # Hopefully this will catch any conditionals that fail for flexible shapes __a : Optional[Any] = { key: tf.keras.Input(shape=val.shape[1:] , dtype=val.dtype , name=_lowercase ) for key, val in model.input_signature.items() if key in model.dummy_inputs } __a : Optional[Any] = model(_lowercase ) self.assertTrue(outputs_dict is not None ) def __magic_name__ ( ): __a : int = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @cached_property def lowerCAmelCase__(self ): '''simple docstring''' return ( EfficientFormerImageProcessor.from_pretrained("""snap-research/efficientformer-l1-300""" ) if is_vision_available() else None ) @slow def lowerCAmelCase__(self ): '''simple docstring''' __a : str = TFEfficientFormerForImageClassification.from_pretrained("""snap-research/efficientformer-l1-300""" ) __a : Optional[Any] = self.default_image_processor __a : List[str] = prepare_img() __a : int = image_processor(images=_lowercase , return_tensors="""tf""" ) # forward pass __a : Optional[Any] = model(**_lowercase , training=_lowercase ) # verify the logits __a : str = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowercase ) __a : Dict = tf.constant([-0.0555, 0.4825, -0.0852] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , _lowercase , atol=1e-4 ) ) @slow def lowerCAmelCase__(self ): '''simple docstring''' __a : Any = TFEfficientFormerForImageClassificationWithTeacher.from_pretrained( """snap-research/efficientformer-l1-300""" ) __a : Any = self.default_image_processor __a : str = prepare_img() __a : str = image_processor(images=_lowercase , return_tensors="""tf""" ) # forward pass __a : List[Any] = model(**_lowercase , training=_lowercase ) # verify the logits __a : int = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowercase ) __a : List[str] = tf.constant([-0.1312, 0.4353, -1.0499] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , _lowercase , atol=1e-4 ) )
63
0
"""simple docstring""" from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_pt_objects import * # noqa F403 else: from .scheduling_consistency_models import CMStochasticIterativeScheduler from .scheduling_ddim import DDIMScheduler from .scheduling_ddim_inverse import DDIMInverseScheduler from .scheduling_ddim_parallel import DDIMParallelScheduler from .scheduling_ddpm import DDPMScheduler from .scheduling_ddpm_parallel import DDPMParallelScheduler from .scheduling_deis_multistep import DEISMultistepScheduler from .scheduling_dpmsolver_multistep import DPMSolverMultistepScheduler from .scheduling_dpmsolver_multistep_inverse import DPMSolverMultistepInverseScheduler from .scheduling_dpmsolver_singlestep import DPMSolverSinglestepScheduler from .scheduling_euler_ancestral_discrete import EulerAncestralDiscreteScheduler from .scheduling_euler_discrete import EulerDiscreteScheduler from .scheduling_heun_discrete import HeunDiscreteScheduler from .scheduling_ipndm import IPNDMScheduler from .scheduling_k_dpm_2_ancestral_discrete import KDPMaAncestralDiscreteScheduler from .scheduling_k_dpm_2_discrete import KDPMaDiscreteScheduler from .scheduling_karras_ve import KarrasVeScheduler from .scheduling_pndm import PNDMScheduler from .scheduling_repaint import RePaintScheduler from .scheduling_sde_ve import ScoreSdeVeScheduler from .scheduling_sde_vp import ScoreSdeVpScheduler from .scheduling_unclip import UnCLIPScheduler from .scheduling_unipc_multistep import UniPCMultistepScheduler from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin from .scheduling_vq_diffusion import VQDiffusionScheduler try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_flax_objects import * # noqa F403 else: from .scheduling_ddim_flax import FlaxDDIMScheduler from .scheduling_ddpm_flax import FlaxDDPMScheduler from .scheduling_dpmsolver_multistep_flax import FlaxDPMSolverMultistepScheduler from .scheduling_karras_ve_flax import FlaxKarrasVeScheduler from .scheduling_lms_discrete_flax import FlaxLMSDiscreteScheduler from .scheduling_pndm_flax import FlaxPNDMScheduler from .scheduling_sde_ve_flax import FlaxScoreSdeVeScheduler from .scheduling_utils_flax import ( FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, broadcast_to_shape_from_left, ) try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .scheduling_lms_discrete import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .scheduling_dpmsolver_sde import DPMSolverSDEScheduler
711
"""simple docstring""" from typing import Callable, Dict, Optional, Tuple import torch from torch import nn from torch.distributions import ( AffineTransform, Distribution, Independent, NegativeBinomial, Normal, StudentT, TransformedDistribution, ) class SCREAMING_SNAKE_CASE__ ( __snake_case ): def __init__(self , _lowercase , _lowercase=None , _lowercase=None , _lowercase=0 ): '''simple docstring''' __a : Any = 1.0 if scale is None else scale __a : str = 0.0 if loc is None else loc super().__init__(_lowercase , [AffineTransform(loc=self.loc , scale=self.scale , event_dim=_lowercase )] ) @property def lowerCAmelCase__(self ): '''simple docstring''' return self.base_dist.mean * self.scale + self.loc @property def lowerCAmelCase__(self ): '''simple docstring''' return self.base_dist.variance * self.scale**2 @property def lowerCAmelCase__(self ): '''simple docstring''' return self.variance.sqrt() class SCREAMING_SNAKE_CASE__ ( nn.Module ): def __init__(self , _lowercase , _lowercase , _lowercase , **_lowercase ): '''simple docstring''' super().__init__(**_lowercase ) __a : str = args_dim __a : List[Any] = nn.ModuleList([nn.Linear(_lowercase , _lowercase ) for dim in args_dim.values()] ) __a : Dict = domain_map def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' __a : List[Any] = [proj(_lowercase ) for proj in self.proj] return self.domain_map(*_lowercase ) class SCREAMING_SNAKE_CASE__ ( nn.Module ): def __init__(self , _lowercase ): '''simple docstring''' super().__init__() __a : Optional[int] = function def lowerCAmelCase__(self , _lowercase , *_lowercase ): '''simple docstring''' return self.function(_lowercase , *_lowercase ) class SCREAMING_SNAKE_CASE__ : _lowerCAmelCase = 42 _lowerCAmelCase = 42 _lowerCAmelCase = 42 def __init__(self , _lowercase = 1 ): '''simple docstring''' __a : Optional[int] = dim __a : str = {k: dim * self.args_dim[k] for k in self.args_dim} def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' if self.dim == 1: return self.distribution_class(*_lowercase ) else: return Independent(self.distribution_class(*_lowercase ) , 1 ) def lowerCAmelCase__(self , _lowercase , _lowercase = None , _lowercase = None , ): '''simple docstring''' __a : Tuple = self._base_distribution(_lowercase ) if loc is None and scale is None: return distr else: return AffineTransformed(_lowercase , loc=_lowercase , scale=_lowercase , event_dim=self.event_dim ) @property def lowerCAmelCase__(self ): '''simple docstring''' return () if self.dim == 1 else (self.dim,) @property def lowerCAmelCase__(self ): '''simple docstring''' return len(self.event_shape ) @property def lowerCAmelCase__(self ): '''simple docstring''' return 0.0 def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' return ParameterProjection( in_features=_lowercase , args_dim=self.args_dim , domain_map=LambdaLayer(self.domain_map ) , ) def lowerCAmelCase__(self , *_lowercase ): '''simple docstring''' raise NotImplementedError() @staticmethod def lowerCAmelCase__(_lowercase ): '''simple docstring''' return (x + torch.sqrt(torch.square(_lowercase ) + 4.0 )) / 2.0 class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = {"df": 1, "loc": 1, "scale": 1} _lowerCAmelCase = StudentT @classmethod def lowerCAmelCase__(cls , _lowercase , _lowercase , _lowercase ): '''simple docstring''' __a : int = cls.squareplus(_lowercase ).clamp_min(torch.finfo(scale.dtype ).eps ) __a : Optional[Any] = 2.0 + cls.squareplus(_lowercase ) return df.squeeze(-1 ), loc.squeeze(-1 ), scale.squeeze(-1 ) class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = {"loc": 1, "scale": 1} _lowerCAmelCase = Normal @classmethod def lowerCAmelCase__(cls , _lowercase , _lowercase ): '''simple docstring''' __a : str = cls.squareplus(_lowercase ).clamp_min(torch.finfo(scale.dtype ).eps ) return loc.squeeze(-1 ), scale.squeeze(-1 ) class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = {"total_count": 1, "logits": 1} _lowerCAmelCase = NegativeBinomial @classmethod def lowerCAmelCase__(cls , _lowercase , _lowercase ): '''simple docstring''' __a : Union[str, Any] = cls.squareplus(_lowercase ) return total_count.squeeze(-1 ), logits.squeeze(-1 ) def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' __a , __a : Optional[Any] = distr_args if self.dim == 1: return self.distribution_class(total_count=_lowercase , logits=_lowercase ) else: return Independent(self.distribution_class(total_count=_lowercase , logits=_lowercase ) , 1 ) def lowerCAmelCase__(self , _lowercase , _lowercase = None , _lowercase = None ): '''simple docstring''' __a , __a : List[Any] = distr_args if scale is not None: # See scaling property of Gamma. logits += scale.log() return self._base_distribution((total_count, logits) )
63
0
"""simple docstring""" import unittest import numpy as np import torch from torch import nn from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import KandinskyVaaPriorPipeline, PriorTransformer, UnCLIPScheduler from diffusers.utils import torch_device from diffusers.utils.testing_utils import enable_full_determinism, skip_mps from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class SCREAMING_SNAKE_CASE__ ( __snake_case , unittest.TestCase ): _lowerCAmelCase = KandinskyVaaPriorPipeline _lowerCAmelCase = ["prompt"] _lowerCAmelCase = ["prompt", "negative_prompt"] _lowerCAmelCase = [ "num_images_per_prompt", "generator", "num_inference_steps", "latents", "negative_prompt", "guidance_scale", "output_type", "return_dict", ] _lowerCAmelCase = False @property def lowerCAmelCase__(self ): '''simple docstring''' return 32 @property def lowerCAmelCase__(self ): '''simple docstring''' return 32 @property def lowerCAmelCase__(self ): '''simple docstring''' return self.time_input_dim @property def lowerCAmelCase__(self ): '''simple docstring''' return self.time_input_dim * 4 @property def lowerCAmelCase__(self ): '''simple docstring''' return 100 @property def lowerCAmelCase__(self ): '''simple docstring''' __a : List[str] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) return tokenizer @property def lowerCAmelCase__(self ): '''simple docstring''' torch.manual_seed(0 ) __a : str = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModelWithProjection(_lowercase ) @property def lowerCAmelCase__(self ): '''simple docstring''' torch.manual_seed(0 ) __a : Dict = { """num_attention_heads""": 2, """attention_head_dim""": 12, """embedding_dim""": self.text_embedder_hidden_size, """num_layers""": 1, } __a : Tuple = PriorTransformer(**_lowercase ) # clip_std and clip_mean is initialized to be 0 so PriorTransformer.post_process_latents will always return 0 - set clip_std to be 1 so it won't return 0 __a : int = nn.Parameter(torch.ones(model.clip_std.shape ) ) return model @property def lowerCAmelCase__(self ): '''simple docstring''' torch.manual_seed(0 ) __a : List[str] = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , image_size=224 , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_channels=3 , num_hidden_layers=5 , patch_size=14 , ) __a : Optional[Any] = CLIPVisionModelWithProjection(_lowercase ) return model @property def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[Any] = CLIPImageProcessor( crop_size=224 , do_center_crop=_lowercase , do_normalize=_lowercase , do_resize=_lowercase , image_mean=[0.4814_5466, 0.457_8275, 0.4082_1073] , image_std=[0.2686_2954, 0.2613_0258, 0.2757_7711] , resample=3 , size=224 , ) return image_processor def lowerCAmelCase__(self ): '''simple docstring''' __a : Union[str, Any] = self.dummy_prior __a : int = self.dummy_image_encoder __a : Any = self.dummy_text_encoder __a : int = self.dummy_tokenizer __a : Optional[Any] = self.dummy_image_processor __a : List[Any] = UnCLIPScheduler( variance_type="""fixed_small_log""" , prediction_type="""sample""" , num_train_timesteps=1000 , clip_sample=_lowercase , clip_sample_range=10.0 , ) __a : List[Any] = { """prior""": prior, """image_encoder""": image_encoder, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """scheduler""": scheduler, """image_processor""": image_processor, } return components def lowerCAmelCase__(self , _lowercase , _lowercase=0 ): '''simple docstring''' if str(_lowercase ).startswith("""mps""" ): __a : Dict = torch.manual_seed(_lowercase ) else: __a : Union[str, Any] = torch.Generator(device=_lowercase ).manual_seed(_lowercase ) __a : Union[str, Any] = { """prompt""": """horse""", """generator""": generator, """guidance_scale""": 4.0, """num_inference_steps""": 2, """output_type""": """np""", } return inputs def lowerCAmelCase__(self ): '''simple docstring''' __a : Union[str, Any] = """cpu""" __a : Union[str, Any] = self.get_dummy_components() __a : Dict = self.pipeline_class(**_lowercase ) __a : Tuple = pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) __a : Optional[int] = pipe(**self.get_dummy_inputs(_lowercase ) ) __a : str = output.image_embeds __a : Any = pipe( **self.get_dummy_inputs(_lowercase ) , return_dict=_lowercase , )[0] __a : List[Any] = image[0, -10:] __a : List[Any] = image_from_tuple[0, -10:] assert image.shape == (1, 32) __a : Optional[Any] = np.array( [-0.0532, 1.7120, 0.3656, -1.0852, -0.8946, -1.1756, 0.4348, 0.2482, 0.5146, -0.1156] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @skip_mps def lowerCAmelCase__(self ): '''simple docstring''' __a : Any = torch_device == """cpu""" __a : Any = True __a : Any = False self._test_inference_batch_single_identical( test_max_difference=_lowercase , relax_max_difference=_lowercase , test_mean_pixel_difference=_lowercase , ) @skip_mps def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[int] = torch_device == """cpu""" __a : Union[str, Any] = False self._test_attention_slicing_forward_pass( test_max_difference=_lowercase , test_mean_pixel_difference=_lowercase , )
712
"""simple docstring""" import unittest import numpy as np import torch from torch import nn from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import KandinskyVaaPriorPipeline, PriorTransformer, UnCLIPScheduler from diffusers.utils import torch_device from diffusers.utils.testing_utils import enable_full_determinism, skip_mps from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class SCREAMING_SNAKE_CASE__ ( __snake_case , unittest.TestCase ): _lowerCAmelCase = KandinskyVaaPriorPipeline _lowerCAmelCase = ["prompt"] _lowerCAmelCase = ["prompt", "negative_prompt"] _lowerCAmelCase = [ "num_images_per_prompt", "generator", "num_inference_steps", "latents", "negative_prompt", "guidance_scale", "output_type", "return_dict", ] _lowerCAmelCase = False @property def lowerCAmelCase__(self ): '''simple docstring''' return 32 @property def lowerCAmelCase__(self ): '''simple docstring''' return 32 @property def lowerCAmelCase__(self ): '''simple docstring''' return self.time_input_dim @property def lowerCAmelCase__(self ): '''simple docstring''' return self.time_input_dim * 4 @property def lowerCAmelCase__(self ): '''simple docstring''' return 100 @property def lowerCAmelCase__(self ): '''simple docstring''' __a : List[str] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) return tokenizer @property def lowerCAmelCase__(self ): '''simple docstring''' torch.manual_seed(0 ) __a : str = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModelWithProjection(_lowercase ) @property def lowerCAmelCase__(self ): '''simple docstring''' torch.manual_seed(0 ) __a : Dict = { """num_attention_heads""": 2, """attention_head_dim""": 12, """embedding_dim""": self.text_embedder_hidden_size, """num_layers""": 1, } __a : Tuple = PriorTransformer(**_lowercase ) # clip_std and clip_mean is initialized to be 0 so PriorTransformer.post_process_latents will always return 0 - set clip_std to be 1 so it won't return 0 __a : int = nn.Parameter(torch.ones(model.clip_std.shape ) ) return model @property def lowerCAmelCase__(self ): '''simple docstring''' torch.manual_seed(0 ) __a : List[str] = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , image_size=224 , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_channels=3 , num_hidden_layers=5 , patch_size=14 , ) __a : Optional[Any] = CLIPVisionModelWithProjection(_lowercase ) return model @property def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[Any] = CLIPImageProcessor( crop_size=224 , do_center_crop=_lowercase , do_normalize=_lowercase , do_resize=_lowercase , image_mean=[0.4814_5466, 0.457_8275, 0.4082_1073] , image_std=[0.2686_2954, 0.2613_0258, 0.2757_7711] , resample=3 , size=224 , ) return image_processor def lowerCAmelCase__(self ): '''simple docstring''' __a : Union[str, Any] = self.dummy_prior __a : int = self.dummy_image_encoder __a : Any = self.dummy_text_encoder __a : int = self.dummy_tokenizer __a : Optional[Any] = self.dummy_image_processor __a : List[Any] = UnCLIPScheduler( variance_type="""fixed_small_log""" , prediction_type="""sample""" , num_train_timesteps=1000 , clip_sample=_lowercase , clip_sample_range=10.0 , ) __a : List[Any] = { """prior""": prior, """image_encoder""": image_encoder, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """scheduler""": scheduler, """image_processor""": image_processor, } return components def lowerCAmelCase__(self , _lowercase , _lowercase=0 ): '''simple docstring''' if str(_lowercase ).startswith("""mps""" ): __a : Dict = torch.manual_seed(_lowercase ) else: __a : Union[str, Any] = torch.Generator(device=_lowercase ).manual_seed(_lowercase ) __a : Union[str, Any] = { """prompt""": """horse""", """generator""": generator, """guidance_scale""": 4.0, """num_inference_steps""": 2, """output_type""": """np""", } return inputs def lowerCAmelCase__(self ): '''simple docstring''' __a : Union[str, Any] = """cpu""" __a : Union[str, Any] = self.get_dummy_components() __a : Dict = self.pipeline_class(**_lowercase ) __a : Tuple = pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) __a : Optional[int] = pipe(**self.get_dummy_inputs(_lowercase ) ) __a : str = output.image_embeds __a : Any = pipe( **self.get_dummy_inputs(_lowercase ) , return_dict=_lowercase , )[0] __a : List[Any] = image[0, -10:] __a : List[Any] = image_from_tuple[0, -10:] assert image.shape == (1, 32) __a : Optional[Any] = np.array( [-0.0532, 1.7120, 0.3656, -1.0852, -0.8946, -1.1756, 0.4348, 0.2482, 0.5146, -0.1156] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @skip_mps def lowerCAmelCase__(self ): '''simple docstring''' __a : Any = torch_device == """cpu""" __a : Any = True __a : Any = False self._test_inference_batch_single_identical( test_max_difference=_lowercase , relax_max_difference=_lowercase , test_mean_pixel_difference=_lowercase , ) @skip_mps def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[int] = torch_device == """cpu""" __a : Union[str, Any] = False self._test_attention_slicing_forward_pass( test_max_difference=_lowercase , test_mean_pixel_difference=_lowercase , )
63
0
"""simple docstring""" from typing import Optional from urllib.parse import quote import huggingface_hub as hfh from packaging import version def __magic_name__ ( _lowerCamelCase : str , _lowerCamelCase : str , _lowerCamelCase : Optional[str] = None ): if version.parse(hfh.__version__ ).release < version.parse("""0.11.0""" ).release: # old versions of hfh don't url-encode the file path __a : List[Any] = quote(_lowerCamelCase ) return hfh.hf_hub_url(_lowerCamelCase , _lowerCamelCase , repo_type="""dataset""" , revision=_lowerCamelCase )
713
"""simple docstring""" import json import os import unittest from transformers import BatchEncoding, LEDTokenizer, LEDTokenizerFast from transformers.models.led.tokenization_led import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class SCREAMING_SNAKE_CASE__ ( __snake_case , unittest.TestCase ): _lowerCAmelCase = LEDTokenizer _lowerCAmelCase = LEDTokenizerFast _lowerCAmelCase = True def lowerCAmelCase__(self ): '''simple docstring''' super().setUp() __a : str = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", ] __a : int = dict(zip(_lowercase , range(len(_lowercase ) ) ) ) __a : Optional[int] = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] __a : List[Any] = {"""unk_token""": """<unk>"""} __a : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) __a : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(_lowercase ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(_lowercase ) ) def lowerCAmelCase__(self , **_lowercase ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **_lowercase ) def lowerCAmelCase__(self , **_lowercase ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **_lowercase ) def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' return "lower newer", "lower newer" @cached_property def lowerCAmelCase__(self ): '''simple docstring''' return LEDTokenizer.from_pretrained("""allenai/led-base-16384""" ) @cached_property def lowerCAmelCase__(self ): '''simple docstring''' return LEDTokenizerFast.from_pretrained("""allenai/led-base-16384""" ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' __a : Any = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] __a : List[str] = [0, 250, 251, 17818, 13, 39186, 1938, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __a : Optional[int] = tokenizer(_lowercase , max_length=len(_lowercase ) , padding=_lowercase , return_tensors="""pt""" ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual((2, 9) , batch.input_ids.shape ) self.assertEqual((2, 9) , batch.attention_mask.shape ) __a : Dict = batch.input_ids.tolist()[0] self.assertListEqual(_lowercase , _lowercase ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' __a : Union[str, Any] = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __a : Tuple = tokenizer(_lowercase , padding=_lowercase , return_tensors="""pt""" ) self.assertIn("""input_ids""" , _lowercase ) self.assertIn("""attention_mask""" , _lowercase ) self.assertNotIn("""labels""" , _lowercase ) self.assertNotIn("""decoder_attention_mask""" , _lowercase ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[Any] = [ """Summary of the text.""", """Another summary.""", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __a : Dict = tokenizer(text_target=_lowercase , max_length=32 , padding="""max_length""" , return_tensors="""pt""" ) self.assertEqual(32 , targets["""input_ids"""].shape[1] ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __a : Optional[int] = tokenizer( ["""I am a small frog""" * 1024, """I am a small frog"""] , padding=_lowercase , truncation=_lowercase , return_tensors="""pt""" ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(batch.input_ids.shape , (2, 5122) ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' __a : Tuple = ["""A long paragraph for summarization."""] __a : Dict = [ """Summary of the text.""", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __a : int = tokenizer(_lowercase , return_tensors="""pt""" ) __a : Dict = tokenizer(text_target=_lowercase , return_tensors="""pt""" ) __a : List[str] = inputs["""input_ids"""] __a : List[Any] = targets["""input_ids"""] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item() ) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item() ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __a : Optional[Any] = ["""Summary of the text.""", """Another summary."""] __a : List[Any] = [[0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, -1, -1]] __a : Union[str, Any] = tokenizer(_lowercase , padding=_lowercase ) __a : Tuple = [[0] * len(_lowercase ) for x in encoded_output["""input_ids"""]] __a : Union[str, Any] = tokenizer.pad(_lowercase ) self.assertSequenceEqual(outputs["""global_attention_mask"""] , _lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' pass def lowerCAmelCase__(self ): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): __a : Dict = self.rust_tokenizer_class.from_pretrained(_lowercase , **_lowercase ) __a : Union[str, Any] = self.tokenizer_class.from_pretrained(_lowercase , **_lowercase ) __a : Union[str, Any] = """A, <mask> AllenNLP sentence.""" __a : Dict = tokenizer_r.encode_plus(_lowercase , add_special_tokens=_lowercase , return_token_type_ids=_lowercase ) __a : Tuple = tokenizer_p.encode_plus(_lowercase , add_special_tokens=_lowercase , return_token_type_ids=_lowercase ) self.assertEqual(sum(tokens_r["""token_type_ids"""] ) , sum(tokens_p["""token_type_ids"""] ) ) self.assertEqual( sum(tokens_r["""attention_mask"""] ) / len(tokens_r["""attention_mask"""] ) , sum(tokens_p["""attention_mask"""] ) / len(tokens_p["""attention_mask"""] ) , ) __a : Tuple = tokenizer_r.convert_ids_to_tokens(tokens_r["""input_ids"""] ) __a : Any = tokenizer_p.convert_ids_to_tokens(tokens_p["""input_ids"""] ) self.assertSequenceEqual(tokens_p["""input_ids"""] , [0, 250, 6, 50264, 3823, 487, 21992, 3645, 4, 2] ) self.assertSequenceEqual(tokens_r["""input_ids"""] , [0, 250, 6, 50264, 3823, 487, 21992, 3645, 4, 2] ) self.assertSequenceEqual( _lowercase , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] ) self.assertSequenceEqual( _lowercase , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] )
63
0
"""simple docstring""" import re import string from collections import Counter import sacrebleu import sacremoses from packaging import version import datasets lowercase__ = "\n@inproceedings{xu-etal-2016-optimizing,\n title = {Optimizing Statistical Machine Translation for Text Simplification},\n authors={Xu, Wei and Napoles, Courtney and Pavlick, Ellie and Chen, Quanze and Callison-Burch, Chris},\n journal = {Transactions of the Association for Computational Linguistics},\n volume = {4},\n year={2016},\n url = {https://www.aclweb.org/anthology/Q16-1029},\n pages = {401--415\n},\n@inproceedings{post-2018-call,\n title = \"A Call for Clarity in Reporting {BLEU} Scores\",\n author = \"Post, Matt\",\n booktitle = \"Proceedings of the Third Conference on Machine Translation: Research Papers\",\n month = oct,\n year = \"2018\",\n address = \"Belgium, Brussels\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/W18-6319\",\n pages = \"186--191\",\n}\n" lowercase__ = "\\nWIKI_SPLIT is the combination of three metrics SARI, EXACT and SACREBLEU\nIt can be used to evaluate the quality of machine-generated texts.\n" lowercase__ = "\nCalculates sari score (between 0 and 100) given a list of source and predicted\nsentences, and a list of lists of reference sentences. It also computes the BLEU score as well as the exact match score.\nArgs:\n sources: list of source sentences where each sentence should be a string.\n predictions: list of predicted sentences where each sentence should be a string.\n references: list of lists of reference sentences where each sentence should be a string.\nReturns:\n sari: sari score\n sacrebleu: sacrebleu score\n exact: exact score\n\nExamples:\n >>> sources=[\"About 95 species are currently accepted .\"]\n >>> predictions=[\"About 95 you now get in .\"]\n >>> references=[[\"About 95 species are currently known .\"]]\n >>> wiki_split = datasets.load_metric(\"wiki_split\")\n >>> results = wiki_split.compute(sources=sources, predictions=predictions, references=references)\n >>> print(results)\n {'sari': 21.805555555555557, 'sacrebleu': 14.535768424205482, 'exact': 0.0}\n" def __magic_name__ ( _lowerCamelCase : Union[str, Any] ): def remove_articles(_lowerCamelCase : List[str] ): __a : str = re.compile(r"""\b(a|an|the)\b""" , re.UNICODE ) return re.sub(_lowerCamelCase , """ """ , _lowerCamelCase ) def white_space_fix(_lowerCamelCase : List[str] ): return " ".join(text.split() ) def remove_punc(_lowerCamelCase : str ): __a : str = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(_lowerCamelCase : Dict ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(_lowerCamelCase ) ) ) ) def __magic_name__ ( _lowerCamelCase : Any , _lowerCamelCase : int ): return int(normalize_answer(_lowerCamelCase ) == normalize_answer(_lowerCamelCase ) ) def __magic_name__ ( _lowerCamelCase : Union[str, Any] , _lowerCamelCase : Tuple ): __a : str = [any(compute_exact(_lowerCamelCase , _lowerCamelCase ) for ref in refs ) for pred, refs in zip(_lowerCamelCase , _lowerCamelCase )] return (sum(_lowerCamelCase ) / len(_lowerCamelCase )) * 1_0_0 def __magic_name__ ( _lowerCamelCase : List[Any] , _lowerCamelCase : int , _lowerCamelCase : Dict , _lowerCamelCase : Optional[Any] ): __a : Tuple = [rgram for rgrams in rgramslist for rgram in rgrams] __a : Union[str, Any] = Counter(_lowerCamelCase ) __a : List[str] = Counter(_lowerCamelCase ) __a : str = Counter() for sgram, scount in sgramcounter.items(): __a : Optional[int] = scount * numref __a : Optional[Any] = Counter(_lowerCamelCase ) __a : Optional[int] = Counter() for cgram, ccount in cgramcounter.items(): __a : Dict = ccount * numref # KEEP __a : List[str] = sgramcounter_rep & cgramcounter_rep __a : Union[str, Any] = keepgramcounter_rep & rgramcounter __a : Union[str, Any] = sgramcounter_rep & rgramcounter __a : int = 0 __a : str = 0 for keepgram in keepgramcountergood_rep: keeptmpscorea += keepgramcountergood_rep[keepgram] / keepgramcounter_rep[keepgram] # Fix an alleged bug [2] in the keep score computation. # keeptmpscore2 += keepgramcountergood_rep[keepgram] / keepgramcounterall_rep[keepgram] keeptmpscorea += keepgramcountergood_rep[keepgram] # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. __a : Optional[Any] = 1 __a : List[str] = 1 if len(_lowerCamelCase ) > 0: __a : List[Any] = keeptmpscorea / len(_lowerCamelCase ) if len(_lowerCamelCase ) > 0: # Fix an alleged bug [2] in the keep score computation. # keepscore_recall = keeptmpscore2 / len(keepgramcounterall_rep) __a : int = keeptmpscorea / sum(keepgramcounterall_rep.values() ) __a : List[Any] = 0 if keepscore_precision > 0 or keepscore_recall > 0: __a : Dict = 2 * keepscore_precision * keepscore_recall / (keepscore_precision + keepscore_recall) # DELETION __a : Optional[Any] = sgramcounter_rep - cgramcounter_rep __a : Tuple = delgramcounter_rep - rgramcounter __a : int = sgramcounter_rep - rgramcounter __a : Any = 0 __a : Union[str, Any] = 0 for delgram in delgramcountergood_rep: deltmpscorea += delgramcountergood_rep[delgram] / delgramcounter_rep[delgram] deltmpscorea += delgramcountergood_rep[delgram] / delgramcounterall_rep[delgram] # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. __a : Any = 1 if len(_lowerCamelCase ) > 0: __a : List[str] = deltmpscorea / len(_lowerCamelCase ) # ADDITION __a : Dict = set(_lowerCamelCase ) - set(_lowerCamelCase ) __a : Optional[int] = set(_lowerCamelCase ) & set(_lowerCamelCase ) __a : List[Any] = set(_lowerCamelCase ) - set(_lowerCamelCase ) __a : Tuple = 0 for addgram in addgramcountergood: addtmpscore += 1 # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. __a : Union[str, Any] = 1 __a : Optional[Any] = 1 if len(_lowerCamelCase ) > 0: __a : Optional[Any] = addtmpscore / len(_lowerCamelCase ) if len(_lowerCamelCase ) > 0: __a : Tuple = addtmpscore / len(_lowerCamelCase ) __a : Tuple = 0 if addscore_precision > 0 or addscore_recall > 0: __a : str = 2 * addscore_precision * addscore_recall / (addscore_precision + addscore_recall) return (keepscore, delscore_precision, addscore) def __magic_name__ ( _lowerCamelCase : Optional[Any] , _lowerCamelCase : Optional[int] , _lowerCamelCase : Optional[int] ): __a : List[str] = len(_lowerCamelCase ) __a : str = ssent.split(""" """ ) __a : str = csent.split(""" """ ) __a : Optional[int] = [] __a : Any = [] __a : Dict = [] __a : Tuple = [] __a : str = [] __a : Union[str, Any] = [] __a : List[Any] = [] __a : Dict = [] __a : Union[str, Any] = [] __a : Tuple = [] for rsent in rsents: __a : str = rsent.split(""" """ ) __a : Optional[int] = [] __a : Optional[int] = [] __a : Union[str, Any] = [] ragramslist.append(_lowerCamelCase ) for i in range(0 , len(_lowerCamelCase ) - 1 ): if i < len(_lowerCamelCase ) - 1: __a : str = ragrams[i] + """ """ + ragrams[i + 1] ragrams.append(_lowerCamelCase ) if i < len(_lowerCamelCase ) - 2: __a : Dict = ragrams[i] + """ """ + ragrams[i + 1] + """ """ + ragrams[i + 2] ragrams.append(_lowerCamelCase ) if i < len(_lowerCamelCase ) - 3: __a : List[str] = ragrams[i] + """ """ + ragrams[i + 1] + """ """ + ragrams[i + 2] + """ """ + ragrams[i + 3] ragrams.append(_lowerCamelCase ) ragramslist.append(_lowerCamelCase ) ragramslist.append(_lowerCamelCase ) ragramslist.append(_lowerCamelCase ) for i in range(0 , len(_lowerCamelCase ) - 1 ): if i < len(_lowerCamelCase ) - 1: __a : List[Any] = sagrams[i] + """ """ + sagrams[i + 1] sagrams.append(_lowerCamelCase ) if i < len(_lowerCamelCase ) - 2: __a : Optional[Any] = sagrams[i] + """ """ + sagrams[i + 1] + """ """ + sagrams[i + 2] sagrams.append(_lowerCamelCase ) if i < len(_lowerCamelCase ) - 3: __a : Tuple = sagrams[i] + """ """ + sagrams[i + 1] + """ """ + sagrams[i + 2] + """ """ + sagrams[i + 3] sagrams.append(_lowerCamelCase ) for i in range(0 , len(_lowerCamelCase ) - 1 ): if i < len(_lowerCamelCase ) - 1: __a : Dict = cagrams[i] + """ """ + cagrams[i + 1] cagrams.append(_lowerCamelCase ) if i < len(_lowerCamelCase ) - 2: __a : Optional[Any] = cagrams[i] + """ """ + cagrams[i + 1] + """ """ + cagrams[i + 2] cagrams.append(_lowerCamelCase ) if i < len(_lowerCamelCase ) - 3: __a : Optional[int] = cagrams[i] + """ """ + cagrams[i + 1] + """ """ + cagrams[i + 2] + """ """ + cagrams[i + 3] cagrams.append(_lowerCamelCase ) (__a) : str = SARIngram(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) (__a) : Optional[Any] = SARIngram(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) (__a) : str = SARIngram(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) (__a) : str = SARIngram(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) __a : Optional[Any] = sum([keepascore, keepascore, keepascore, keepascore] ) / 4 __a : List[str] = sum([delascore, delascore, delascore, delascore] ) / 4 __a : Union[str, Any] = sum([addascore, addascore, addascore, addascore] ) / 4 __a : int = (avgkeepscore + avgdelscore + avgaddscore) / 3 return finalscore def __magic_name__ ( _lowerCamelCase : Tuple , _lowerCamelCase : bool = True , _lowerCamelCase : str = "13a" , _lowerCamelCase : bool = True ): # Normalization is requried for the ASSET dataset (one of the primary # datasets in sentence simplification) to allow using space # to split the sentence. Even though Wiki-Auto and TURK datasets, # do not require normalization, we do it for consistency. # Code adapted from the EASSE library [1] written by the authors of the ASSET dataset. # [1] https://github.com/feralvam/easse/blob/580bba7e1378fc8289c663f864e0487188fe8067/easse/utils/preprocessing.py#L7 if lowercase: __a : Optional[Any] = sentence.lower() if tokenizer in ["13a", "intl"]: if version.parse(sacrebleu.__version__ ).major >= 2: __a : Tuple = sacrebleu.metrics.bleu._get_tokenizer(_lowerCamelCase )()(_lowerCamelCase ) else: __a : List[Any] = sacrebleu.TOKENIZERS[tokenizer]()(_lowerCamelCase ) elif tokenizer == "moses": __a : Tuple = sacremoses.MosesTokenizer().tokenize(_lowerCamelCase , return_str=_lowerCamelCase , escape=_lowerCamelCase ) elif tokenizer == "penn": __a : Tuple = sacremoses.MosesTokenizer().penn_tokenize(_lowerCamelCase , return_str=_lowerCamelCase ) else: __a : str = sentence if not return_str: __a : Tuple = normalized_sent.split() return normalized_sent def __magic_name__ ( _lowerCamelCase : Optional[int] , _lowerCamelCase : int , _lowerCamelCase : List[str] ): if not (len(_lowerCamelCase ) == len(_lowerCamelCase ) == len(_lowerCamelCase )): raise ValueError("""Sources length must match predictions and references lengths.""" ) __a : Optional[Any] = 0 for src, pred, refs in zip(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): sari_score += SARIsent(normalize(_lowerCamelCase ) , normalize(_lowerCamelCase ) , [normalize(_lowerCamelCase ) for sent in refs] ) __a : Optional[int] = sari_score / len(_lowerCamelCase ) return 1_0_0 * sari_score def __magic_name__ ( _lowerCamelCase : Optional[int] , _lowerCamelCase : Union[str, Any] , _lowerCamelCase : Dict="exp" , _lowerCamelCase : Optional[Any]=None , _lowerCamelCase : Optional[Any]=False , _lowerCamelCase : Union[str, Any]=False , _lowerCamelCase : List[str]=False , ): __a : Any = len(references[0] ) if any(len(_lowerCamelCase ) != references_per_prediction for refs in references ): raise ValueError("""Sacrebleu requires the same number of references for each prediction""" ) __a : int = [[refs[i] for refs in references] for i in range(_lowerCamelCase )] __a : Any = sacrebleu.corpus_bleu( _lowerCamelCase , _lowerCamelCase , smooth_method=_lowerCamelCase , smooth_value=_lowerCamelCase , force=_lowerCamelCase , lowercase=_lowerCamelCase , use_effective_order=_lowerCamelCase , ) return output.score @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class SCREAMING_SNAKE_CASE__ ( datasets.Metric ): def lowerCAmelCase__(self ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Sequence(datasets.Value("""string""" , id="""sequence""" ) , id="""references""" ), } ) , codebase_urls=[ """https://github.com/huggingface/transformers/blob/master/src/transformers/data/metrics/squad_metrics.py""", """https://github.com/cocoxu/simplification/blob/master/SARI.py""", """https://github.com/tensorflow/tensor2tensor/blob/master/tensor2tensor/utils/sari_hook.py""", """https://github.com/mjpost/sacreBLEU""", ] , reference_urls=[ """https://www.aclweb.org/anthology/Q16-1029.pdf""", """https://github.com/mjpost/sacreBLEU""", """https://en.wikipedia.org/wiki/BLEU""", """https://towardsdatascience.com/evaluating-text-output-in-nlp-bleu-at-your-own-risk-e8609665a213""", ] , ) def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase ): '''simple docstring''' __a : Any = {} result.update({"""sari""": compute_sari(sources=_lowercase , predictions=_lowercase , references=_lowercase )} ) result.update({"""sacrebleu""": compute_sacrebleu(predictions=_lowercase , references=_lowercase )} ) result.update({"""exact""": compute_em(predictions=_lowercase , references=_lowercase )} ) return result
714
"""simple docstring""" import argparse from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_controlnet_from_original_ckpt if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() parser.add_argument( "--checkpoint_path", default=None, type=str, required=True, help="Path to the checkpoint to convert." ) parser.add_argument( "--original_config_file", type=str, required=True, help="The YAML config file corresponding to the original architecture.", ) parser.add_argument( "--num_in_channels", default=None, type=int, help="The number of input channels. If `None` number of input channels will be automatically inferred.", ) parser.add_argument( "--image_size", default=512, type=int, help=( "The image size that the model was trained on. Use 512 for Stable Diffusion v1.X and Stable Siffusion v2" " Base. Use 768 for Stable Diffusion v2." ), ) parser.add_argument( "--extract_ema", action="store_true", help=( "Only relevant for checkpoints that have both EMA and non-EMA weights. Whether to extract the EMA weights" " or not. Defaults to `False`. Add `--extract_ema` to extract the EMA weights. EMA weights usually yield" " higher quality images for inference. Non-EMA weights are usually better to continue fine-tuning." ), ) parser.add_argument( "--upcast_attention", action="store_true", help=( "Whether the attention computation should always be upcasted. This is necessary when running stable" " diffusion 2.1." ), ) parser.add_argument( "--from_safetensors", action="store_true", help="If `--checkpoint_path` is in `safetensors` format, load checkpoint with safetensors instead of PyTorch.", ) parser.add_argument( "--to_safetensors", action="store_true", help="Whether to store pipeline in safetensors format or not.", ) parser.add_argument("--dump_path", default=None, type=str, required=True, help="Path to the output model.") parser.add_argument("--device", type=str, help="Device to use (e.g. cpu, cuda:0, cuda:1, etc.)") def __magic_name__ ( _lowerCamelCase : Optional[Any] ): if string == "True": return True elif string == "False": return False else: raise ValueError(F'''could not parse string as bool {string}''' ) parser.add_argument( "--use_linear_projection", help="Override for use linear projection", required=False, type=parse_bool ) parser.add_argument("--cross_attention_dim", help="Override for cross attention_dim", required=False, type=int) lowercase__ = parser.parse_args() lowercase__ = download_controlnet_from_original_ckpt( checkpoint_path=args.checkpoint_path, original_config_file=args.original_config_file, image_size=args.image_size, extract_ema=args.extract_ema, num_in_channels=args.num_in_channels, upcast_attention=args.upcast_attention, from_safetensors=args.from_safetensors, device=args.device, use_linear_projection=args.use_linear_projection, cross_attention_dim=args.cross_attention_dim, ) controlnet.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
63
0
"""simple docstring""" import collections import gzip import os import urllib import numpy from tensorflow.python.framework import dtypes, random_seed from tensorflow.python.platform import gfile from tensorflow.python.util.deprecation import deprecated lowercase__ = collections.namedtuple("_Datasets", ["train", "validation", "test"]) # CVDF mirror of http://yann.lecun.com/exdb/mnist/ lowercase__ = "https://storage.googleapis.com/cvdf-datasets/mnist/" def __magic_name__ ( _lowerCamelCase : List[str] ): __a : int = numpy.dtype(numpy.uintaa ).newbyteorder(""">""" ) return numpy.frombuffer(bytestream.read(4 ) , dtype=_lowerCamelCase )[0] @deprecated(_lowerCamelCase , """Please use tf.data to implement this functionality.""" ) def __magic_name__ ( _lowerCamelCase : Optional[int] ): print("""Extracting""" , f.name ) with gzip.GzipFile(fileobj=_lowerCamelCase ) as bytestream: __a : int = _readaa(_lowerCamelCase ) if magic != 2_0_5_1: raise ValueError( """Invalid magic number %d in MNIST image file: %s""" % (magic, f.name) ) __a : List[Any] = _readaa(_lowerCamelCase ) __a : List[Any] = _readaa(_lowerCamelCase ) __a : int = _readaa(_lowerCamelCase ) __a : int = bytestream.read(rows * cols * num_images ) __a : Dict = numpy.frombuffer(_lowerCamelCase , dtype=numpy.uinta ) __a : Any = data.reshape(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , 1 ) return data @deprecated(_lowerCamelCase , """Please use tf.one_hot on tensors.""" ) def __magic_name__ ( _lowerCamelCase : Optional[int] , _lowerCamelCase : List[Any] ): __a : List[str] = labels_dense.shape[0] __a : int = numpy.arange(_lowerCamelCase ) * num_classes __a : Union[str, Any] = numpy.zeros((num_labels, num_classes) ) __a : int = 1 return labels_one_hot @deprecated(_lowerCamelCase , """Please use tf.data to implement this functionality.""" ) def __magic_name__ ( _lowerCamelCase : Dict , _lowerCamelCase : Union[str, Any]=False , _lowerCamelCase : Dict=1_0 ): print("""Extracting""" , f.name ) with gzip.GzipFile(fileobj=_lowerCamelCase ) as bytestream: __a : Optional[Any] = _readaa(_lowerCamelCase ) if magic != 2_0_4_9: raise ValueError( """Invalid magic number %d in MNIST label file: %s""" % (magic, f.name) ) __a : Union[str, Any] = _readaa(_lowerCamelCase ) __a : List[str] = bytestream.read(_lowerCamelCase ) __a : List[Any] = numpy.frombuffer(_lowerCamelCase , dtype=numpy.uinta ) if one_hot: return _dense_to_one_hot(_lowerCamelCase , _lowerCamelCase ) return labels class SCREAMING_SNAKE_CASE__ : @deprecated( _lowercase , """Please use alternatives such as official/mnist/_DataSet.py""" """ from tensorflow/models.""" , ) def __init__(self , _lowercase , _lowercase , _lowercase=False , _lowercase=False , _lowercase=dtypes.floataa , _lowercase=True , _lowercase=None , ): '''simple docstring''' __a : List[str] = random_seed.get_seed(_lowercase ) # If op level seed is not set, use whatever graph level seed is returned numpy.random.seed(seeda if seed is None else seeda ) __a : Any = dtypes.as_dtype(_lowercase ).base_dtype if dtype not in (dtypes.uinta, dtypes.floataa): raise TypeError("""Invalid image dtype %r, expected uint8 or float32""" % dtype ) if fake_data: __a : Union[str, Any] = 10000 __a : Optional[Any] = one_hot else: assert ( images.shape[0] == labels.shape[0] ), F'''images.shape: {images.shape} labels.shape: {labels.shape}''' __a : int = images.shape[0] # Convert shape from [num examples, rows, columns, depth] # to [num examples, rows*columns] (assuming depth == 1) if reshape: assert images.shape[3] == 1 __a : List[Any] = images.reshape( images.shape[0] , images.shape[1] * images.shape[2] ) if dtype == dtypes.floataa: # Convert from [0, 255] -> [0.0, 1.0]. __a : Dict = images.astype(numpy.floataa ) __a : List[Any] = numpy.multiply(_lowercase , 1.0 / 255.0 ) __a : Dict = images __a : Dict = labels __a : int = 0 __a : Optional[int] = 0 @property def lowerCAmelCase__(self ): '''simple docstring''' return self._images @property def lowerCAmelCase__(self ): '''simple docstring''' return self._labels @property def lowerCAmelCase__(self ): '''simple docstring''' return self._num_examples @property def lowerCAmelCase__(self ): '''simple docstring''' return self._epochs_completed def lowerCAmelCase__(self , _lowercase , _lowercase=False , _lowercase=True ): '''simple docstring''' if fake_data: __a : int = [1] * 784 __a : Dict = [1] + [0] * 9 if self.one_hot else 0 return ( [fake_image for _ in range(_lowercase )], [fake_label for _ in range(_lowercase )], ) __a : Optional[Any] = self._index_in_epoch # Shuffle for the first epoch if self._epochs_completed == 0 and start == 0 and shuffle: __a : Tuple = numpy.arange(self._num_examples ) numpy.random.shuffle(_lowercase ) __a : Dict = self.images[perma] __a : Optional[int] = self.labels[perma] # Go to the next epoch if start + batch_size > self._num_examples: # Finished epoch self._epochs_completed += 1 # Get the rest examples in this epoch __a : Union[str, Any] = self._num_examples - start __a : Union[str, Any] = self._images[start : self._num_examples] __a : int = self._labels[start : self._num_examples] # Shuffle the data if shuffle: __a : Union[str, Any] = numpy.arange(self._num_examples ) numpy.random.shuffle(_lowercase ) __a : Union[str, Any] = self.images[perm] __a : int = self.labels[perm] # Start next epoch __a : List[str] = 0 __a : Dict = batch_size - rest_num_examples __a : Tuple = self._index_in_epoch __a : Any = self._images[start:end] __a : List[Any] = self._labels[start:end] return ( numpy.concatenate((images_rest_part, images_new_part) , axis=0 ), numpy.concatenate((labels_rest_part, labels_new_part) , axis=0 ), ) else: self._index_in_epoch += batch_size __a : Optional[Any] = self._index_in_epoch return self._images[start:end], self._labels[start:end] @deprecated(_lowerCamelCase , """Please write your own downloading logic.""" ) def __magic_name__ ( _lowerCamelCase : Dict , _lowerCamelCase : Union[str, Any] , _lowerCamelCase : Optional[Any] ): if not gfile.Exists(_lowerCamelCase ): gfile.MakeDirs(_lowerCamelCase ) __a : List[str] = os.path.join(_lowerCamelCase , _lowerCamelCase ) if not gfile.Exists(_lowerCamelCase ): urllib.request.urlretrieve(_lowerCamelCase , _lowerCamelCase ) # noqa: S310 with gfile.GFile(_lowerCamelCase ) as f: __a : List[str] = f.size() print("""Successfully downloaded""" , _lowerCamelCase , _lowerCamelCase , """bytes.""" ) return filepath @deprecated( _lowerCamelCase , """Please use alternatives such as:""" """ tensorflow_datasets.load('mnist')""" ) def __magic_name__ ( _lowerCamelCase : List[str] , _lowerCamelCase : Optional[int]=False , _lowerCamelCase : Union[str, Any]=False , _lowerCamelCase : Union[str, Any]=dtypes.floataa , _lowerCamelCase : Optional[Any]=True , _lowerCamelCase : List[Any]=5_0_0_0 , _lowerCamelCase : Tuple=None , _lowerCamelCase : Tuple=DEFAULT_SOURCE_URL , ): if fake_data: def fake(): return _DataSet( [] , [] , fake_data=_lowerCamelCase , one_hot=_lowerCamelCase , dtype=_lowerCamelCase , seed=_lowerCamelCase ) __a : Any = fake() __a : Optional[int] = fake() __a : str = fake() return _Datasets(train=_lowerCamelCase , validation=_lowerCamelCase , test=_lowerCamelCase ) if not source_url: # empty string check __a : str = DEFAULT_SOURCE_URL __a : int = """train-images-idx3-ubyte.gz""" __a : Union[str, Any] = """train-labels-idx1-ubyte.gz""" __a : str = """t10k-images-idx3-ubyte.gz""" __a : Optional[int] = """t10k-labels-idx1-ubyte.gz""" __a : Optional[Any] = _maybe_download( _lowerCamelCase , _lowerCamelCase , source_url + train_images_file ) with gfile.Open(_lowerCamelCase , """rb""" ) as f: __a : List[Any] = _extract_images(_lowerCamelCase ) __a : str = _maybe_download( _lowerCamelCase , _lowerCamelCase , source_url + train_labels_file ) with gfile.Open(_lowerCamelCase , """rb""" ) as f: __a : Optional[int] = _extract_labels(_lowerCamelCase , one_hot=_lowerCamelCase ) __a : Optional[int] = _maybe_download( _lowerCamelCase , _lowerCamelCase , source_url + test_images_file ) with gfile.Open(_lowerCamelCase , """rb""" ) as f: __a : Optional[Any] = _extract_images(_lowerCamelCase ) __a : Any = _maybe_download( _lowerCamelCase , _lowerCamelCase , source_url + test_labels_file ) with gfile.Open(_lowerCamelCase , """rb""" ) as f: __a : str = _extract_labels(_lowerCamelCase , one_hot=_lowerCamelCase ) if not 0 <= validation_size <= len(_lowerCamelCase ): __a : Dict = ( """Validation size should be between 0 and """ F'''{len(_lowerCamelCase )}. Received: {validation_size}.''' ) raise ValueError(_lowerCamelCase ) __a : int = train_images[:validation_size] __a : Optional[Any] = train_labels[:validation_size] __a : Optional[Any] = train_images[validation_size:] __a : Tuple = train_labels[validation_size:] __a : int = {"""dtype""": dtype, """reshape""": reshape, """seed""": seed} __a : List[str] = _DataSet(_lowerCamelCase , _lowerCamelCase , **_lowerCamelCase ) __a : List[Any] = _DataSet(_lowerCamelCase , _lowerCamelCase , **_lowerCamelCase ) __a : Any = _DataSet(_lowerCamelCase , _lowerCamelCase , **_lowerCamelCase ) return _Datasets(train=_lowerCamelCase , validation=_lowerCamelCase , test=_lowerCamelCase )
715
"""simple docstring""" import torch from diffusers import DiffusionPipeline class SCREAMING_SNAKE_CASE__ ( __snake_case ): def __init__(self , _lowercase , _lowercase ): '''simple docstring''' super().__init__() self.register_modules(unet=_lowercase , scheduler=_lowercase ) def __call__(self ): '''simple docstring''' __a : Dict = torch.randn( (1, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , ) __a : Optional[Any] = 1 __a : List[str] = self.unet(_lowercase , _lowercase ).sample __a : Union[str, Any] = self.scheduler.step(_lowercase , _lowercase , _lowercase ).prev_sample __a : Optional[int] = scheduler_output - scheduler_output + torch.ones_like(_lowercase ) return result
63
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowercase__ = {"configuration_xglm": ["XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP", "XGLMConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = ["XGLMTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = ["XGLMTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ "XGLM_PRETRAINED_MODEL_ARCHIVE_LIST", "XGLMForCausalLM", "XGLMModel", "XGLMPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ "FlaxXGLMForCausalLM", "FlaxXGLMModel", "FlaxXGLMPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ "TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXGLMForCausalLM", "TFXGLMModel", "TFXGLMPreTrainedModel", ] if TYPE_CHECKING: from .configuration_xglm import XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XGLMConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm import XGLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm_fast import XGLMTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xglm import XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, XGLMForCausalLM, XGLMModel, XGLMPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xglm import FlaxXGLMForCausalLM, FlaxXGLMModel, FlaxXGLMPreTrainedModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, TFXGLMPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()["__file__"], _import_structure)
716
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { "sayakpaul/vit-msn-base": "https://huggingface.co/sayakpaul/vit-msn-base/resolve/main/config.json", # See all ViT MSN models at https://huggingface.co/models?filter=vit_msn } class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = "vit_msn" def __init__(self , _lowercase=768 , _lowercase=12 , _lowercase=12 , _lowercase=3072 , _lowercase="gelu" , _lowercase=0.0 , _lowercase=0.0 , _lowercase=0.02 , _lowercase=1e-06 , _lowercase=224 , _lowercase=16 , _lowercase=3 , _lowercase=True , **_lowercase , ): '''simple docstring''' super().__init__(**_lowercase ) __a : int = hidden_size __a : str = num_hidden_layers __a : str = num_attention_heads __a : Optional[Any] = intermediate_size __a : Union[str, Any] = hidden_act __a : Tuple = hidden_dropout_prob __a : Any = attention_probs_dropout_prob __a : List[Any] = initializer_range __a : Any = layer_norm_eps __a : Dict = image_size __a : List[Any] = patch_size __a : Dict = num_channels __a : Optional[Any] = qkv_bias
63
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available lowercase__ = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = ["GPTSw3Tokenizer"] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_swa import GPTSwaTokenizer else: import sys lowercase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
717
"""simple docstring""" import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert_fast import BertTokenizerFast from .tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer, DPRReaderTokenizer lowercase__ = logging.get_logger(__name__) lowercase__ = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} lowercase__ = { "vocab_file": { "facebook/dpr-ctx_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-ctx_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-ctx_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-ctx_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/tokenizer.json" ), }, } lowercase__ = { "vocab_file": { "facebook/dpr-question_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-question_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-question_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-question_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/tokenizer.json" ), }, } lowercase__ = { "vocab_file": { "facebook/dpr-reader-single-nq-base": ( "https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-reader-multiset-base": ( "https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-reader-single-nq-base": ( "https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-reader-multiset-base": ( "https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/tokenizer.json" ), }, } lowercase__ = { "facebook/dpr-ctx_encoder-single-nq-base": 512, "facebook/dpr-ctx_encoder-multiset-base": 512, } lowercase__ = { "facebook/dpr-question_encoder-single-nq-base": 512, "facebook/dpr-question_encoder-multiset-base": 512, } lowercase__ = { "facebook/dpr-reader-single-nq-base": 512, "facebook/dpr-reader-multiset-base": 512, } lowercase__ = { "facebook/dpr-ctx_encoder-single-nq-base": {"do_lower_case": True}, "facebook/dpr-ctx_encoder-multiset-base": {"do_lower_case": True}, } lowercase__ = { "facebook/dpr-question_encoder-single-nq-base": {"do_lower_case": True}, "facebook/dpr-question_encoder-multiset-base": {"do_lower_case": True}, } lowercase__ = { "facebook/dpr-reader-single-nq-base": {"do_lower_case": True}, "facebook/dpr-reader-multiset-base": {"do_lower_case": True}, } class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = VOCAB_FILES_NAMES _lowerCAmelCase = CONTEXT_ENCODER_PRETRAINED_VOCAB_FILES_MAP _lowerCAmelCase = CONTEXT_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCAmelCase = CONTEXT_ENCODER_PRETRAINED_INIT_CONFIGURATION _lowerCAmelCase = DPRContextEncoderTokenizer class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = VOCAB_FILES_NAMES _lowerCAmelCase = QUESTION_ENCODER_PRETRAINED_VOCAB_FILES_MAP _lowerCAmelCase = QUESTION_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCAmelCase = QUESTION_ENCODER_PRETRAINED_INIT_CONFIGURATION _lowerCAmelCase = DPRQuestionEncoderTokenizer lowercase__ = collections.namedtuple( "DPRSpanPrediction", ["span_score", "relevance_score", "doc_id", "start_index", "end_index", "text"] ) lowercase__ = collections.namedtuple("DPRReaderOutput", ["start_logits", "end_logits", "relevance_logits"]) lowercase__ = R"\n Return a dictionary with the token ids of the input strings and other information to give to `.decode_best_spans`.\n It converts the strings of a question and different passages (title and text) in a sequence of IDs (integers),\n using the tokenizer and vocabulary. The resulting `input_ids` is a matrix of size `(n_passages, sequence_length)`\n with the format:\n\n [CLS] <question token ids> [SEP] <titles ids> [SEP] <texts ids>\n\n Args:\n questions (`str` or `List[str]`):\n The questions to be encoded. You can specify one question for many passages. In this case, the question\n will be duplicated like `[questions] * n_passages`. Otherwise you have to specify as many questions as in\n `titles` or `texts`.\n titles (`str` or `List[str]`):\n The passages titles to be encoded. This can be a string or a list of strings if there are several passages.\n texts (`str` or `List[str]`):\n The passages texts to be encoded. This can be a string or a list of strings if there are several passages.\n padding (`bool`, `str` or [`~utils.PaddingStrategy`], *optional*, defaults to `False`):\n Activates and controls padding. Accepts the following values:\n\n - `True` or `'longest'`: Pad to the longest sequence in the batch (or no padding if only a single sequence\n if provided).\n - `'max_length'`: Pad to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided.\n - `False` or `'do_not_pad'` (default): No padding (i.e., can output a batch with sequences of different\n lengths).\n truncation (`bool`, `str` or [`~tokenization_utils_base.TruncationStrategy`], *optional*, defaults to `False`):\n Activates and controls truncation. Accepts the following values:\n\n - `True` or `'longest_first'`: Truncate to a maximum length specified with the argument `max_length` or to\n the maximum acceptable input length for the model if that argument is not provided. This will truncate\n token by token, removing a token from the longest sequence in the pair if a pair of sequences (or a batch\n of pairs) is provided.\n - `'only_first'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the first\n sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `'only_second'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the\n second sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `False` or `'do_not_truncate'` (default): No truncation (i.e., can output batch with sequence lengths\n greater than the model maximum admissible input size).\n max_length (`int`, *optional*):\n Controls the maximum length to use by one of the truncation/padding parameters.\n\n If left unset or set to `None`, this will use the predefined model maximum length if a maximum length\n is required by one of the truncation/padding parameters. If the model has no specific maximum input\n length (like XLNet) truncation/padding to a maximum length will be deactivated.\n return_tensors (`str` or [`~utils.TensorType`], *optional*):\n If set, will return tensors instead of list of python integers. Acceptable values are:\n\n - `'tf'`: Return TensorFlow `tf.constant` objects.\n - `'pt'`: Return PyTorch `torch.Tensor` objects.\n - `'np'`: Return Numpy `np.ndarray` objects.\n return_attention_mask (`bool`, *optional*):\n Whether or not to return the attention mask. If not set, will return the attention mask according to the\n specific tokenizer's default, defined by the `return_outputs` attribute.\n\n [What are attention masks?](../glossary#attention-mask)\n\n Return:\n `Dict[str, List[List[int]]]`: A dictionary with the following keys:\n\n - `input_ids`: List of token ids to be fed to a model.\n - `attention_mask`: List of indices specifying which tokens should be attended to by the model.\n " @add_start_docstrings(__snake_case ) class SCREAMING_SNAKE_CASE__ : def __call__(self , _lowercase , _lowercase = None , _lowercase = None , _lowercase = False , _lowercase = False , _lowercase = None , _lowercase = None , _lowercase = None , **_lowercase , ): '''simple docstring''' if titles is None and texts is None: return super().__call__( _lowercase , padding=_lowercase , truncation=_lowercase , max_length=_lowercase , return_tensors=_lowercase , return_attention_mask=_lowercase , **_lowercase , ) elif titles is None or texts is None: __a : str = titles if texts is None else texts return super().__call__( _lowercase , _lowercase , padding=_lowercase , truncation=_lowercase , max_length=_lowercase , return_tensors=_lowercase , return_attention_mask=_lowercase , **_lowercase , ) __a : str = titles if not isinstance(_lowercase , _lowercase ) else [titles] __a : Optional[Any] = texts if not isinstance(_lowercase , _lowercase ) else [texts] __a : Tuple = len(_lowercase ) __a : Dict = questions if not isinstance(_lowercase , _lowercase ) else [questions] * n_passages assert len(_lowercase ) == len( _lowercase ), F'''There should be as many titles than texts but got {len(_lowercase )} titles and {len(_lowercase )} texts.''' __a : Optional[Any] = super().__call__(_lowercase , _lowercase , padding=_lowercase , truncation=_lowercase )["""input_ids"""] __a : str = super().__call__(_lowercase , add_special_tokens=_lowercase , padding=_lowercase , truncation=_lowercase )["""input_ids"""] __a : Union[str, Any] = { """input_ids""": [ (encoded_question_and_title + encoded_text)[:max_length] if max_length is not None and truncation else encoded_question_and_title + encoded_text for encoded_question_and_title, encoded_text in zip(_lowercase , _lowercase ) ] } if return_attention_mask is not False: __a : Optional[int] = [] for input_ids in encoded_inputs["input_ids"]: attention_mask.append([int(input_id != self.pad_token_id ) for input_id in input_ids] ) __a : str = attention_mask return self.pad(_lowercase , padding=_lowercase , max_length=_lowercase , return_tensors=_lowercase ) def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase = 16 , _lowercase = 64 , _lowercase = 4 , ): '''simple docstring''' __a : Union[str, Any] = reader_input["""input_ids"""] __a , __a , __a : Optional[int] = reader_output[:3] __a : int = len(_lowercase ) __a : Any = sorted(range(_lowercase ) , reverse=_lowercase , key=relevance_logits.__getitem__ ) __a : List[DPRReaderOutput] = [] for doc_id in sorted_docs: __a : Optional[int] = list(input_ids[doc_id] ) # assuming question & title information is at the beginning of the sequence __a : Dict = sequence_ids.index(self.sep_token_id , 2 ) + 1 # second sep id if sequence_ids[-1] == self.pad_token_id: __a : int = sequence_ids.index(self.pad_token_id ) else: __a : Optional[Any] = len(_lowercase ) __a : List[Any] = self._get_best_spans( start_logits=start_logits[doc_id][passage_offset:sequence_len] , end_logits=end_logits[doc_id][passage_offset:sequence_len] , max_answer_length=_lowercase , top_spans=_lowercase , ) for start_index, end_index in best_spans: start_index += passage_offset end_index += passage_offset nbest_spans_predictions.append( DPRSpanPrediction( span_score=start_logits[doc_id][start_index] + end_logits[doc_id][end_index] , relevance_score=relevance_logits[doc_id] , doc_id=_lowercase , start_index=_lowercase , end_index=_lowercase , text=self.decode(sequence_ids[start_index : end_index + 1] ) , ) ) if len(_lowercase ) >= num_spans: break return nbest_spans_predictions[:num_spans] def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase , _lowercase , ): '''simple docstring''' __a : Tuple = [] for start_index, start_score in enumerate(_lowercase ): for answer_length, end_score in enumerate(end_logits[start_index : start_index + max_answer_length] ): scores.append(((start_index, start_index + answer_length), start_score + end_score) ) __a : str = sorted(_lowercase , key=lambda _lowercase : x[1] , reverse=_lowercase ) __a : Union[str, Any] = [] for (start_index, end_index), score in scores: assert start_index <= end_index, F'''Wrong span indices: [{start_index}:{end_index}]''' __a : List[str] = end_index - start_index + 1 assert length <= max_answer_length, F'''Span is too long: {length} > {max_answer_length}''' if any( start_index <= prev_start_index <= prev_end_index <= end_index or prev_start_index <= start_index <= end_index <= prev_end_index for (prev_start_index, prev_end_index) in chosen_span_intervals ): continue chosen_span_intervals.append((start_index, end_index) ) if len(_lowercase ) == top_spans: break return chosen_span_intervals @add_end_docstrings(__snake_case ) class SCREAMING_SNAKE_CASE__ ( __snake_case , __snake_case ): _lowerCAmelCase = VOCAB_FILES_NAMES _lowerCAmelCase = READER_PRETRAINED_VOCAB_FILES_MAP _lowerCAmelCase = READER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCAmelCase = READER_PRETRAINED_INIT_CONFIGURATION _lowerCAmelCase = ["input_ids", "attention_mask"] _lowerCAmelCase = DPRReaderTokenizer
63
0
"""simple docstring""" class SCREAMING_SNAKE_CASE__ : def __init__(self ): '''simple docstring''' __a : dict[str, TrieNode] = {} # Mapping from char to TrieNode __a : List[Any] = False def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' for word in words: self.insert(_lowercase ) def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' __a : Any = self for char in word: if char not in curr.nodes: __a : Optional[int] = TrieNode() __a : List[str] = curr.nodes[char] __a : List[Any] = True def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' __a : Tuple = self for char in word: if char not in curr.nodes: return False __a : Optional[int] = curr.nodes[char] return curr.is_leaf def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' def _delete(_lowercase , _lowercase , _lowercase ) -> bool: if index == len(_lowercase ): # If word does not exist if not curr.is_leaf: return False __a : Optional[Any] = False return len(curr.nodes ) == 0 __a : Optional[int] = word[index] __a : Any = curr.nodes.get(_lowercase ) # If char not in current trie node if not char_node: return False # Flag to check if node can be deleted __a : Union[str, Any] = _delete(_lowercase , _lowercase , index + 1 ) if delete_curr: del curr.nodes[char] return len(curr.nodes ) == 0 return delete_curr _delete(self , _lowercase , 0 ) def __magic_name__ ( _lowerCamelCase : TrieNode , _lowerCamelCase : str ): if node.is_leaf: print(_lowerCamelCase , end=""" """ ) for key, value in node.nodes.items(): print_words(_lowerCamelCase , word + key ) def __magic_name__ ( ): __a : Union[str, Any] = """banana bananas bandana band apple all beast""".split() __a : List[Any] = TrieNode() root.insert_many(_lowerCamelCase ) # print_words(root, "") assert all(root.find(_lowerCamelCase ) for word in words ) assert root.find("""banana""" ) assert not root.find("""bandanas""" ) assert not root.find("""apps""" ) assert root.find("""apple""" ) assert root.find("""all""" ) root.delete("""all""" ) assert not root.find("""all""" ) root.delete("""banana""" ) assert not root.find("""banana""" ) assert root.find("""bananas""" ) return True def __magic_name__ ( _lowerCamelCase : str , _lowerCamelCase : bool ): print(str(_lowerCamelCase ) , """works!""" if passes else """doesn't work :(""" ) def __magic_name__ ( ): assert test_trie() def __magic_name__ ( ): print_results("""Testing trie functionality""" , test_trie() ) if __name__ == "__main__": main()
718
"""simple docstring""" import os def __magic_name__ ( _lowerCamelCase : Dict ): __a : List[str] = len(grid[0] ) __a : int = len(_lowerCamelCase ) __a : Tuple = 0 __a : List[Any] = 0 __a : List[str] = 0 # Check vertically, horizontally, diagonally at the same time (only works # for nxn grid) for i in range(_lowerCamelCase ): for j in range(n_rows - 3 ): __a : List[Any] = grid[j][i] * grid[j + 1][i] * grid[j + 2][i] * grid[j + 3][i] __a : Tuple = grid[i][j] * grid[i][j + 1] * grid[i][j + 2] * grid[i][j + 3] # Left-to-right diagonal (\) product if i < n_columns - 3: __a : List[Any] = ( grid[i][j] * grid[i + 1][j + 1] * grid[i + 2][j + 2] * grid[i + 3][j + 3] ) # Right-to-left diagonal(/) product if i > 2: __a : List[Any] = ( grid[i][j] * grid[i - 1][j + 1] * grid[i - 2][j + 2] * grid[i - 3][j + 3] ) __a : str = max( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) if max_product > largest: __a : Optional[Any] = max_product return largest def __magic_name__ ( ): __a : Tuple = [] with open(os.path.dirname(_lowerCamelCase ) + """/grid.txt""" ) as file: for line in file: grid.append(line.strip("""\n""" ).split(""" """ ) ) __a : Tuple = [[int(_lowerCamelCase ) for i in grid[j]] for j in range(len(_lowerCamelCase ) )] return largest_product(_lowerCamelCase ) if __name__ == "__main__": print(solution())
63
0
"""simple docstring""" import unittest from huggingface_hub import hf_hub_download from transformers import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEFeatureExtractor from transformers.pipelines import VideoClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_decord, require_tf, require_torch, require_torch_or_tf, require_vision, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf @require_vision @require_decord class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): _lowerCAmelCase = MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase ): '''simple docstring''' __a : str = hf_hub_download( repo_id="""nateraw/video-demo""" , filename="""archery.mp4""" , repo_type="""dataset""" ) __a : str = VideoClassificationPipeline(model=_lowercase , image_processor=_lowercase , top_k=2 ) __a : Any = [ example_video_filepath, """https://huggingface.co/datasets/nateraw/video-demo/resolve/main/archery.mp4""", ] return video_classifier, examples def lowerCAmelCase__(self , _lowercase , _lowercase ): '''simple docstring''' for example in examples: __a : Dict = video_classifier(_lowercase ) self.assertEqual( _lowercase , [ {"""score""": ANY(_lowercase ), """label""": ANY(_lowercase )}, {"""score""": ANY(_lowercase ), """label""": ANY(_lowercase )}, ] , ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[Any] = """hf-internal-testing/tiny-random-VideoMAEForVideoClassification""" __a : Tuple = VideoMAEFeatureExtractor( size={"""shortest_edge""": 10} , crop_size={"""height""": 10, """width""": 10} ) __a : Any = pipeline( """video-classification""" , model=_lowercase , feature_extractor=_lowercase , frame_sampling_rate=4 ) __a : Dict = hf_hub_download(repo_id="""nateraw/video-demo""" , filename="""archery.mp4""" , repo_type="""dataset""" ) __a : str = video_classifier(_lowercase , top_k=2 ) self.assertEqual( nested_simplify(_lowercase , decimals=4 ) , [{"""score""": 0.5199, """label""": """LABEL_0"""}, {"""score""": 0.4801, """label""": """LABEL_1"""}] , ) __a : List[Any] = video_classifier( [ video_file_path, video_file_path, ] , top_k=2 , ) self.assertEqual( nested_simplify(_lowercase , decimals=4 ) , [ [{"""score""": 0.5199, """label""": """LABEL_0"""}, {"""score""": 0.4801, """label""": """LABEL_1"""}], [{"""score""": 0.5199, """label""": """LABEL_0"""}, {"""score""": 0.4801, """label""": """LABEL_1"""}], ] , ) @require_tf def lowerCAmelCase__(self ): '''simple docstring''' pass
719
"""simple docstring""" from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import BaseOutput, is_torch_available, is_transformers_available @dataclass class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = 42 _lowerCAmelCase = 42 if is_transformers_available() and is_torch_available(): from .pipeline_semantic_stable_diffusion import SemanticStableDiffusionPipeline
63
0
"""simple docstring""" import qiskit def __magic_name__ ( _lowerCamelCase : int , _lowerCamelCase : int ): __a : Optional[int] = qiskit.Aer.get_backend("""aer_simulator""" ) # Create a Quantum Circuit acting on the q register __a : Union[str, Any] = qiskit.QuantumCircuit(_lowerCamelCase , _lowerCamelCase ) # Apply X (NOT) Gate to Qubits 0 & 1 circuit.x(0 ) circuit.x(1 ) # Map the quantum measurement to the classical bits circuit.measure([0, 1] , [0, 1] ) # Execute the circuit on the qasm simulator __a : Dict = qiskit.execute(_lowerCamelCase , _lowerCamelCase , shots=1_0_0_0 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(_lowerCamelCase ) if __name__ == "__main__": lowercase__ = single_qubit_measure(2, 2) print(f'Total count for various states are: {counts}')
720
"""simple docstring""" import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TextClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. lowercase__ = {"LayoutLMv2Config", "LayoutLMv3Config"} @is_pipeline_test class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): _lowerCAmelCase = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING _lowerCAmelCase = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: _lowerCAmelCase = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: _lowerCAmelCase = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } @require_torch def lowerCAmelCase__(self ): '''simple docstring''' __a : int = pipeline( task="""text-classification""" , model="""hf-internal-testing/tiny-random-distilbert""" , framework="""pt""" ) __a : Tuple = text_classifier("""This is great !""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """LABEL_0""", """score""": 0.504}] ) __a : Optional[Any] = text_classifier("""This is great !""" , top_k=2 ) self.assertEqual( nested_simplify(_lowercase ) , [{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}] ) __a : int = text_classifier(["""This is great !""", """This is bad"""] , top_k=2 ) self.assertEqual( nested_simplify(_lowercase ) , [ [{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}], [{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}], ] , ) __a : List[str] = text_classifier("""This is great !""" , top_k=1 ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """LABEL_0""", """score""": 0.504}] ) # Legacy behavior __a : Optional[int] = text_classifier("""This is great !""" , return_all_scores=_lowercase ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """LABEL_0""", """score""": 0.504}] ) __a : Tuple = text_classifier("""This is great !""" , return_all_scores=_lowercase ) self.assertEqual( nested_simplify(_lowercase ) , [[{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}]] ) __a : Any = text_classifier(["""This is great !""", """Something else"""] , return_all_scores=_lowercase ) self.assertEqual( nested_simplify(_lowercase ) , [ [{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}], [{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}], ] , ) __a : Union[str, Any] = text_classifier(["""This is great !""", """Something else"""] , return_all_scores=_lowercase ) self.assertEqual( nested_simplify(_lowercase ) , [ {"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_0""", """score""": 0.504}, ] , ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' import torch __a : Any = pipeline( task="""text-classification""" , model="""hf-internal-testing/tiny-random-distilbert""" , framework="""pt""" , device=torch.device("""cpu""" ) , ) __a : Optional[int] = text_classifier("""This is great !""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """LABEL_0""", """score""": 0.504}] ) @require_tf def lowerCAmelCase__(self ): '''simple docstring''' __a : List[Any] = pipeline( task="""text-classification""" , model="""hf-internal-testing/tiny-random-distilbert""" , framework="""tf""" ) __a : List[str] = text_classifier("""This is great !""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """LABEL_0""", """score""": 0.504}] ) @slow @require_torch def lowerCAmelCase__(self ): '''simple docstring''' __a : Tuple = pipeline("""text-classification""" ) __a : Tuple = text_classifier("""This is great !""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """POSITIVE""", """score""": 1.0}] ) __a : Optional[int] = text_classifier("""This is bad !""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """NEGATIVE""", """score""": 1.0}] ) __a : Union[str, Any] = text_classifier("""Birds are a type of animal""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """POSITIVE""", """score""": 0.988}] ) @slow @require_tf def lowerCAmelCase__(self ): '''simple docstring''' __a : List[str] = pipeline("""text-classification""" , framework="""tf""" ) __a : str = text_classifier("""This is great !""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """POSITIVE""", """score""": 1.0}] ) __a : Tuple = text_classifier("""This is bad !""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """NEGATIVE""", """score""": 1.0}] ) __a : str = text_classifier("""Birds are a type of animal""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """POSITIVE""", """score""": 0.988}] ) def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase ): '''simple docstring''' __a : Dict = TextClassificationPipeline(model=_lowercase , tokenizer=_lowercase ) return text_classifier, ["HuggingFace is in", "This is another test"] def lowerCAmelCase__(self , _lowercase , _lowercase ): '''simple docstring''' __a : List[str] = text_classifier.model # Small inputs because BartTokenizer tiny has maximum position embeddings = 22 __a : Union[str, Any] = """HuggingFace is in""" __a : List[str] = text_classifier(_lowercase ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": ANY(_lowercase ), """score""": ANY(_lowercase )}] ) self.assertTrue(outputs[0]["""label"""] in model.config.idalabel.values() ) __a : Optional[int] = ["""HuggingFace is in """, """Paris is in France"""] __a : Dict = text_classifier(_lowercase ) self.assertEqual( nested_simplify(_lowercase ) , [{"""label""": ANY(_lowercase ), """score""": ANY(_lowercase )}, {"""label""": ANY(_lowercase ), """score""": ANY(_lowercase )}] , ) self.assertTrue(outputs[0]["""label"""] in model.config.idalabel.values() ) self.assertTrue(outputs[1]["""label"""] in model.config.idalabel.values() ) # Forcing to get all results with `top_k=None` # This is NOT the legacy format __a : Dict = text_classifier(_lowercase , top_k=_lowercase ) __a : Dict = len(model.config.idalabel.values() ) self.assertEqual( nested_simplify(_lowercase ) , [[{"""label""": ANY(_lowercase ), """score""": ANY(_lowercase )}] * N, [{"""label""": ANY(_lowercase ), """score""": ANY(_lowercase )}] * N] , ) __a : Dict = {"""text""": """HuggingFace is in """, """text_pair""": """Paris is in France"""} __a : Any = text_classifier(_lowercase ) self.assertEqual( nested_simplify(_lowercase ) , {"""label""": ANY(_lowercase ), """score""": ANY(_lowercase )} , ) self.assertTrue(outputs["""label"""] in model.config.idalabel.values() ) # This might be used a text pair, but tokenizer + pipe interaction # makes it hard to understand that it's not using the pair properly # https://github.com/huggingface/transformers/issues/17305 # We disabled this usage instead as it was outputting wrong outputs. __a : Dict = [["""HuggingFace is in """, """Paris is in France"""]] with self.assertRaises(_lowercase ): text_classifier(_lowercase ) # This used to be valid for doing text pairs # We're keeping it working because of backward compatibility __a : Optional[int] = text_classifier([[["""HuggingFace is in """, """Paris is in France"""]]] ) self.assertEqual( nested_simplify(_lowercase ) , [{"""label""": ANY(_lowercase ), """score""": ANY(_lowercase )}] , ) self.assertTrue(outputs[0]["""label"""] in model.config.idalabel.values() )
63
0
"""simple docstring""" def __magic_name__ ( _lowerCamelCase : Tuple , _lowerCamelCase : Tuple ): print("""\nThe shortest path matrix using Floyd Warshall algorithm\n""" ) for i in range(_lowerCamelCase ): for j in range(_lowerCamelCase ): if dist[i][j] != float("""inf""" ): print(int(dist[i][j] ) , end="""\t""" ) else: print("""INF""" , end="""\t""" ) print() def __magic_name__ ( _lowerCamelCase : List[str] , _lowerCamelCase : Union[str, Any] ): __a : Optional[int] = [[float("""inf""" ) for _ in range(_lowerCamelCase )] for _ in range(_lowerCamelCase )] for i in range(_lowerCamelCase ): for j in range(_lowerCamelCase ): __a : Dict = graph[i][j] # check vertex k against all other vertices (i, j) for k in range(_lowerCamelCase ): # looping through rows of graph array for i in range(_lowerCamelCase ): # looping through columns of graph array for j in range(_lowerCamelCase ): if ( dist[i][k] != float("""inf""" ) and dist[k][j] != float("""inf""" ) and dist[i][k] + dist[k][j] < dist[i][j] ): __a : Any = dist[i][k] + dist[k][j] _print_dist(_lowerCamelCase , _lowerCamelCase ) return dist, v if __name__ == "__main__": lowercase__ = int(input("Enter number of vertices: ")) lowercase__ = int(input("Enter number of edges: ")) lowercase__ = [[float("inf") for i in range(v)] for j in range(v)] for i in range(v): lowercase__ = 0.0 # src and dst are indices that must be within the array size graph[e][v] # failure to follow this will result in an error for i in range(e): print("\nEdge ", i + 1) lowercase__ = int(input("Enter source:")) lowercase__ = int(input("Enter destination:")) lowercase__ = float(input("Enter weight:")) lowercase__ = weight floyd_warshall(graph, v) # Example Input # Enter number of vertices: 3 # Enter number of edges: 2 # # generated graph from vertex and edge inputs # [[inf, inf, inf], [inf, inf, inf], [inf, inf, inf]] # [[0.0, inf, inf], [inf, 0.0, inf], [inf, inf, 0.0]] # specify source, destination and weight for edge #1 # Edge 1 # Enter source:1 # Enter destination:2 # Enter weight:2 # specify source, destination and weight for edge #2 # Edge 2 # Enter source:2 # Enter destination:1 # Enter weight:1 # # Expected Output from the vertice, edge and src, dst, weight inputs!! # 0 INF INF # INF 0 2 # INF 1 0
721
"""simple docstring""" import unittest from knapsack import knapsack as k class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def lowerCAmelCase__(self ): '''simple docstring''' __a : str = 0 __a : Optional[Any] = [0] __a : int = [0] __a : str = len(_lowercase ) self.assertEqual(k.knapsack(_lowercase , _lowercase , _lowercase , _lowercase ) , 0 ) __a : int = [60] __a : Union[str, Any] = [10] __a : Tuple = len(_lowercase ) self.assertEqual(k.knapsack(_lowercase , _lowercase , _lowercase , _lowercase ) , 0 ) def lowerCAmelCase__(self ): '''simple docstring''' __a : int = 3 __a : str = [1, 2, 3] __a : Optional[Any] = [3, 2, 1] __a : int = len(_lowercase ) self.assertEqual(k.knapsack(_lowercase , _lowercase , _lowercase , _lowercase ) , 5 ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Dict = 50 __a : Tuple = [60, 100, 120] __a : List[str] = [10, 20, 30] __a : Union[str, Any] = len(_lowercase ) self.assertEqual(k.knapsack(_lowercase , _lowercase , _lowercase , _lowercase ) , 220 ) if __name__ == "__main__": unittest.main()
63
0
"""simple docstring""" from maths.is_square_free import is_square_free from maths.prime_factors import prime_factors def __magic_name__ ( _lowerCamelCase : int ): __a : Optional[int] = prime_factors(_lowerCamelCase ) if is_square_free(_lowerCamelCase ): return -1 if len(_lowerCamelCase ) % 2 else 1 return 0 if __name__ == "__main__": import doctest doctest.testmod()
700
"""simple docstring""" from manim import * class SCREAMING_SNAKE_CASE__ ( __snake_case ): def lowerCAmelCase__(self ): '''simple docstring''' __a : List[str] = Rectangle(height=0.5 , width=0.5 ) __a : Union[str, Any] = Rectangle(height=0.25 , width=0.25 ) __a : Dict = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) __a : Dict = [mem.copy() for i in range(6 )] __a : str = [mem.copy() for i in range(6 )] __a : Tuple = VGroup(*_lowercase ).arrange(_lowercase , buff=0 ) __a : List[Any] = VGroup(*_lowercase ).arrange(_lowercase , buff=0 ) __a : List[Any] = VGroup(_lowercase , _lowercase ).arrange(_lowercase , buff=0 ) __a : Union[str, Any] = Text("""CPU""" , font_size=24 ) __a : Dict = Group(_lowercase , _lowercase ).arrange(_lowercase , buff=0.5 , aligned_edge=_lowercase ) cpu.move_to([-2.5, -0.5, 0] ) self.add(_lowercase ) __a : Optional[Any] = [mem.copy() for i in range(4 )] __a : Dict = VGroup(*_lowercase ).arrange(_lowercase , buff=0 ) __a : List[str] = Text("""GPU""" , font_size=24 ) __a : Any = Group(_lowercase , _lowercase ).arrange(_lowercase , buff=0.5 , aligned_edge=_lowercase ) gpu.move_to([-1, -1, 0] ) self.add(_lowercase ) __a : List[Any] = [mem.copy() for i in range(6 )] __a : Any = VGroup(*_lowercase ).arrange(_lowercase , buff=0 ) __a : Optional[Any] = Text("""Model""" , font_size=24 ) __a : Any = Group(_lowercase , _lowercase ).arrange(_lowercase , buff=0.5 , aligned_edge=_lowercase ) model.move_to([3, -1.0, 0] ) self.add(_lowercase ) __a : Tuple = [] __a : Tuple = [] __a : Optional[int] = [] for i, rect in enumerate(_lowercase ): rect.set_stroke(_lowercase ) __a : str = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(_lowercase , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=_lowercase ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(model_cpu_arr[0] , direction=_lowercase , buff=0.0 ) else: cpu_target.next_to(model_cpu_arr[i - 1] , direction=_lowercase , buff=0.0 ) self.add(_lowercase ) model_cpu_arr.append(_lowercase ) self.add(*_lowercase , *_lowercase , *_lowercase ) __a : Optional[Any] = [mem.copy() for i in range(6 )] __a : Union[str, Any] = VGroup(*_lowercase ).arrange(_lowercase , buff=0 ) __a : Any = Text("""Loaded Checkpoint""" , font_size=24 ) __a : str = Group(_lowercase , _lowercase ).arrange(_lowercase , buff=0.5 , aligned_edge=_lowercase ) checkpoint.move_to([3, 0.5, 0] ) self.add(_lowercase ) __a : Dict = [] __a : int = [] for i, rect in enumerate(_lowercase ): __a : List[str] = fill.copy().set_fill(_lowercase , opacity=0.7 ) target.move_to(_lowercase ) ckpt_arr.append(_lowercase ) __a : Union[str, Any] = target.copy() if i < 5: cpu_target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.move_to(cpu_right_col_base[i - 5] ) ckpt_cpu_arr.append(_lowercase ) self.add(*_lowercase , *_lowercase ) __a : List[str] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) __a : List[Any] = MarkupText( F'''<b>Key:</b>\n\n<span fgcolor=\'{YELLOW}\'>●</span> Empty Model''' , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(_lowercase , _lowercase ) __a : str = MarkupText( F'''<span fgcolor=\'{BLUE}\'>●</span> Checkpoint''' , font_size=18 , ) blue_text.next_to(_lowercase , DOWN * 2.4 , aligned_edge=key_text.get_left() ) self.add(_lowercase ) __a : Optional[int] = MarkupText( F'''Based on the passed in configuration, weights are stored in\na variety of np.memmaps on disk or to a particular device.''' , font_size=24 , ) step_a.move_to([2, 2, 0] ) __a : List[Any] = [meta_mem.copy() for i in range(6 )] __a : Optional[int] = [meta_mem.copy() for i in range(6 )] __a : List[Any] = VGroup(*_lowercase ).arrange(_lowercase , buff=0 ) __a : List[str] = VGroup(*_lowercase ).arrange(_lowercase , buff=0 ) __a : Tuple = VGroup(_lowercase , _lowercase ).arrange(_lowercase , buff=0 ) __a : Dict = Text("""Disk""" , font_size=24 ) __a : Dict = Group(_lowercase , _lowercase ).arrange(_lowercase , buff=0.5 , aligned_edge=_lowercase ) disk.move_to([-4.0, -1.25, 0] ) self.play(Write(_lowercase , run_time=3 ) , Write(_lowercase , run_time=1 ) , Create(_lowercase , run_time=1 ) ) __a : Optional[Any] = [] for i, rect in enumerate(_lowercase ): __a : List[str] = rect.copy() target.generate_target() target.target.move_to(disk_left_col_base[i] ).scale(0.5 ) animations.append(MoveToTarget(_lowercase , run_time=1.5 ) ) self.play(*_lowercase ) self.play(FadeOut(_lowercase ) ) __a : List[str] = MarkupText(F'''Then, the checkpoint is removed from memory\nthrough garbage collection.''' , font_size=24 ) step_a.move_to([2, 2, 0] ) self.play(Write(_lowercase , run_time=3 ) ) self.play( FadeOut(_lowercase , _lowercase , *_lowercase , *_lowercase ) , ) self.wait()
63
0
"""simple docstring""" lowercase__ = { "a": "AAAAA", "b": "AAAAB", "c": "AAABA", "d": "AAABB", "e": "AABAA", "f": "AABAB", "g": "AABBA", "h": "AABBB", "i": "ABAAA", "j": "BBBAA", "k": "ABAAB", "l": "ABABA", "m": "ABABB", "n": "ABBAA", "o": "ABBAB", "p": "ABBBA", "q": "ABBBB", "r": "BAAAA", "s": "BAAAB", "t": "BAABA", "u": "BAABB", "v": "BBBAB", "w": "BABAA", "x": "BABAB", "y": "BABBA", "z": "BABBB", " ": " ", } lowercase__ = {value: key for key, value in encode_dict.items()} def __magic_name__ ( _lowerCamelCase : str ): __a : List[Any] = """""" for letter in word.lower(): if letter.isalpha() or letter == " ": encoded += encode_dict[letter] else: raise Exception("""encode() accepts only letters of the alphabet and spaces""" ) return encoded def __magic_name__ ( _lowerCamelCase : str ): if set(_lowerCamelCase ) - {"A", "B", " "} != set(): raise Exception("""decode() accepts only 'A', 'B' and spaces""" ) __a : Tuple = """""" for word in coded.split(): while len(_lowerCamelCase ) != 0: decoded += decode_dict[word[:5]] __a : Optional[int] = word[5:] decoded += " " return decoded.strip() if __name__ == "__main__": from doctest import testmod testmod()
701
"""simple docstring""" def __magic_name__ ( _lowerCamelCase : int , _lowerCamelCase : float , _lowerCamelCase : float ): return round(float(moles / volume ) * nfactor ) def __magic_name__ ( _lowerCamelCase : float , _lowerCamelCase : float , _lowerCamelCase : float ): return round(float((moles * 0.08_21 * temperature) / (volume) ) ) def __magic_name__ ( _lowerCamelCase : float , _lowerCamelCase : float , _lowerCamelCase : float ): return round(float((moles * 0.08_21 * temperature) / (pressure) ) ) def __magic_name__ ( _lowerCamelCase : float , _lowerCamelCase : float , _lowerCamelCase : float ): return round(float((pressure * volume) / (0.08_21 * moles) ) ) if __name__ == "__main__": import doctest doctest.testmod()
63
0
import math from enum import Enum from typing import Optional, Union from torch.optim import Optimizer from torch.optim.lr_scheduler import LambdaLR from .utils import logging lowercase__ = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = "linear" _lowerCAmelCase = "cosine" _lowerCAmelCase = "cosine_with_restarts" _lowerCAmelCase = "polynomial" _lowerCAmelCase = "constant" _lowerCAmelCase = "constant_with_warmup" _lowerCAmelCase = "piecewise_constant" def __magic_name__ ( _lowerCamelCase : Optimizer , _lowerCamelCase : int = -1 ): return LambdaLR(_lowerCamelCase , lambda _lowerCamelCase : 1 , last_epoch=_lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : Optimizer , _lowerCamelCase : int , _lowerCamelCase : int = -1 ): def lr_lambda(_lowerCamelCase : int ): if current_step < num_warmup_steps: return float(_lowerCamelCase ) / float(max(1.0 , _lowerCamelCase ) ) return 1.0 return LambdaLR(_lowerCamelCase , _lowerCamelCase , last_epoch=_lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : Optimizer , _lowerCamelCase : str , _lowerCamelCase : int = -1 ): __a : Optional[int] = {} __a : Any = step_rules.split(""",""" ) for rule_str in rule_list[:-1]: __a : int = rule_str.split(""":""" ) __a : Optional[int] = int(_lowerCamelCase ) __a : str = float(_lowerCamelCase ) __a : int = value __a : Dict = float(rule_list[-1] ) def create_rules_function(_lowerCamelCase : str , _lowerCamelCase : Tuple ): def rule_func(_lowerCamelCase : int ) -> float: __a : Optional[Any] = sorted(rules_dict.keys() ) for i, sorted_step in enumerate(_lowerCamelCase ): if steps < sorted_step: return rules_dict[sorted_steps[i]] return last_lr_multiple return rule_func __a : Optional[int] = create_rules_function(_lowerCamelCase , _lowerCamelCase ) return LambdaLR(_lowerCamelCase , _lowerCamelCase , last_epoch=_lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : Optional[Any] , _lowerCamelCase : Any , _lowerCamelCase : List[str] , _lowerCamelCase : str=-1 ): def lr_lambda(_lowerCamelCase : int ): if current_step < num_warmup_steps: return float(_lowerCamelCase ) / float(max(1 , _lowerCamelCase ) ) return max( 0.0 , float(num_training_steps - current_step ) / float(max(1 , num_training_steps - num_warmup_steps ) ) ) return LambdaLR(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : Optimizer , _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : float = 0.5 , _lowerCamelCase : int = -1 ): def lr_lambda(_lowerCamelCase : Any ): if current_step < num_warmup_steps: return float(_lowerCamelCase ) / float(max(1 , _lowerCamelCase ) ) __a : Dict = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * float(_lowerCamelCase ) * 2.0 * progress )) ) return LambdaLR(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : Optimizer , _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : int = 1 , _lowerCamelCase : int = -1 ): def lr_lambda(_lowerCamelCase : Optional[int] ): if current_step < num_warmup_steps: return float(_lowerCamelCase ) / float(max(1 , _lowerCamelCase ) ) __a : Dict = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) if progress >= 1.0: return 0.0 return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * ((float(_lowerCamelCase ) * progress) % 1.0) )) ) return LambdaLR(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : Any , _lowerCamelCase : Any , _lowerCamelCase : Optional[Any] , _lowerCamelCase : Optional[Any]=1E-7 , _lowerCamelCase : Optional[int]=1.0 , _lowerCamelCase : Optional[int]=-1 ): __a : Union[str, Any] = optimizer.defaults["""lr"""] if not (lr_init > lr_end): raise ValueError(F'''lr_end ({lr_end}) must be be smaller than initial lr ({lr_init})''' ) def lr_lambda(_lowerCamelCase : int ): if current_step < num_warmup_steps: return float(_lowerCamelCase ) / float(max(1 , _lowerCamelCase ) ) elif current_step > num_training_steps: return lr_end / lr_init # as LambdaLR multiplies by lr_init else: __a : Tuple = lr_init - lr_end __a : int = num_training_steps - num_warmup_steps __a : Optional[int] = 1 - (current_step - num_warmup_steps) / decay_steps __a : List[str] = lr_range * pct_remaining**power + lr_end return decay / lr_init # as LambdaLR multiplies by lr_init return LambdaLR(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) lowercase__ = { SchedulerType.LINEAR: get_linear_schedule_with_warmup, SchedulerType.COSINE: get_cosine_schedule_with_warmup, SchedulerType.COSINE_WITH_RESTARTS: get_cosine_with_hard_restarts_schedule_with_warmup, SchedulerType.POLYNOMIAL: get_polynomial_decay_schedule_with_warmup, SchedulerType.CONSTANT: get_constant_schedule, SchedulerType.CONSTANT_WITH_WARMUP: get_constant_schedule_with_warmup, SchedulerType.PIECEWISE_CONSTANT: get_piecewise_constant_schedule, } def __magic_name__ ( _lowerCamelCase : Union[str, SchedulerType] , _lowerCamelCase : Optimizer , _lowerCamelCase : Optional[str] = None , _lowerCamelCase : Optional[int] = None , _lowerCamelCase : Optional[int] = None , _lowerCamelCase : int = 1 , _lowerCamelCase : float = 1.0 , _lowerCamelCase : int = -1 , ): __a : int = SchedulerType(_lowerCamelCase ) __a : Optional[int] = TYPE_TO_SCHEDULER_FUNCTION[name] if name == SchedulerType.CONSTANT: return schedule_func(_lowerCamelCase , last_epoch=_lowerCamelCase ) if name == SchedulerType.PIECEWISE_CONSTANT: return schedule_func(_lowerCamelCase , step_rules=_lowerCamelCase , last_epoch=_lowerCamelCase ) # All other schedulers require `num_warmup_steps` if num_warmup_steps is None: raise ValueError(F'''{name} requires `num_warmup_steps`, please provide that argument.''' ) if name == SchedulerType.CONSTANT_WITH_WARMUP: return schedule_func(_lowerCamelCase , num_warmup_steps=_lowerCamelCase , last_epoch=_lowerCamelCase ) # All other schedulers require `num_training_steps` if num_training_steps is None: raise ValueError(F'''{name} requires `num_training_steps`, please provide that argument.''' ) if name == SchedulerType.COSINE_WITH_RESTARTS: return schedule_func( _lowerCamelCase , num_warmup_steps=_lowerCamelCase , num_training_steps=_lowerCamelCase , num_cycles=_lowerCamelCase , last_epoch=_lowerCamelCase , ) if name == SchedulerType.POLYNOMIAL: return schedule_func( _lowerCamelCase , num_warmup_steps=_lowerCamelCase , num_training_steps=_lowerCamelCase , power=_lowerCamelCase , last_epoch=_lowerCamelCase , ) return schedule_func( _lowerCamelCase , num_warmup_steps=_lowerCamelCase , num_training_steps=_lowerCamelCase , last_epoch=_lowerCamelCase )
702
"""simple docstring""" def __magic_name__ ( _lowerCamelCase : list[int] ): if not nums: # Makes sure that the list is not empty raise ValueError("""List is empty""" ) __a : Any = sum(_lowerCamelCase ) / len(_lowerCamelCase ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(_lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
63
0
"""simple docstring""" from copy import deepcopy import torch import torch.nn.functional as F from torch.optim import AdamW from torch.optim.lr_scheduler import LambdaLR from torch.utils.data import DataLoader from accelerate.accelerator import Accelerator from accelerate.state import GradientState from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import DistributedType, is_torch_version, set_seed def __magic_name__ ( _lowerCamelCase : Dict , _lowerCamelCase : Any , _lowerCamelCase : str , _lowerCamelCase : Optional[int] ): for param, grad_param in zip(model_a.parameters() , model_b.parameters() ): if not param.requires_grad: continue if not did_step: # Grads should not be in sync assert ( torch.allclose(param.grad , grad_param.grad ) is False ), F'''Gradients in sync when they should not be at iteration {iteration}:\nmodel_a grad ({param.grad}) == model_b grad ({grad_param.grad})''' else: # Grads should be in sync assert ( torch.allclose(param.grad , grad_param.grad ) is True ), F'''Gradients not in sync when they should be at iteration {iteration}:\nmodel_a grad ({param.grad}) != model_b grad ({grad_param.grad})''' def __magic_name__ ( _lowerCamelCase : Optional[int] , _lowerCamelCase : str , _lowerCamelCase : Dict , _lowerCamelCase : Union[str, Any] , _lowerCamelCase : Dict=True ): model.train() __a : Optional[int] = model(_lowerCamelCase ) __a : Union[str, Any] = F.mse_loss(_lowerCamelCase , target.to(output.device ) ) if not do_backward: loss /= accelerator.gradient_accumulation_steps loss.backward() else: accelerator.backward(_lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : List[str] , _lowerCamelCase : Any=False ): set_seed(4_2 ) __a : Dict = RegressionModel() __a : Optional[Any] = deepcopy(_lowerCamelCase ) __a : str = RegressionDataset(length=8_0 ) __a : int = DataLoader(_lowerCamelCase , batch_size=1_6 ) model.to(accelerator.device ) if sched: __a : List[str] = AdamW(params=model.parameters() , lr=1E-3 ) __a : Dict = AdamW(params=ddp_model.parameters() , lr=1E-3 ) __a : Tuple = LambdaLR(_lowerCamelCase , lr_lambda=lambda _lowerCamelCase : epoch**0.65 ) __a : Union[str, Any] = LambdaLR(_lowerCamelCase , lr_lambda=lambda _lowerCamelCase : epoch**0.65 ) # Make a copy of `model` if sched: __a : List[Any] = accelerator.prepare(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) else: __a : Any = accelerator.prepare(_lowerCamelCase , _lowerCamelCase ) if sched: return (model, opt, sched, dataloader, ddp_model, ddp_opt, ddp_sched) return model, ddp_model, dataloader def __magic_name__ ( _lowerCamelCase : List[Any] ): # Test when on a single CPU or GPU that the context manager does nothing __a : str = get_training_setup(_lowerCamelCase ) # Use a single batch __a : List[str] = next(iter(_lowerCamelCase ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model __a : int = accelerator.gather((ddp_input, ddp_target) ) __a : str = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(_lowerCamelCase ): step_model(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) else: # Sync grads step_model(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # Since `no_sync` is a noop, `ddp_model` and `model` grads should always be in sync check_model_parameters(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue assert torch.allclose( param.grad , ddp_param.grad ), F'''Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1_3_3_7 + iteration ) __a : List[Any] = ddp_input[torch.randperm(len(_lowerCamelCase ) )] def __magic_name__ ( _lowerCamelCase : Dict ): # Test on distributed setup that context manager behaves properly __a : List[Any] = get_training_setup(_lowerCamelCase ) # Use a single batch __a : List[Any] = next(iter(_lowerCamelCase ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model __a : List[str] = accelerator.gather((ddp_input, ddp_target) ) __a : List[Any] = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(_lowerCamelCase ): step_model(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) else: # Sync grads step_model(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue if iteration % 2 == 0: # Grads should not be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is False ), F'''Gradients in sync when they should not be:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})''' else: # Grads should be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is True ), F'''Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1_3_3_7 + iteration ) __a : Optional[Any] = ddp_input[torch.randperm(len(_lowerCamelCase ) )] def __magic_name__ ( _lowerCamelCase : Optional[int]=False , _lowerCamelCase : List[str]=False ): __a : Tuple = Accelerator( split_batches=_lowerCamelCase , dispatch_batches=_lowerCamelCase , gradient_accumulation_steps=2 ) # Test that context manager behaves properly __a : List[Any] = get_training_setup(_lowerCamelCase ) for iteration, batch in enumerate(_lowerCamelCase ): __a : Optional[Any] = batch.values() # Gather the distributed inputs and targs for the base model __a : Dict = accelerator.gather((ddp_input, ddp_target) ) __a : Any = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # Do "gradient accumulation" (noop) with accelerator.accumulate(_lowerCamelCase ): step_model(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue if ((iteration + 1) % 2 == 0) or (iteration == len(_lowerCamelCase ) - 1): # Grads should be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is True ), F'''Gradients not in sync when they should be at iteration {iteration}:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' else: # Grads should not be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is False ), F'''Gradients in sync when they should not be at iteration {iteration}:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1_3_3_7 + iteration ) __a : int = ddp_input[torch.randperm(len(_lowerCamelCase ) )] GradientState._reset_state() def __magic_name__ ( _lowerCamelCase : Optional[Any]=False , _lowerCamelCase : int=False ): __a : Any = Accelerator( split_batches=_lowerCamelCase , dispatch_batches=_lowerCamelCase , gradient_accumulation_steps=2 ) # Test that context manager behaves properly __a : Any = get_training_setup(_lowerCamelCase , _lowerCamelCase ) for iteration, batch in enumerate(_lowerCamelCase ): __a : Optional[int] = batch.values() # Gather the distributed inputs and targs for the base model __a : Optional[Any] = accelerator.gather((ddp_input, ddp_target) ) __a : Optional[int] = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" model.train() ddp_model.train() step_model(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) opt.step() if ((iteration + 1) % 2 == 0) or ((iteration + 1) == len(_lowerCamelCase )): if split_batches: sched.step() else: for _ in range(accelerator.num_processes ): sched.step() opt.zero_grad() # Perform gradient accumulation under wrapper with accelerator.accumulate(_lowerCamelCase ): step_model(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) ddp_opt.step() ddp_sched.step() ddp_opt.zero_grad() # Learning rates should be the same assert ( opt.param_groups[0]["lr"] == ddp_opt.param_groups[0]["lr"] ), F'''Learning rates found in each optimizer did not align\nopt: {opt.param_groups[0]['lr']}\nDDP opt: {ddp_opt.param_groups[0]['lr']}\n''' __a : Optional[int] = (((iteration + 1) % 2) == 0) or ((iteration + 1) == len(_lowerCamelCase )) if accelerator.num_processes > 1: check_model_parameters(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # Shuffle ddp_input on each iteration torch.manual_seed(1_3_3_7 + iteration ) GradientState._reset_state() def __magic_name__ ( ): __a : List[Any] = Accelerator() __a : str = RegressionDataset(length=8_0 ) __a : Dict = DataLoader(_lowerCamelCase , batch_size=1_6 ) __a : int = RegressionDataset(length=9_6 ) __a : List[Any] = DataLoader(_lowerCamelCase , batch_size=1_6 ) __a : List[str] = accelerator.prepare(_lowerCamelCase , _lowerCamelCase ) assert accelerator.gradient_state.active_dataloader is None for iteration, _ in enumerate(_lowerCamelCase ): assert id(accelerator.gradient_state.active_dataloader ) == id(_lowerCamelCase ) if iteration < len(_lowerCamelCase ) - 1: assert not accelerator.gradient_state.end_of_dataloader if iteration == 1: for batch_num, _ in enumerate(_lowerCamelCase ): assert id(accelerator.gradient_state.active_dataloader ) == id(_lowerCamelCase ) if batch_num < len(_lowerCamelCase ) - 1: assert not accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader assert accelerator.gradient_state.active_dataloader is None def __magic_name__ ( ): __a : int = Accelerator() __a : str = accelerator.state if state.local_process_index == 0: print("""**Test `accumulate` gradient accumulation with dataloader break**""" ) test_dataloader_break() if state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print("""**Test NOOP `no_sync` context manager**""" ) test_noop_sync(_lowerCamelCase ) if state.distributed_type in (DistributedType.MULTI_GPU, DistributedType.MULTI_CPU): if state.local_process_index == 0: print("""**Test Distributed `no_sync` context manager**""" ) test_distributed_sync(_lowerCamelCase ) if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation, """ , F'''`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**''' , ) test_gradient_accumulation(_lowerCamelCase , _lowerCamelCase ) # Currently will break on torch 2.0 +, need to investigate why if is_torch_version("""<""" , """2.0""" ) or state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation with optimizer and scheduler, """ , """`split_batches=False`, `dispatch_batches=False`**""" , ) test_gradient_accumulation_with_opt_and_scheduler() if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if not split_batch and not dispatch_batches: continue if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation with optimizer and scheduler, """ , F'''`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**''' , ) test_gradient_accumulation_with_opt_and_scheduler(_lowerCamelCase , _lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : Union[str, Any] ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
703
"""simple docstring""" import math import sys import cva import numpy as np def __magic_name__ ( _lowerCamelCase : np.ndarray , _lowerCamelCase : float ): # For applying gaussian function for each element in matrix. __a : int = math.sqrt(_lowerCamelCase ) __a : Any = 1 / (sigma * math.sqrt(2 * math.pi )) return cons * np.exp(-((img / sigma) ** 2) * 0.5 ) def __magic_name__ ( _lowerCamelCase : np.ndarray , _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : int ): __a : Any = kernel_size // 2 return img[x - half : x + half + 1, y - half : y + half + 1] def __magic_name__ ( _lowerCamelCase : int , _lowerCamelCase : float ): # Creates a gaussian kernel of given dimension. __a : int = np.zeros((kernel_size, kernel_size) ) for i in range(0 , _lowerCamelCase ): for j in range(0 , _lowerCamelCase ): __a : Any = math.sqrt( abs(i - kernel_size // 2 ) ** 2 + abs(j - kernel_size // 2 ) ** 2 ) return vec_gaussian(_lowerCamelCase , _lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : np.ndarray , _lowerCamelCase : float , _lowerCamelCase : float , _lowerCamelCase : int , ): __a : Tuple = np.zeros(img.shape ) __a : Optional[int] = get_gauss_kernel(_lowerCamelCase , _lowerCamelCase ) __a , __a : int = img.shape for i in range(kernel_size // 2 , size_x - kernel_size // 2 ): for j in range(kernel_size // 2 , size_y - kernel_size // 2 ): __a : List[str] = get_slice(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) __a : Any = img_s - img_s[kernel_size // 2, kernel_size // 2] __a : Optional[Any] = vec_gaussian(_lowerCamelCase , _lowerCamelCase ) __a : Optional[Any] = np.multiply(_lowerCamelCase , _lowerCamelCase ) __a : Any = np.multiply(_lowerCamelCase , _lowerCamelCase ) __a : Tuple = np.sum(_lowerCamelCase ) / np.sum(_lowerCamelCase ) __a : Optional[Any] = val return imga def __magic_name__ ( _lowerCamelCase : list ): __a : Optional[Any] = args[1] if args[1:] else """../image_data/lena.jpg""" __a : Union[str, Any] = float(args[2] ) if args[2:] else 1.0 __a : Optional[int] = float(args[3] ) if args[3:] else 1.0 if args[4:]: __a : Any = int(args[4] ) __a : Any = kernel_size + abs(kernel_size % 2 - 1 ) else: __a : Optional[int] = 5 return filename, spatial_variance, intensity_variance, kernel_size if __name__ == "__main__": lowercase__ , lowercase__ , lowercase__ , lowercase__ = parse_args(sys.argv) lowercase__ = cva.imread(filename, 0) cva.imshow("input image", img) lowercase__ = img / 255 lowercase__ = out.astype("float32") lowercase__ = bilateral_filter(out, spatial_variance, intensity_variance, kernel_size) lowercase__ = out * 255 lowercase__ = np.uinta(out) cva.imshow("output image", out) cva.waitKey(0) cva.destroyAllWindows()
63
0
"""simple docstring""" import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class SCREAMING_SNAKE_CASE__ ( __snake_case , unittest.TestCase ): _lowerCAmelCase = BarthezTokenizer _lowerCAmelCase = BarthezTokenizerFast _lowerCAmelCase = True _lowerCAmelCase = True def lowerCAmelCase__(self ): '''simple docstring''' super().setUp() __a : Tuple = BarthezTokenizerFast.from_pretrained("""moussaKam/mbarthez""" ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=_lowercase ) __a : List[str] = tokenizer def lowerCAmelCase__(self ): '''simple docstring''' __a : Tuple = """<pad>""" __a : Any = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_lowercase ) , _lowercase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_lowercase ) , _lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Any = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(vocab_keys[-1] , """<mask>""" ) self.assertEqual(len(_lowercase ) , 101122 ) def lowerCAmelCase__(self ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 101122 ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' __a : Dict = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] __a : str = [0, 57, 3018, 70307, 91, 2] __a : Any = self.tokenizer( _lowercase , max_length=len(_lowercase ) , padding=_lowercase , truncation=_lowercase , return_tensors="""pt""" ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) __a : Optional[int] = batch.input_ids.tolist()[0] self.assertListEqual(_lowercase , _lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' if not self.test_rust_tokenizer: return __a : Optional[Any] = self.get_tokenizer() __a : List[str] = self.get_rust_tokenizer() __a : List[str] = """I was born in 92000, and this is falsé.""" __a : Optional[Any] = tokenizer.tokenize(_lowercase ) __a : Tuple = rust_tokenizer.tokenize(_lowercase ) self.assertListEqual(_lowercase , _lowercase ) __a : Any = tokenizer.encode(_lowercase , add_special_tokens=_lowercase ) __a : List[Any] = rust_tokenizer.encode(_lowercase , add_special_tokens=_lowercase ) self.assertListEqual(_lowercase , _lowercase ) __a : Dict = self.get_rust_tokenizer() __a : Optional[Any] = tokenizer.encode(_lowercase ) __a : Tuple = rust_tokenizer.encode(_lowercase ) self.assertListEqual(_lowercase , _lowercase ) @slow def lowerCAmelCase__(self ): '''simple docstring''' __a : List[Any] = {"""input_ids""": [[0, 490, 14328, 4507, 354, 47, 43669, 95, 25, 78117, 20215, 19779, 190, 22, 400, 4, 35343, 80310, 603, 86, 24937, 105, 33438, 94762, 196, 39642, 7, 15, 15933, 173, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 10534, 87, 25, 66, 3358, 196, 55289, 8, 82961, 81, 2204, 75203, 7, 15, 763, 12956, 216, 178, 14328, 9595, 1377, 69693, 7, 448, 71021, 196, 18106, 1437, 13974, 108, 9083, 4, 49315, 7, 39, 86, 1326, 2793, 46333, 4, 448, 196, 74588, 7, 49315, 7, 39, 21, 822, 38470, 74, 21, 66723, 62480, 8, 22050, 5, 2]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. __a : Union[str, Any] = [ """Le transformeur est un modèle d'apprentissage profond introduit en 2017, """ """utilisé principalement dans le domaine du traitement automatique des langues (TAL).""", """À l'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus """ """pour gérer des données séquentielles, telles que le langage naturel, pour des tâches """ """telles que la traduction et la synthèse de texte.""", ] self.tokenizer_integration_test_util( expected_encoding=_lowercase , model_name="""moussaKam/mbarthez""" , revision="""c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6""" , sequences=_lowercase , )
704
"""simple docstring""" from unittest import TestCase from datasets import Dataset from minhash_deduplication import deduplicate_dataset, make_duplicate_clusters def __magic_name__ ( ): __a : Dict = { """repo_name""": ["""test_repo1""", """test_repo2""", """test_repo3"""], """path""": ["""test_1.py""", """test_2.py""", """unit_test.py"""], """content""": ["""a """ * 2_0, """a """ * 3_0, """b """ * 7], } __a : Optional[Any] = Dataset.from_dict(_lowerCamelCase ) return dataset class SCREAMING_SNAKE_CASE__ ( __snake_case ): def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[int] = get_dataset() __a : List[Any] = make_duplicate_clusters(_lowercase , 0.85 ) self.assertEqual(len(duplicate_clusters[0] ) , 2 ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Tuple = get_dataset() __a , __a : Optional[Any] = deduplicate_dataset(_lowercase ) self.assertEqual(len(_lowercase ) , 2 ) print(_lowercase ) self.assertEqual(duplicate_clusters[0][0]["""copies"""] , 2 ) self.assertEqual(duplicate_clusters[0][0]["""is_extreme"""] , _lowercase )
63
0
"""simple docstring""" import argparse from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_controlnet_from_original_ckpt if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() parser.add_argument( "--checkpoint_path", default=None, type=str, required=True, help="Path to the checkpoint to convert." ) parser.add_argument( "--original_config_file", type=str, required=True, help="The YAML config file corresponding to the original architecture.", ) parser.add_argument( "--num_in_channels", default=None, type=int, help="The number of input channels. If `None` number of input channels will be automatically inferred.", ) parser.add_argument( "--image_size", default=512, type=int, help=( "The image size that the model was trained on. Use 512 for Stable Diffusion v1.X and Stable Siffusion v2" " Base. Use 768 for Stable Diffusion v2." ), ) parser.add_argument( "--extract_ema", action="store_true", help=( "Only relevant for checkpoints that have both EMA and non-EMA weights. Whether to extract the EMA weights" " or not. Defaults to `False`. Add `--extract_ema` to extract the EMA weights. EMA weights usually yield" " higher quality images for inference. Non-EMA weights are usually better to continue fine-tuning." ), ) parser.add_argument( "--upcast_attention", action="store_true", help=( "Whether the attention computation should always be upcasted. This is necessary when running stable" " diffusion 2.1." ), ) parser.add_argument( "--from_safetensors", action="store_true", help="If `--checkpoint_path` is in `safetensors` format, load checkpoint with safetensors instead of PyTorch.", ) parser.add_argument( "--to_safetensors", action="store_true", help="Whether to store pipeline in safetensors format or not.", ) parser.add_argument("--dump_path", default=None, type=str, required=True, help="Path to the output model.") parser.add_argument("--device", type=str, help="Device to use (e.g. cpu, cuda:0, cuda:1, etc.)") def __magic_name__ ( _lowerCamelCase : Optional[Any] ): if string == "True": return True elif string == "False": return False else: raise ValueError(F'''could not parse string as bool {string}''' ) parser.add_argument( "--use_linear_projection", help="Override for use linear projection", required=False, type=parse_bool ) parser.add_argument("--cross_attention_dim", help="Override for cross attention_dim", required=False, type=int) lowercase__ = parser.parse_args() lowercase__ = download_controlnet_from_original_ckpt( checkpoint_path=args.checkpoint_path, original_config_file=args.original_config_file, image_size=args.image_size, extract_ema=args.extract_ema, num_in_channels=args.num_in_channels, upcast_attention=args.upcast_attention, from_safetensors=args.from_safetensors, device=args.device, use_linear_projection=args.use_linear_projection, cross_attention_dim=args.cross_attention_dim, ) controlnet.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
705
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tensorflow_text_available, is_torch_available lowercase__ = { "configuration_ernie": ["ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP", "ErnieConfig", "ErnieOnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ "ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST", "ErnieForCausalLM", "ErnieForMaskedLM", "ErnieForMultipleChoice", "ErnieForNextSentencePrediction", "ErnieForPreTraining", "ErnieForQuestionAnswering", "ErnieForSequenceClassification", "ErnieForTokenClassification", "ErnieModel", "ErniePreTrainedModel", ] if TYPE_CHECKING: from .configuration_ernie import ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP, ErnieConfig, ErnieOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ernie import ( ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST, ErnieForCausalLM, ErnieForMaskedLM, ErnieForMultipleChoice, ErnieForNextSentencePrediction, ErnieForPreTraining, ErnieForQuestionAnswering, ErnieForSequenceClassification, ErnieForTokenClassification, ErnieModel, ErniePreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
63
0
"""simple docstring""" import qiskit def __magic_name__ ( _lowerCamelCase : int , _lowerCamelCase : int ): __a : List[Any] = qiskit.Aer.get_backend("""aer_simulator""" ) # Create a Quantum Circuit acting on the q register __a : int = qiskit.QuantumCircuit(_lowerCamelCase , _lowerCamelCase ) # Map the quantum measurement to the classical bits circuit.measure([0] , [0] ) # Execute the circuit on the simulator __a : str = qiskit.execute(_lowerCamelCase , _lowerCamelCase , shots=1_0_0_0 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(_lowerCamelCase ) if __name__ == "__main__": print(f'Total count for various states are: {single_qubit_measure(1, 1)}')
706
"""simple docstring""" import math from enum import Enum from typing import Optional, Union from torch.optim import Optimizer from torch.optim.lr_scheduler import LambdaLR from .utils import logging lowercase__ = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = "linear" _lowerCAmelCase = "cosine" _lowerCAmelCase = "cosine_with_restarts" _lowerCAmelCase = "polynomial" _lowerCAmelCase = "constant" _lowerCAmelCase = "constant_with_warmup" _lowerCAmelCase = "piecewise_constant" def __magic_name__ ( _lowerCamelCase : Optimizer , _lowerCamelCase : int = -1 ): return LambdaLR(_lowerCamelCase , lambda _lowerCamelCase : 1 , last_epoch=_lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : Optimizer , _lowerCamelCase : int , _lowerCamelCase : int = -1 ): def lr_lambda(_lowerCamelCase : int ): if current_step < num_warmup_steps: return float(_lowerCamelCase ) / float(max(1.0 , _lowerCamelCase ) ) return 1.0 return LambdaLR(_lowerCamelCase , _lowerCamelCase , last_epoch=_lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : Optimizer , _lowerCamelCase : str , _lowerCamelCase : int = -1 ): __a : Optional[int] = {} __a : Any = step_rules.split(""",""" ) for rule_str in rule_list[:-1]: __a , __a : int = rule_str.split(""":""" ) __a : Optional[int] = int(_lowerCamelCase ) __a : str = float(_lowerCamelCase ) __a : int = value __a : Dict = float(rule_list[-1] ) def create_rules_function(_lowerCamelCase : str , _lowerCamelCase : Tuple ): def rule_func(_lowerCamelCase : int ) -> float: __a : Optional[Any] = sorted(rules_dict.keys() ) for i, sorted_step in enumerate(_lowerCamelCase ): if steps < sorted_step: return rules_dict[sorted_steps[i]] return last_lr_multiple return rule_func __a : Optional[int] = create_rules_function(_lowerCamelCase , _lowerCamelCase ) return LambdaLR(_lowerCamelCase , _lowerCamelCase , last_epoch=_lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : Optional[Any] , _lowerCamelCase : Any , _lowerCamelCase : List[str] , _lowerCamelCase : str=-1 ): def lr_lambda(_lowerCamelCase : int ): if current_step < num_warmup_steps: return float(_lowerCamelCase ) / float(max(1 , _lowerCamelCase ) ) return max( 0.0 , float(num_training_steps - current_step ) / float(max(1 , num_training_steps - num_warmup_steps ) ) ) return LambdaLR(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : Optimizer , _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : float = 0.5 , _lowerCamelCase : int = -1 ): def lr_lambda(_lowerCamelCase : Any ): if current_step < num_warmup_steps: return float(_lowerCamelCase ) / float(max(1 , _lowerCamelCase ) ) __a : Dict = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * float(_lowerCamelCase ) * 2.0 * progress )) ) return LambdaLR(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : Optimizer , _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : int = 1 , _lowerCamelCase : int = -1 ): def lr_lambda(_lowerCamelCase : Optional[int] ): if current_step < num_warmup_steps: return float(_lowerCamelCase ) / float(max(1 , _lowerCamelCase ) ) __a : Dict = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) if progress >= 1.0: return 0.0 return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * ((float(_lowerCamelCase ) * progress) % 1.0) )) ) return LambdaLR(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : Any , _lowerCamelCase : Any , _lowerCamelCase : Optional[Any] , _lowerCamelCase : Optional[Any]=1E-7 , _lowerCamelCase : Optional[int]=1.0 , _lowerCamelCase : Optional[int]=-1 ): __a : Union[str, Any] = optimizer.defaults["""lr"""] if not (lr_init > lr_end): raise ValueError(F'''lr_end ({lr_end}) must be be smaller than initial lr ({lr_init})''' ) def lr_lambda(_lowerCamelCase : int ): if current_step < num_warmup_steps: return float(_lowerCamelCase ) / float(max(1 , _lowerCamelCase ) ) elif current_step > num_training_steps: return lr_end / lr_init # as LambdaLR multiplies by lr_init else: __a : Tuple = lr_init - lr_end __a : int = num_training_steps - num_warmup_steps __a : Optional[int] = 1 - (current_step - num_warmup_steps) / decay_steps __a : List[str] = lr_range * pct_remaining**power + lr_end return decay / lr_init # as LambdaLR multiplies by lr_init return LambdaLR(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) lowercase__ = { SchedulerType.LINEAR: get_linear_schedule_with_warmup, SchedulerType.COSINE: get_cosine_schedule_with_warmup, SchedulerType.COSINE_WITH_RESTARTS: get_cosine_with_hard_restarts_schedule_with_warmup, SchedulerType.POLYNOMIAL: get_polynomial_decay_schedule_with_warmup, SchedulerType.CONSTANT: get_constant_schedule, SchedulerType.CONSTANT_WITH_WARMUP: get_constant_schedule_with_warmup, SchedulerType.PIECEWISE_CONSTANT: get_piecewise_constant_schedule, } def __magic_name__ ( _lowerCamelCase : Union[str, SchedulerType] , _lowerCamelCase : Optimizer , _lowerCamelCase : Optional[str] = None , _lowerCamelCase : Optional[int] = None , _lowerCamelCase : Optional[int] = None , _lowerCamelCase : int = 1 , _lowerCamelCase : float = 1.0 , _lowerCamelCase : int = -1 , ): __a : int = SchedulerType(_lowerCamelCase ) __a : Optional[int] = TYPE_TO_SCHEDULER_FUNCTION[name] if name == SchedulerType.CONSTANT: return schedule_func(_lowerCamelCase , last_epoch=_lowerCamelCase ) if name == SchedulerType.PIECEWISE_CONSTANT: return schedule_func(_lowerCamelCase , step_rules=_lowerCamelCase , last_epoch=_lowerCamelCase ) # All other schedulers require `num_warmup_steps` if num_warmup_steps is None: raise ValueError(F'''{name} requires `num_warmup_steps`, please provide that argument.''' ) if name == SchedulerType.CONSTANT_WITH_WARMUP: return schedule_func(_lowerCamelCase , num_warmup_steps=_lowerCamelCase , last_epoch=_lowerCamelCase ) # All other schedulers require `num_training_steps` if num_training_steps is None: raise ValueError(F'''{name} requires `num_training_steps`, please provide that argument.''' ) if name == SchedulerType.COSINE_WITH_RESTARTS: return schedule_func( _lowerCamelCase , num_warmup_steps=_lowerCamelCase , num_training_steps=_lowerCamelCase , num_cycles=_lowerCamelCase , last_epoch=_lowerCamelCase , ) if name == SchedulerType.POLYNOMIAL: return schedule_func( _lowerCamelCase , num_warmup_steps=_lowerCamelCase , num_training_steps=_lowerCamelCase , power=_lowerCamelCase , last_epoch=_lowerCamelCase , ) return schedule_func( _lowerCamelCase , num_warmup_steps=_lowerCamelCase , num_training_steps=_lowerCamelCase , last_epoch=_lowerCamelCase )
63
0
"""simple docstring""" from tempfile import TemporaryDirectory from unittest import TestCase from unittest.mock import MagicMock, patch from transformers import AutoModel, TFAutoModel from transformers.onnx import FeaturesManager from transformers.testing_utils import SMALL_MODEL_IDENTIFIER, require_tf, require_torch @require_torch @require_tf class SCREAMING_SNAKE_CASE__ ( __snake_case ): def lowerCAmelCase__(self ): '''simple docstring''' __a : Union[str, Any] = SMALL_MODEL_IDENTIFIER __a : List[Any] = """pt""" __a : Union[str, Any] = """tf""" def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' __a : Tuple = AutoModel.from_pretrained(self.test_model ) model_pt.save_pretrained(_lowercase ) def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' __a : List[str] = TFAutoModel.from_pretrained(self.test_model , from_pt=_lowercase ) model_tf.save_pretrained(_lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Any = """mock_framework""" # Framework provided - return whatever the user provides __a : str = FeaturesManager.determine_framework(self.test_model , _lowercase ) self.assertEqual(_lowercase , _lowercase ) # Local checkpoint and framework provided - return provided framework # PyTorch checkpoint with TemporaryDirectory() as local_pt_ckpt: self._setup_pt_ckpt(_lowercase ) __a : int = FeaturesManager.determine_framework(_lowercase , _lowercase ) self.assertEqual(_lowercase , _lowercase ) # TensorFlow checkpoint with TemporaryDirectory() as local_tf_ckpt: self._setup_tf_ckpt(_lowercase ) __a : List[str] = FeaturesManager.determine_framework(_lowercase , _lowercase ) self.assertEqual(_lowercase , _lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' with TemporaryDirectory() as local_pt_ckpt: self._setup_pt_ckpt(_lowercase ) __a : List[Any] = FeaturesManager.determine_framework(_lowercase ) self.assertEqual(_lowercase , self.framework_pt ) # TensorFlow checkpoint with TemporaryDirectory() as local_tf_ckpt: self._setup_tf_ckpt(_lowercase ) __a : str = FeaturesManager.determine_framework(_lowercase ) self.assertEqual(_lowercase , self.framework_tf ) # Invalid local checkpoint with TemporaryDirectory() as local_invalid_ckpt: with self.assertRaises(_lowercase ): __a : str = FeaturesManager.determine_framework(_lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' __a : List[str] = MagicMock(return_value=_lowercase ) with patch("""transformers.onnx.features.is_tf_available""" , _lowercase ): __a : Dict = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_lowercase , self.framework_pt ) # PyTorch not in environment -> use TensorFlow __a : Tuple = MagicMock(return_value=_lowercase ) with patch("""transformers.onnx.features.is_torch_available""" , _lowercase ): __a : Union[str, Any] = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_lowercase , self.framework_tf ) # Both in environment -> use PyTorch __a : List[Any] = MagicMock(return_value=_lowercase ) __a : List[Any] = MagicMock(return_value=_lowercase ) with patch("""transformers.onnx.features.is_tf_available""" , _lowercase ), patch( """transformers.onnx.features.is_torch_available""" , _lowercase ): __a : int = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_lowercase , self.framework_pt ) # Both not in environment -> raise error __a : int = MagicMock(return_value=_lowercase ) __a : List[str] = MagicMock(return_value=_lowercase ) with patch("""transformers.onnx.features.is_tf_available""" , _lowercase ), patch( """transformers.onnx.features.is_torch_available""" , _lowercase ): with self.assertRaises(_lowercase ): __a : Optional[Any] = FeaturesManager.determine_framework(self.test_model )
707
"""simple docstring""" import importlib import torch import yaml from omegaconf import OmegaConf from taming.models.vqgan import VQModel def __magic_name__ ( _lowerCamelCase : int , _lowerCamelCase : Optional[Any]=False ): __a : Dict = OmegaConf.load(_lowerCamelCase ) if display: print(yaml.dump(OmegaConf.to_container(_lowerCamelCase ) ) ) return config def __magic_name__ ( _lowerCamelCase : str , _lowerCamelCase : Union[str, Any]=None , _lowerCamelCase : int=None ): if conf_path is None: __a : str = """./model_checkpoints/vqgan_only.yaml""" __a : List[Any] = load_config(_lowerCamelCase , display=_lowerCamelCase ) __a : Dict = VQModel(**config.model.params ) if ckpt_path is None: __a : List[Any] = """./model_checkpoints/vqgan_only.pt""" __a : Tuple = torch.load(_lowerCamelCase , map_location=_lowerCamelCase ) if ".ckpt" in ckpt_path: __a : List[str] = sd["""state_dict"""] model.load_state_dict(_lowerCamelCase , strict=_lowerCamelCase ) model.to(_lowerCamelCase ) del sd return model def __magic_name__ ( _lowerCamelCase : Optional[int] , _lowerCamelCase : List[str] ): __a , __a , __a : Tuple = model.encode(_lowerCamelCase ) print(F'''VQGAN --- {model.__class__.__name__}: latent shape: {z.shape[2:]}''' ) __a : Union[str, Any] = model.decode(_lowerCamelCase ) return xrec def __magic_name__ ( _lowerCamelCase : Optional[int] , _lowerCamelCase : Union[str, Any]=False ): __a , __a : Optional[Any] = string.rsplit(""".""" , 1 ) if reload: __a : Optional[Any] = importlib.import_module(_lowerCamelCase ) importlib.reload(_lowerCamelCase ) return getattr(importlib.import_module(_lowerCamelCase , package=_lowerCamelCase ) , cls ) def __magic_name__ ( _lowerCamelCase : Any ): if "target" not in config: raise KeyError("""Expected key `target` to instantiate.""" ) return get_obj_from_str(config["""target"""] )(**config.get("""params""" , {} ) ) def __magic_name__ ( _lowerCamelCase : Any , _lowerCamelCase : Dict , _lowerCamelCase : int=True , _lowerCamelCase : int=True ): __a : Union[str, Any] = instantiate_from_config(_lowerCamelCase ) if sd is not None: model.load_state_dict(_lowerCamelCase ) if gpu: model.cuda() if eval_mode: model.eval() return {"model": model} def __magic_name__ ( _lowerCamelCase : Any , _lowerCamelCase : Union[str, Any] , _lowerCamelCase : str , _lowerCamelCase : int ): # load the specified checkpoint if ckpt: __a : List[str] = torch.load(_lowerCamelCase , map_location="""cpu""" ) __a : Any = pl_sd["""global_step"""] print(F'''loaded model from global step {global_step}.''' ) else: __a : List[Any] = {"""state_dict""": None} __a : Any = None __a : Union[str, Any] = load_model_from_config(config.model , pl_sd["""state_dict"""] , gpu=_lowerCamelCase , eval_mode=_lowerCamelCase )["""model"""] return model, global_step
63
0
"""simple docstring""" import warnings from ...utils import logging from .image_processing_yolos import YolosImageProcessor lowercase__ = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( __snake_case ): def __init__(self , *_lowercase , **_lowercase ): '''simple docstring''' warnings.warn( """The class YolosFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use YolosImageProcessor instead.""" , _lowercase , ) super().__init__(*_lowercase , **_lowercase )
708
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) lowercase__ = { "configuration_llama": ["LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP", "LlamaConfig"], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = ["LlamaTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = ["LlamaTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ "LlamaForCausalLM", "LlamaModel", "LlamaPreTrainedModel", "LlamaForSequenceClassification", ] if TYPE_CHECKING: from .configuration_llama import LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP, LlamaConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama import LlamaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama_fast import LlamaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_llama import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaPreTrainedModel else: import sys lowercase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
63
0
"""simple docstring""" from __future__ import annotations from random import random class SCREAMING_SNAKE_CASE__ : def __init__(self , _lowercase = None ): '''simple docstring''' __a : List[Any] = value __a : str = random() __a : Node | None = None __a : Node | None = None def __repr__(self ): '''simple docstring''' from pprint import pformat if self.left is None and self.right is None: return F'''\'{self.value}: {self.prior:.5}\'''' else: return pformat( {F'''{self.value}: {self.prior:.5}''': (self.left, self.right)} , indent=1 ) def __str__(self ): '''simple docstring''' __a : str = str(self.value ) + """ """ __a : List[Any] = str(self.left or """""" ) __a : Optional[Any] = str(self.right or """""" ) return value + left + right def __magic_name__ ( _lowerCamelCase : Node | None , _lowerCamelCase : int ): if root is None: # None tree is split into 2 Nones return None, None elif root.value is None: return None, None else: if value < root.value: __a : Union[str, Any] = split(root.left , _lowerCamelCase ) return left, root else: __a : Any = split(root.right , _lowerCamelCase ) return root, right def __magic_name__ ( _lowerCamelCase : Node | None , _lowerCamelCase : Node | None ): if (not left) or (not right): # If one node is None, return the other return left or right elif left.prior < right.prior: __a : str = merge(left.right , _lowerCamelCase ) return left else: __a : List[Any] = merge(_lowerCamelCase , right.left ) return right def __magic_name__ ( _lowerCamelCase : Node | None , _lowerCamelCase : int ): __a : Optional[int] = Node(_lowerCamelCase ) __a : Dict = split(_lowerCamelCase , _lowerCamelCase ) return merge(merge(_lowerCamelCase , _lowerCamelCase ) , _lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : Node | None , _lowerCamelCase : int ): __a : Any = split(_lowerCamelCase , value - 1 ) __a : List[Any] = split(_lowerCamelCase , _lowerCamelCase ) return merge(_lowerCamelCase , _lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : Node | None ): if not root: # None return else: inorder(root.left ) print(root.value , end=""",""" ) inorder(root.right ) def __magic_name__ ( _lowerCamelCase : Node | None , _lowerCamelCase : str ): for arg in args.split(): if arg[0] == "+": __a : Optional[int] = insert(_lowerCamelCase , int(arg[1:] ) ) elif arg[0] == "-": __a : Any = erase(_lowerCamelCase , int(arg[1:] ) ) else: print("""Unknown command""" ) return root def __magic_name__ ( ): __a : Optional[Any] = None print( """enter numbers to create a tree, + value to add value into treap, """ """- value to erase all nodes with value. 'q' to quit. """ ) __a : Tuple = input() while args != "q": __a : Tuple = interact_treap(_lowerCamelCase , _lowerCamelCase ) print(_lowerCamelCase ) __a : Dict = input() print("""good by!""" ) if __name__ == "__main__": import doctest doctest.testmod() main()
709
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { "microsoft/unispeech-large-1500h-cv": ( "https://huggingface.co/microsoft/unispeech-large-1500h-cv/resolve/main/config.json" ), # See all UniSpeech models at https://huggingface.co/models?filter=unispeech } class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = "unispeech" def __init__(self , _lowercase=32 , _lowercase=768 , _lowercase=12 , _lowercase=12 , _lowercase=3072 , _lowercase="gelu" , _lowercase=0.1 , _lowercase=0.1 , _lowercase=0.1 , _lowercase=0.0 , _lowercase=0.0 , _lowercase=0.1 , _lowercase=0.1 , _lowercase=0.02 , _lowercase=1e-5 , _lowercase="group" , _lowercase="gelu" , _lowercase=(512, 512, 512, 512, 512, 512, 512) , _lowercase=(5, 2, 2, 2, 2, 2, 2) , _lowercase=(10, 3, 3, 3, 3, 2, 2) , _lowercase=False , _lowercase=128 , _lowercase=16 , _lowercase=False , _lowercase=True , _lowercase=0.05 , _lowercase=10 , _lowercase=2 , _lowercase=0.0 , _lowercase=10 , _lowercase=0 , _lowercase=320 , _lowercase=2 , _lowercase=0.1 , _lowercase=100 , _lowercase=256 , _lowercase=256 , _lowercase=0.1 , _lowercase="mean" , _lowercase=False , _lowercase=False , _lowercase=256 , _lowercase=80 , _lowercase=0 , _lowercase=1 , _lowercase=2 , _lowercase=0.5 , **_lowercase , ): '''simple docstring''' super().__init__(**_lowercase , pad_token_id=_lowercase , bos_token_id=_lowercase , eos_token_id=_lowercase ) __a : Union[str, Any] = hidden_size __a : Any = feat_extract_norm __a : Union[str, Any] = feat_extract_activation __a : Tuple = list(_lowercase ) __a : Dict = list(_lowercase ) __a : List[Any] = list(_lowercase ) __a : List[Any] = conv_bias __a : Optional[Any] = num_conv_pos_embeddings __a : Union[str, Any] = num_conv_pos_embedding_groups __a : Dict = len(self.conv_dim ) __a : Dict = num_hidden_layers __a : Union[str, Any] = intermediate_size __a : List[str] = hidden_act __a : int = num_attention_heads __a : int = hidden_dropout __a : Any = attention_dropout __a : List[Any] = activation_dropout __a : List[Any] = feat_proj_dropout __a : Union[str, Any] = final_dropout __a : str = layerdrop __a : Dict = layer_norm_eps __a : Dict = initializer_range __a : Union[str, Any] = num_ctc_classes __a : List[Any] = vocab_size __a : Any = do_stable_layer_norm __a : List[str] = use_weighted_layer_sum __a : List[str] = classifier_proj_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==""" """ `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =""" F''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' F''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __a : Dict = apply_spec_augment __a : Union[str, Any] = mask_time_prob __a : List[str] = mask_time_length __a : Dict = mask_time_min_masks __a : List[Any] = mask_feature_prob __a : Tuple = mask_feature_length __a : int = mask_feature_min_masks # parameters for pretraining with codevector quantized representations __a : List[Any] = num_codevectors_per_group __a : Union[str, Any] = num_codevector_groups __a : List[Any] = contrastive_logits_temperature __a : Any = feat_quantizer_dropout __a : Optional[int] = num_negatives __a : List[str] = codevector_dim __a : List[Any] = proj_codevector_dim __a : Tuple = diversity_loss_weight # ctc loss __a : Any = ctc_loss_reduction __a : List[str] = ctc_zero_infinity # pretraining loss __a : Tuple = replace_prob @property def lowerCAmelCase__(self ): '''simple docstring''' return functools.reduce(operator.mul , self.conv_stride , 1 )
63
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { "facebook/s2t-wav2vec2-large-en-de": ( "https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/config.json" ), # See all Speech2Text models at https://huggingface.co/models?filter=speech2text2 } class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = "speech_to_text_2" _lowerCAmelCase = ["past_key_values"] _lowerCAmelCase = {"num_attention_heads": "decoder_attention_heads", "hidden_size": "d_model"} def __init__(self , _lowercase=10000 , _lowercase=6 , _lowercase=2048 , _lowercase=4 , _lowercase=0.0 , _lowercase=True , _lowercase="relu" , _lowercase=256 , _lowercase=0.1 , _lowercase=0.0 , _lowercase=0.0 , _lowercase=0.02 , _lowercase=2 , _lowercase=True , _lowercase=1 , _lowercase=0 , _lowercase=2 , _lowercase=1024 , **_lowercase , ): '''simple docstring''' __a : Optional[int] = vocab_size __a : Any = d_model __a : List[str] = decoder_ffn_dim __a : str = decoder_layers __a : Optional[Any] = decoder_attention_heads __a : Optional[int] = dropout __a : Tuple = attention_dropout __a : int = activation_dropout __a : List[str] = activation_function __a : int = init_std __a : Dict = decoder_layerdrop __a : Dict = use_cache __a : Tuple = decoder_layers __a : Any = scale_embedding # scale factor will be sqrt(d_model) if True __a : Optional[Any] = max_target_positions super().__init__( pad_token_id=_lowercase , bos_token_id=_lowercase , eos_token_id=_lowercase , decoder_start_token_id=_lowercase , **_lowercase , )
710
"""simple docstring""" import inspect import unittest from typing import List import numpy as np from transformers import EfficientFormerConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, ) from transformers.models.efficientformer.modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_vision_available(): from PIL import Image from transformers import EfficientFormerImageProcessor class SCREAMING_SNAKE_CASE__ : def __init__(self , _lowercase , _lowercase = 13 , _lowercase = 64 , _lowercase = 2 , _lowercase = 3 , _lowercase = 3 , _lowercase = True , _lowercase = True , _lowercase = 128 , _lowercase=[16, 32, 64, 128] , _lowercase = 7 , _lowercase = 4 , _lowercase = 37 , _lowercase = "gelu" , _lowercase = 0.1 , _lowercase = 0.1 , _lowercase = 10 , _lowercase = 0.02 , _lowercase = 2 , _lowercase = 1 , _lowercase = 128 , _lowercase = [2, 2, 2, 2] , _lowercase = 2 , _lowercase = 2 , ): '''simple docstring''' __a : str = parent __a : List[Any] = batch_size __a : int = image_size __a : Tuple = patch_size __a : str = num_channels __a : Union[str, Any] = is_training __a : List[Any] = use_labels __a : int = hidden_size __a : Optional[Any] = num_hidden_layers __a : List[Any] = num_attention_heads __a : Dict = intermediate_size __a : str = hidden_act __a : Dict = hidden_dropout_prob __a : str = attention_probs_dropout_prob __a : Optional[int] = type_sequence_label_size __a : Dict = initializer_range __a : Dict = encoder_stride __a : int = num_attention_outputs __a : List[Any] = embed_dim __a : Optional[Any] = embed_dim + 1 __a : Optional[Any] = resolution __a : Optional[Any] = depths __a : Union[str, Any] = hidden_sizes __a : List[str] = dim __a : Any = mlp_expansion_ratio def lowerCAmelCase__(self ): '''simple docstring''' __a : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __a : str = None if self.use_labels: __a : List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __a : List[str] = self.get_config() return config, pixel_values, labels def lowerCAmelCase__(self ): '''simple docstring''' return EfficientFormerConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowercase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , resolution=self.resolution , depths=self.depths , hidden_sizes=self.hidden_sizes , dim=self.dim , mlp_expansion_ratio=self.mlp_expansion_ratio , ) def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase ): '''simple docstring''' __a : Optional[Any] = TFEfficientFormerModel(config=_lowercase ) __a : List[Any] = model(_lowercase , training=_lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase ): '''simple docstring''' __a : Optional[Any] = self.type_sequence_label_size __a : Any = TFEfficientFormerForImageClassification(_lowercase ) __a : Union[str, Any] = model(_lowercase , labels=_lowercase , training=_lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __a : Optional[Any] = 1 __a : int = TFEfficientFormerForImageClassification(_lowercase ) __a : List[str] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __a : str = model(_lowercase , labels=_lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Any = self.prepare_config_and_inputs() __a , __a , __a : Tuple = config_and_inputs __a : Tuple = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class SCREAMING_SNAKE_CASE__ ( __snake_case , __snake_case , unittest.TestCase ): _lowerCAmelCase = ( ( TFEfficientFormerModel, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerForImageClassification, ) if is_tf_available() else () ) _lowerCAmelCase = ( { "feature-extraction": TFEfficientFormerModel, "image-classification": ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, ), } if is_tf_available() else {} ) _lowerCAmelCase = False _lowerCAmelCase = False _lowerCAmelCase = False _lowerCAmelCase = False _lowerCAmelCase = False def lowerCAmelCase__(self ): '''simple docstring''' __a : Tuple = TFEfficientFormerModelTester(self ) __a : Any = ConfigTester( self , config_class=_lowercase , has_text_modality=_lowercase , hidden_size=37 ) def lowerCAmelCase__(self ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="""EfficientFormer does not use inputs_embeds""" ) def lowerCAmelCase__(self ): '''simple docstring''' pass @unittest.skip(reason="""EfficientFormer does not support input and output embeddings""" ) def lowerCAmelCase__(self ): '''simple docstring''' pass def lowerCAmelCase__(self ): '''simple docstring''' __a , __a : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a : Dict = model_class(_lowercase ) __a : Optional[Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __a : Optional[Any] = [*signature.parameters.keys()] __a : Union[str, Any] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' def check_hidden_states_output(_lowercase , _lowercase , _lowercase ): __a : Tuple = model_class(_lowercase ) __a : int = model(**self._prepare_for_class(_lowercase , _lowercase ) , training=_lowercase ) __a : Tuple = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __a : str = getattr( self.model_tester , """expected_num_hidden_layers""" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(_lowercase ) , _lowercase ) if hasattr(self.model_tester , """encoder_seq_length""" ): __a : Any = self.model_tester.encoder_seq_length if hasattr(self.model_tester , """chunk_length""" ) and self.model_tester.chunk_length > 1: __a : int = seq_length * self.model_tester.chunk_length else: __a : Any = self.model_tester.seq_length self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) if config.is_encoder_decoder: __a : Optional[int] = outputs.decoder_hidden_states self.asseretIsInstance(_lowercase , (list, tuple) ) self.assertEqual(len(_lowercase ) , _lowercase ) __a : Any = getattr(self.model_tester , """seq_length""" , _lowercase ) __a : List[Any] = getattr(self.model_tester , """decoder_seq_length""" , _lowercase ) self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [decoder_seq_length, self.model_tester.hidden_size] , ) __a , __a : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a : Dict = True check_hidden_states_output(_lowercase , _lowercase , _lowercase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __a : int = True check_hidden_states_output(_lowercase , _lowercase , _lowercase ) def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase=False ): '''simple docstring''' __a : Any = super()._prepare_for_class(_lowercase , _lowercase , return_labels=_lowercase ) if return_labels: if model_class.__name__ == "TFEfficientFormerForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowercase ) @unittest.skip(reason="""EfficientFormer does not implement masked image modeling yet""" ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowercase ) @slow def lowerCAmelCase__(self ): '''simple docstring''' for model_name in TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a : Union[str, Any] = TFEfficientFormerModel.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' __a , __a : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() __a : int = True __a : Optional[int] = getattr(self.model_tester , """seq_length""" , _lowercase ) __a : Dict = getattr(self.model_tester , """encoder_seq_length""" , _lowercase ) __a : Dict = getattr(self.model_tester , """key_length""" , _lowercase ) __a : int = getattr(self.model_tester , """chunk_length""" , _lowercase ) if chunk_length is not None and hasattr(self.model_tester , """num_hashes""" ): __a : List[str] = encoder_seq_length * self.model_tester.num_hashes for model_class in self.all_model_classes: __a : List[Any] = True __a : Tuple = False __a : List[Any] = True __a : int = model_class(_lowercase ) __a : List[Any] = model(**self._prepare_for_class(_lowercase , _lowercase ) , training=_lowercase ) __a : Dict = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(_lowercase ) , self.model_tester.num_attention_outputs ) # check that output_attentions also work using config del inputs_dict["output_attentions"] __a : Optional[Any] = True __a : List[str] = model_class(_lowercase ) __a : Dict = model(**self._prepare_for_class(_lowercase , _lowercase ) , training=_lowercase ) __a : int = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(_lowercase ) , self.model_tester.num_attention_outputs ) if chunk_length is not None: self.assertListEqual( list(attentions[0].shape[-4:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, chunk_length, encoder_key_length] , ) else: self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, encoder_key_length] , ) def lowerCAmelCase__(self ): '''simple docstring''' __a , __a : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # Prepare our model __a : Dict = model_class(_lowercase ) # These are maximally general inputs for the model, with multiple None dimensions # Hopefully this will catch any conditionals that fail for flexible shapes __a : Optional[Any] = { key: tf.keras.Input(shape=val.shape[1:] , dtype=val.dtype , name=_lowercase ) for key, val in model.input_signature.items() if key in model.dummy_inputs } __a : Optional[Any] = model(_lowercase ) self.assertTrue(outputs_dict is not None ) def __magic_name__ ( ): __a : int = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @cached_property def lowerCAmelCase__(self ): '''simple docstring''' return ( EfficientFormerImageProcessor.from_pretrained("""snap-research/efficientformer-l1-300""" ) if is_vision_available() else None ) @slow def lowerCAmelCase__(self ): '''simple docstring''' __a : str = TFEfficientFormerForImageClassification.from_pretrained("""snap-research/efficientformer-l1-300""" ) __a : Optional[Any] = self.default_image_processor __a : List[str] = prepare_img() __a : int = image_processor(images=_lowercase , return_tensors="""tf""" ) # forward pass __a : Optional[Any] = model(**_lowercase , training=_lowercase ) # verify the logits __a : str = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowercase ) __a : Dict = tf.constant([-0.0555, 0.4825, -0.0852] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , _lowercase , atol=1e-4 ) ) @slow def lowerCAmelCase__(self ): '''simple docstring''' __a : Any = TFEfficientFormerForImageClassificationWithTeacher.from_pretrained( """snap-research/efficientformer-l1-300""" ) __a : Any = self.default_image_processor __a : str = prepare_img() __a : str = image_processor(images=_lowercase , return_tensors="""tf""" ) # forward pass __a : List[Any] = model(**_lowercase , training=_lowercase ) # verify the logits __a : int = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowercase ) __a : List[str] = tf.constant([-0.1312, 0.4353, -1.0499] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , _lowercase , atol=1e-4 ) )
63
0
"""simple docstring""" import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## lowercase__ = 16 lowercase__ = 32 def __magic_name__ ( _lowerCamelCase : Accelerator , _lowerCamelCase : int = 1_6 ): __a : Union[str, Any] = AutoTokenizer.from_pretrained("""bert-base-cased""" ) __a : List[str] = load_dataset("""glue""" , """mrpc""" ) def tokenize_function(_lowerCamelCase : List[Any] ): # max_length=None => use the model max length (it's actually the default) __a : Union[str, Any] = tokenizer(examples["""sentence1"""] , examples["""sentence2"""] , truncation=_lowerCamelCase , max_length=_lowerCamelCase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): __a : List[str] = datasets.map( _lowerCamelCase , batched=_lowerCamelCase , remove_columns=["""idx""", """sentence1""", """sentence2"""] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library __a : Optional[int] = tokenized_datasets.rename_column("""label""" , """labels""" ) def collate_fn(_lowerCamelCase : List[str] ): # On TPU it's best to pad everything to the same length or training will be very slow. __a : List[str] = 1_2_8 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": __a : List[Any] = 1_6 elif accelerator.mixed_precision != "no": __a : int = 8 else: __a : Dict = None return tokenizer.pad( _lowerCamelCase , padding="""longest""" , max_length=_lowerCamelCase , pad_to_multiple_of=_lowerCamelCase , return_tensors="""pt""" , ) # Instantiate dataloaders. __a : List[Any] = DataLoader( tokenized_datasets["""train"""] , shuffle=_lowerCamelCase , collate_fn=_lowerCamelCase , batch_size=_lowerCamelCase ) __a : int = DataLoader( tokenized_datasets["""validation"""] , shuffle=_lowerCamelCase , collate_fn=_lowerCamelCase , batch_size=_lowerCamelCase ) return train_dataloader, eval_dataloader # For testing only if os.environ.get("TESTING_MOCKED_DATALOADERS", None) == "1": from accelerate.test_utils.training import mocked_dataloaders lowercase__ = mocked_dataloaders # noqa: F811 def __magic_name__ ( _lowerCamelCase : Tuple , _lowerCamelCase : Dict ): # For testing only if os.environ.get("""TESTING_MOCKED_DATALOADERS""" , _lowerCamelCase ) == "1": __a : List[Any] = 2 # Initialize accelerator __a : int = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs __a : Dict = config["""lr"""] __a : Any = int(config["""num_epochs"""] ) __a : Optional[int] = int(config["""seed"""] ) __a : List[Any] = int(config["""batch_size"""] ) __a : Optional[Any] = evaluate.load("""glue""" , """mrpc""" ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=_lowerCamelCase ) def inner_training_loop(_lowerCamelCase : Optional[Any] ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(_lowerCamelCase ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) __a : str = AutoModelForSequenceClassification.from_pretrained("""bert-base-cased""" , return_dict=_lowerCamelCase ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). __a : str = model.to(accelerator.device ) # Instantiate optimizer __a : Optional[Any] = AdamW(params=model.parameters() , lr=_lowerCamelCase ) __a : List[Any] = get_dataloaders(_lowerCamelCase , _lowerCamelCase ) # Instantiate scheduler __a : Dict = get_linear_schedule_with_warmup( optimizer=_lowerCamelCase , num_warmup_steps=1_0_0 , num_training_steps=(len(_lowerCamelCase ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. __a : Optional[Any] = accelerator.prepare( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # Now we train the model for epoch in range(_lowerCamelCase ): model.train() for step, batch in enumerate(_lowerCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) __a : Any = model(**_lowerCamelCase ) __a : Union[str, Any] = outputs.loss accelerator.backward(_lowerCamelCase ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(_lowerCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): __a : Optional[int] = model(**_lowerCamelCase ) __a : Tuple = outputs.logits.argmax(dim=-1 ) __a : Any = accelerator.gather_for_metrics((predictions, batch["""labels"""]) ) metric.add_batch( predictions=_lowerCamelCase , references=_lowerCamelCase , ) __a : int = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F'''epoch {epoch}:''' , _lowerCamelCase ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def __magic_name__ ( ): __a : Optional[int] = argparse.ArgumentParser(description="""Simple example of training script.""" ) parser.add_argument( """--mixed_precision""" , type=_lowerCamelCase , default=_lowerCamelCase , choices=["""no""", """fp16""", """bf16""", """fp8"""] , help="""Whether to use mixed precision. Choose""" """between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.""" """and an Nvidia Ampere GPU.""" , ) parser.add_argument("""--cpu""" , action="""store_true""" , help="""If passed, will train on the CPU.""" ) __a : Optional[Any] = parser.parse_args() __a : Any = {"""lr""": 2E-5, """num_epochs""": 3, """seed""": 4_2, """batch_size""": 1_6} training_function(_lowerCamelCase , _lowerCamelCase ) if __name__ == "__main__": main()
711
"""simple docstring""" from typing import Callable, Dict, Optional, Tuple import torch from torch import nn from torch.distributions import ( AffineTransform, Distribution, Independent, NegativeBinomial, Normal, StudentT, TransformedDistribution, ) class SCREAMING_SNAKE_CASE__ ( __snake_case ): def __init__(self , _lowercase , _lowercase=None , _lowercase=None , _lowercase=0 ): '''simple docstring''' __a : Any = 1.0 if scale is None else scale __a : str = 0.0 if loc is None else loc super().__init__(_lowercase , [AffineTransform(loc=self.loc , scale=self.scale , event_dim=_lowercase )] ) @property def lowerCAmelCase__(self ): '''simple docstring''' return self.base_dist.mean * self.scale + self.loc @property def lowerCAmelCase__(self ): '''simple docstring''' return self.base_dist.variance * self.scale**2 @property def lowerCAmelCase__(self ): '''simple docstring''' return self.variance.sqrt() class SCREAMING_SNAKE_CASE__ ( nn.Module ): def __init__(self , _lowercase , _lowercase , _lowercase , **_lowercase ): '''simple docstring''' super().__init__(**_lowercase ) __a : str = args_dim __a : List[Any] = nn.ModuleList([nn.Linear(_lowercase , _lowercase ) for dim in args_dim.values()] ) __a : Dict = domain_map def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' __a : List[Any] = [proj(_lowercase ) for proj in self.proj] return self.domain_map(*_lowercase ) class SCREAMING_SNAKE_CASE__ ( nn.Module ): def __init__(self , _lowercase ): '''simple docstring''' super().__init__() __a : Optional[int] = function def lowerCAmelCase__(self , _lowercase , *_lowercase ): '''simple docstring''' return self.function(_lowercase , *_lowercase ) class SCREAMING_SNAKE_CASE__ : _lowerCAmelCase = 42 _lowerCAmelCase = 42 _lowerCAmelCase = 42 def __init__(self , _lowercase = 1 ): '''simple docstring''' __a : Optional[int] = dim __a : str = {k: dim * self.args_dim[k] for k in self.args_dim} def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' if self.dim == 1: return self.distribution_class(*_lowercase ) else: return Independent(self.distribution_class(*_lowercase ) , 1 ) def lowerCAmelCase__(self , _lowercase , _lowercase = None , _lowercase = None , ): '''simple docstring''' __a : Tuple = self._base_distribution(_lowercase ) if loc is None and scale is None: return distr else: return AffineTransformed(_lowercase , loc=_lowercase , scale=_lowercase , event_dim=self.event_dim ) @property def lowerCAmelCase__(self ): '''simple docstring''' return () if self.dim == 1 else (self.dim,) @property def lowerCAmelCase__(self ): '''simple docstring''' return len(self.event_shape ) @property def lowerCAmelCase__(self ): '''simple docstring''' return 0.0 def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' return ParameterProjection( in_features=_lowercase , args_dim=self.args_dim , domain_map=LambdaLayer(self.domain_map ) , ) def lowerCAmelCase__(self , *_lowercase ): '''simple docstring''' raise NotImplementedError() @staticmethod def lowerCAmelCase__(_lowercase ): '''simple docstring''' return (x + torch.sqrt(torch.square(_lowercase ) + 4.0 )) / 2.0 class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = {"df": 1, "loc": 1, "scale": 1} _lowerCAmelCase = StudentT @classmethod def lowerCAmelCase__(cls , _lowercase , _lowercase , _lowercase ): '''simple docstring''' __a : int = cls.squareplus(_lowercase ).clamp_min(torch.finfo(scale.dtype ).eps ) __a : Optional[Any] = 2.0 + cls.squareplus(_lowercase ) return df.squeeze(-1 ), loc.squeeze(-1 ), scale.squeeze(-1 ) class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = {"loc": 1, "scale": 1} _lowerCAmelCase = Normal @classmethod def lowerCAmelCase__(cls , _lowercase , _lowercase ): '''simple docstring''' __a : str = cls.squareplus(_lowercase ).clamp_min(torch.finfo(scale.dtype ).eps ) return loc.squeeze(-1 ), scale.squeeze(-1 ) class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = {"total_count": 1, "logits": 1} _lowerCAmelCase = NegativeBinomial @classmethod def lowerCAmelCase__(cls , _lowercase , _lowercase ): '''simple docstring''' __a : Union[str, Any] = cls.squareplus(_lowercase ) return total_count.squeeze(-1 ), logits.squeeze(-1 ) def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' __a , __a : Optional[Any] = distr_args if self.dim == 1: return self.distribution_class(total_count=_lowercase , logits=_lowercase ) else: return Independent(self.distribution_class(total_count=_lowercase , logits=_lowercase ) , 1 ) def lowerCAmelCase__(self , _lowercase , _lowercase = None , _lowercase = None ): '''simple docstring''' __a , __a : List[Any] = distr_args if scale is not None: # See scaling property of Gamma. logits += scale.log() return self._base_distribution((total_count, logits) )
63
0
"""simple docstring""" from collections import Counter from pathlib import Path from typing import Optional, Tuple import yaml class SCREAMING_SNAKE_CASE__ ( yaml.SafeLoader ): def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' __a : Optional[Any] = [self.constructed_objects[key_node] for key_node, _ in node.value] __a : Tuple = [tuple(_lowercase ) if isinstance(_lowercase , _lowercase ) else key for key in keys] __a : List[Any] = Counter(_lowercase ) __a : Tuple = [key for key in counter if counter[key] > 1] if duplicate_keys: raise TypeError(F'''Got duplicate yaml keys: {duplicate_keys}''' ) def lowerCAmelCase__(self , _lowercase , _lowercase=False ): '''simple docstring''' __a : Optional[Any] = super().construct_mapping(_lowercase , deep=_lowercase ) self._check_no_duplicates_on_constructed_node(_lowercase ) return mapping def __magic_name__ ( _lowerCamelCase : str ): __a : Optional[Any] = list(readme_content.splitlines() ) if full_content and full_content[0] == "---" and "---" in full_content[1:]: __a : Union[str, Any] = full_content[1:].index("""---""" ) + 1 __a : int = """\n""".join(full_content[1:sep_idx] ) return yamlblock, "\n".join(full_content[sep_idx + 1 :] ) return None, "\n".join(_lowerCamelCase ) class SCREAMING_SNAKE_CASE__ ( __snake_case ): # class attributes _lowerCAmelCase = {"train_eval_index"} # train-eval-index in the YAML metadata @classmethod def lowerCAmelCase__(cls , _lowercase ): '''simple docstring''' with open(_lowercase , encoding="""utf-8""" ) as readme_file: __a : Optional[Any] = _split_yaml_from_readme(readme_file.read() ) if yaml_string is not None: return cls.from_yaml_string(_lowercase ) else: return cls() def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' if path.exists(): with open(_lowercase , encoding="""utf-8""" ) as readme_file: __a : str = readme_file.read() else: __a : List[str] = None __a : Optional[Any] = self._to_readme(_lowercase ) with open(_lowercase , """w""" , encoding="""utf-8""" ) as readme_file: readme_file.write(_lowercase ) def lowerCAmelCase__(self , _lowercase = None ): '''simple docstring''' if readme_content is not None: __a : Tuple = _split_yaml_from_readme(_lowercase ) __a : str = """---\n""" + self.to_yaml_string() + """---\n""" + content else: __a : Union[str, Any] = """---\n""" + self.to_yaml_string() + """---\n""" return full_content @classmethod def lowerCAmelCase__(cls , _lowercase ): '''simple docstring''' __a : int = yaml.load(_lowercase , Loader=_NoDuplicateSafeLoader ) or {} # Convert the YAML keys to DatasetMetadata fields __a : Tuple = { (key.replace("""-""" , """_""" ) if key.replace("""-""" , """_""" ) in cls._FIELDS_WITH_DASHES else key): value for key, value in metadata_dict.items() } return cls(**_lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' return yaml.safe_dump( { (key.replace("""_""" , """-""" ) if key in self._FIELDS_WITH_DASHES else key): value for key, value in self.items() } , sort_keys=_lowercase , allow_unicode=_lowercase , encoding="""utf-8""" , ).decode("""utf-8""" ) lowercase__ = { "image-classification": [], "translation": [], "image-segmentation": [], "fill-mask": [], "automatic-speech-recognition": [], "token-classification": [], "sentence-similarity": [], "audio-classification": [], "question-answering": [], "summarization": [], "zero-shot-classification": [], "table-to-text": [], "feature-extraction": [], "other": [], "multiple-choice": [], "text-classification": [], "text-to-image": [], "text2text-generation": [], "zero-shot-image-classification": [], "tabular-classification": [], "tabular-regression": [], "image-to-image": [], "tabular-to-text": [], "unconditional-image-generation": [], "text-retrieval": [], "text-to-speech": [], "object-detection": [], "audio-to-audio": [], "text-generation": [], "conversational": [], "table-question-answering": [], "visual-question-answering": [], "image-to-text": [], "reinforcement-learning": [], "voice-activity-detection": [], "time-series-forecasting": [], "document-question-answering": [], } if __name__ == "__main__": from argparse import ArgumentParser lowercase__ = ArgumentParser(usage="Validate the yaml metadata block of a README.md file.") ap.add_argument("readme_filepath") lowercase__ = ap.parse_args() lowercase__ = Path(args.readme_filepath) lowercase__ = DatasetMetadata.from_readme(readme_filepath) print(dataset_metadata) dataset_metadata.to_readme(readme_filepath)
712
"""simple docstring""" import unittest import numpy as np import torch from torch import nn from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import KandinskyVaaPriorPipeline, PriorTransformer, UnCLIPScheduler from diffusers.utils import torch_device from diffusers.utils.testing_utils import enable_full_determinism, skip_mps from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class SCREAMING_SNAKE_CASE__ ( __snake_case , unittest.TestCase ): _lowerCAmelCase = KandinskyVaaPriorPipeline _lowerCAmelCase = ["prompt"] _lowerCAmelCase = ["prompt", "negative_prompt"] _lowerCAmelCase = [ "num_images_per_prompt", "generator", "num_inference_steps", "latents", "negative_prompt", "guidance_scale", "output_type", "return_dict", ] _lowerCAmelCase = False @property def lowerCAmelCase__(self ): '''simple docstring''' return 32 @property def lowerCAmelCase__(self ): '''simple docstring''' return 32 @property def lowerCAmelCase__(self ): '''simple docstring''' return self.time_input_dim @property def lowerCAmelCase__(self ): '''simple docstring''' return self.time_input_dim * 4 @property def lowerCAmelCase__(self ): '''simple docstring''' return 100 @property def lowerCAmelCase__(self ): '''simple docstring''' __a : List[str] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) return tokenizer @property def lowerCAmelCase__(self ): '''simple docstring''' torch.manual_seed(0 ) __a : str = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModelWithProjection(_lowercase ) @property def lowerCAmelCase__(self ): '''simple docstring''' torch.manual_seed(0 ) __a : Dict = { """num_attention_heads""": 2, """attention_head_dim""": 12, """embedding_dim""": self.text_embedder_hidden_size, """num_layers""": 1, } __a : Tuple = PriorTransformer(**_lowercase ) # clip_std and clip_mean is initialized to be 0 so PriorTransformer.post_process_latents will always return 0 - set clip_std to be 1 so it won't return 0 __a : int = nn.Parameter(torch.ones(model.clip_std.shape ) ) return model @property def lowerCAmelCase__(self ): '''simple docstring''' torch.manual_seed(0 ) __a : List[str] = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , image_size=224 , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_channels=3 , num_hidden_layers=5 , patch_size=14 , ) __a : Optional[Any] = CLIPVisionModelWithProjection(_lowercase ) return model @property def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[Any] = CLIPImageProcessor( crop_size=224 , do_center_crop=_lowercase , do_normalize=_lowercase , do_resize=_lowercase , image_mean=[0.4814_5466, 0.457_8275, 0.4082_1073] , image_std=[0.2686_2954, 0.2613_0258, 0.2757_7711] , resample=3 , size=224 , ) return image_processor def lowerCAmelCase__(self ): '''simple docstring''' __a : Union[str, Any] = self.dummy_prior __a : int = self.dummy_image_encoder __a : Any = self.dummy_text_encoder __a : int = self.dummy_tokenizer __a : Optional[Any] = self.dummy_image_processor __a : List[Any] = UnCLIPScheduler( variance_type="""fixed_small_log""" , prediction_type="""sample""" , num_train_timesteps=1000 , clip_sample=_lowercase , clip_sample_range=10.0 , ) __a : List[Any] = { """prior""": prior, """image_encoder""": image_encoder, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """scheduler""": scheduler, """image_processor""": image_processor, } return components def lowerCAmelCase__(self , _lowercase , _lowercase=0 ): '''simple docstring''' if str(_lowercase ).startswith("""mps""" ): __a : Dict = torch.manual_seed(_lowercase ) else: __a : Union[str, Any] = torch.Generator(device=_lowercase ).manual_seed(_lowercase ) __a : Union[str, Any] = { """prompt""": """horse""", """generator""": generator, """guidance_scale""": 4.0, """num_inference_steps""": 2, """output_type""": """np""", } return inputs def lowerCAmelCase__(self ): '''simple docstring''' __a : Union[str, Any] = """cpu""" __a : Union[str, Any] = self.get_dummy_components() __a : Dict = self.pipeline_class(**_lowercase ) __a : Tuple = pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) __a : Optional[int] = pipe(**self.get_dummy_inputs(_lowercase ) ) __a : str = output.image_embeds __a : Any = pipe( **self.get_dummy_inputs(_lowercase ) , return_dict=_lowercase , )[0] __a : List[Any] = image[0, -10:] __a : List[Any] = image_from_tuple[0, -10:] assert image.shape == (1, 32) __a : Optional[Any] = np.array( [-0.0532, 1.7120, 0.3656, -1.0852, -0.8946, -1.1756, 0.4348, 0.2482, 0.5146, -0.1156] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @skip_mps def lowerCAmelCase__(self ): '''simple docstring''' __a : Any = torch_device == """cpu""" __a : Any = True __a : Any = False self._test_inference_batch_single_identical( test_max_difference=_lowercase , relax_max_difference=_lowercase , test_mean_pixel_difference=_lowercase , ) @skip_mps def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[int] = torch_device == """cpu""" __a : Union[str, Any] = False self._test_attention_slicing_forward_pass( test_max_difference=_lowercase , test_mean_pixel_difference=_lowercase , )
63
0
"""simple docstring""" class SCREAMING_SNAKE_CASE__ : def __init__(self , _lowercase ): '''simple docstring''' __a : Union[str, Any] = size __a : Dict = [0] * size __a : Tuple = [0] * size @staticmethod def lowerCAmelCase__(_lowercase ): '''simple docstring''' return index | (index + 1) @staticmethod def lowerCAmelCase__(_lowercase ): '''simple docstring''' return (index & (index + 1)) - 1 def lowerCAmelCase__(self , _lowercase , _lowercase ): '''simple docstring''' __a : Union[str, Any] = value while index < self.size: __a : str = self.get_prev(_lowercase ) + 1 if current_left_border == index: __a : Union[str, Any] = value else: __a : str = max(_lowercase , _lowercase , _lowercase ) __a : Optional[int] = self.get_next(_lowercase ) def lowerCAmelCase__(self , _lowercase , _lowercase ): '''simple docstring''' right -= 1 # Because of right is exclusive __a : int = 0 while left <= right: __a : int = self.get_prev(_lowercase ) if left <= current_left: __a : Optional[Any] = max(_lowercase , self.tree[right] ) __a : List[str] = current_left else: __a : List[str] = max(_lowercase , self.arr[right] ) right -= 1 return result if __name__ == "__main__": import doctest doctest.testmod()
713
"""simple docstring""" import json import os import unittest from transformers import BatchEncoding, LEDTokenizer, LEDTokenizerFast from transformers.models.led.tokenization_led import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class SCREAMING_SNAKE_CASE__ ( __snake_case , unittest.TestCase ): _lowerCAmelCase = LEDTokenizer _lowerCAmelCase = LEDTokenizerFast _lowerCAmelCase = True def lowerCAmelCase__(self ): '''simple docstring''' super().setUp() __a : str = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", ] __a : int = dict(zip(_lowercase , range(len(_lowercase ) ) ) ) __a : Optional[int] = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] __a : List[Any] = {"""unk_token""": """<unk>"""} __a : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) __a : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(_lowercase ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(_lowercase ) ) def lowerCAmelCase__(self , **_lowercase ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **_lowercase ) def lowerCAmelCase__(self , **_lowercase ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **_lowercase ) def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' return "lower newer", "lower newer" @cached_property def lowerCAmelCase__(self ): '''simple docstring''' return LEDTokenizer.from_pretrained("""allenai/led-base-16384""" ) @cached_property def lowerCAmelCase__(self ): '''simple docstring''' return LEDTokenizerFast.from_pretrained("""allenai/led-base-16384""" ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' __a : Any = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] __a : List[str] = [0, 250, 251, 17818, 13, 39186, 1938, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __a : Optional[int] = tokenizer(_lowercase , max_length=len(_lowercase ) , padding=_lowercase , return_tensors="""pt""" ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual((2, 9) , batch.input_ids.shape ) self.assertEqual((2, 9) , batch.attention_mask.shape ) __a : Dict = batch.input_ids.tolist()[0] self.assertListEqual(_lowercase , _lowercase ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' __a : Union[str, Any] = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __a : Tuple = tokenizer(_lowercase , padding=_lowercase , return_tensors="""pt""" ) self.assertIn("""input_ids""" , _lowercase ) self.assertIn("""attention_mask""" , _lowercase ) self.assertNotIn("""labels""" , _lowercase ) self.assertNotIn("""decoder_attention_mask""" , _lowercase ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[Any] = [ """Summary of the text.""", """Another summary.""", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __a : Dict = tokenizer(text_target=_lowercase , max_length=32 , padding="""max_length""" , return_tensors="""pt""" ) self.assertEqual(32 , targets["""input_ids"""].shape[1] ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __a : Optional[int] = tokenizer( ["""I am a small frog""" * 1024, """I am a small frog"""] , padding=_lowercase , truncation=_lowercase , return_tensors="""pt""" ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(batch.input_ids.shape , (2, 5122) ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' __a : Tuple = ["""A long paragraph for summarization."""] __a : Dict = [ """Summary of the text.""", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __a : int = tokenizer(_lowercase , return_tensors="""pt""" ) __a : Dict = tokenizer(text_target=_lowercase , return_tensors="""pt""" ) __a : List[str] = inputs["""input_ids"""] __a : List[Any] = targets["""input_ids"""] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item() ) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item() ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __a : Optional[Any] = ["""Summary of the text.""", """Another summary."""] __a : List[Any] = [[0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, -1, -1]] __a : Union[str, Any] = tokenizer(_lowercase , padding=_lowercase ) __a : Tuple = [[0] * len(_lowercase ) for x in encoded_output["""input_ids"""]] __a : Union[str, Any] = tokenizer.pad(_lowercase ) self.assertSequenceEqual(outputs["""global_attention_mask"""] , _lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' pass def lowerCAmelCase__(self ): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): __a : Dict = self.rust_tokenizer_class.from_pretrained(_lowercase , **_lowercase ) __a : Union[str, Any] = self.tokenizer_class.from_pretrained(_lowercase , **_lowercase ) __a : Union[str, Any] = """A, <mask> AllenNLP sentence.""" __a : Dict = tokenizer_r.encode_plus(_lowercase , add_special_tokens=_lowercase , return_token_type_ids=_lowercase ) __a : Tuple = tokenizer_p.encode_plus(_lowercase , add_special_tokens=_lowercase , return_token_type_ids=_lowercase ) self.assertEqual(sum(tokens_r["""token_type_ids"""] ) , sum(tokens_p["""token_type_ids"""] ) ) self.assertEqual( sum(tokens_r["""attention_mask"""] ) / len(tokens_r["""attention_mask"""] ) , sum(tokens_p["""attention_mask"""] ) / len(tokens_p["""attention_mask"""] ) , ) __a : Tuple = tokenizer_r.convert_ids_to_tokens(tokens_r["""input_ids"""] ) __a : Any = tokenizer_p.convert_ids_to_tokens(tokens_p["""input_ids"""] ) self.assertSequenceEqual(tokens_p["""input_ids"""] , [0, 250, 6, 50264, 3823, 487, 21992, 3645, 4, 2] ) self.assertSequenceEqual(tokens_r["""input_ids"""] , [0, 250, 6, 50264, 3823, 487, 21992, 3645, 4, 2] ) self.assertSequenceEqual( _lowercase , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] ) self.assertSequenceEqual( _lowercase , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] )
63
0
"""simple docstring""" from __future__ import annotations def __magic_name__ ( _lowerCamelCase : str ): return [ord(_lowerCamelCase ) - 9_6 for elem in plain] def __magic_name__ ( _lowerCamelCase : list[int] ): return "".join(chr(elem + 9_6 ) for elem in encoded ) def __magic_name__ ( ): __a : List[str] = encode(input("""-> """ ).strip().lower() ) print("""Encoded: """ , _lowerCamelCase ) print("""Decoded:""" , decode(_lowerCamelCase ) ) if __name__ == "__main__": main()
714
"""simple docstring""" import argparse from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_controlnet_from_original_ckpt if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() parser.add_argument( "--checkpoint_path", default=None, type=str, required=True, help="Path to the checkpoint to convert." ) parser.add_argument( "--original_config_file", type=str, required=True, help="The YAML config file corresponding to the original architecture.", ) parser.add_argument( "--num_in_channels", default=None, type=int, help="The number of input channels. If `None` number of input channels will be automatically inferred.", ) parser.add_argument( "--image_size", default=512, type=int, help=( "The image size that the model was trained on. Use 512 for Stable Diffusion v1.X and Stable Siffusion v2" " Base. Use 768 for Stable Diffusion v2." ), ) parser.add_argument( "--extract_ema", action="store_true", help=( "Only relevant for checkpoints that have both EMA and non-EMA weights. Whether to extract the EMA weights" " or not. Defaults to `False`. Add `--extract_ema` to extract the EMA weights. EMA weights usually yield" " higher quality images for inference. Non-EMA weights are usually better to continue fine-tuning." ), ) parser.add_argument( "--upcast_attention", action="store_true", help=( "Whether the attention computation should always be upcasted. This is necessary when running stable" " diffusion 2.1." ), ) parser.add_argument( "--from_safetensors", action="store_true", help="If `--checkpoint_path` is in `safetensors` format, load checkpoint with safetensors instead of PyTorch.", ) parser.add_argument( "--to_safetensors", action="store_true", help="Whether to store pipeline in safetensors format or not.", ) parser.add_argument("--dump_path", default=None, type=str, required=True, help="Path to the output model.") parser.add_argument("--device", type=str, help="Device to use (e.g. cpu, cuda:0, cuda:1, etc.)") def __magic_name__ ( _lowerCamelCase : Optional[Any] ): if string == "True": return True elif string == "False": return False else: raise ValueError(F'''could not parse string as bool {string}''' ) parser.add_argument( "--use_linear_projection", help="Override for use linear projection", required=False, type=parse_bool ) parser.add_argument("--cross_attention_dim", help="Override for cross attention_dim", required=False, type=int) lowercase__ = parser.parse_args() lowercase__ = download_controlnet_from_original_ckpt( checkpoint_path=args.checkpoint_path, original_config_file=args.original_config_file, image_size=args.image_size, extract_ema=args.extract_ema, num_in_channels=args.num_in_channels, upcast_attention=args.upcast_attention, from_safetensors=args.from_safetensors, device=args.device, use_linear_projection=args.use_linear_projection, cross_attention_dim=args.cross_attention_dim, ) controlnet.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
63
0
"""simple docstring""" def __magic_name__ ( _lowerCamelCase : list ): if len(_lowerCamelCase ) <= 1: return lst __a : Tuple = 1 while i < len(_lowerCamelCase ): if lst[i - 1] <= lst[i]: i += 1 else: __a : List[Any] = lst[i], lst[i - 1] i -= 1 if i == 0: __a : Union[str, Any] = 1 return lst if __name__ == "__main__": lowercase__ = input("Enter numbers separated by a comma:\n").strip() lowercase__ = [int(item) for item in user_input.split(",")] print(gnome_sort(unsorted))
715
"""simple docstring""" import torch from diffusers import DiffusionPipeline class SCREAMING_SNAKE_CASE__ ( __snake_case ): def __init__(self , _lowercase , _lowercase ): '''simple docstring''' super().__init__() self.register_modules(unet=_lowercase , scheduler=_lowercase ) def __call__(self ): '''simple docstring''' __a : Dict = torch.randn( (1, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , ) __a : Optional[Any] = 1 __a : List[str] = self.unet(_lowercase , _lowercase ).sample __a : Union[str, Any] = self.scheduler.step(_lowercase , _lowercase , _lowercase ).prev_sample __a : Optional[int] = scheduler_output - scheduler_output + torch.ones_like(_lowercase ) return result
63
0
"""simple docstring""" from collections import deque def __magic_name__ ( _lowerCamelCase : int ): __a : Optional[Any] = len(_lowerCamelCase ) __a : int = deque() __a : List[str] = [False for _ in range(_lowerCamelCase )] __a : int = [-1 for _ in range(_lowerCamelCase )] __a : Optional[Any] = index_of[:] def strong_connect(_lowerCamelCase : List[str] , _lowerCamelCase : List[str] , _lowerCamelCase : Dict ): __a : Tuple = index # the number when this node is seen __a : List[str] = index # lowest rank node reachable from here index += 1 stack.append(_lowerCamelCase ) __a : Union[str, Any] = True for w in g[v]: if index_of[w] == -1: __a : Optional[int] = strong_connect(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) __a : Dict = ( lowlink_of[w] if lowlink_of[w] < lowlink_of[v] else lowlink_of[v] ) elif on_stack[w]: __a : Optional[int] = ( lowlink_of[w] if lowlink_of[w] < lowlink_of[v] else lowlink_of[v] ) if lowlink_of[v] == index_of[v]: __a : Union[str, Any] = [] __a : Any = stack.pop() __a : Optional[int] = False component.append(_lowerCamelCase ) while w != v: __a : str = stack.pop() __a : Dict = False component.append(_lowerCamelCase ) components.append(_lowerCamelCase ) return index __a : Union[str, Any] = [] for v in range(_lowerCamelCase ): if index_of[v] == -1: strong_connect(_lowerCamelCase , 0 , _lowerCamelCase ) return components def __magic_name__ ( _lowerCamelCase : Any , _lowerCamelCase : List[Any] ): __a : int = [[] for _ in range(_lowerCamelCase )] for u, v in edges: g[u].append(_lowerCamelCase ) return g if __name__ == "__main__": # Test lowercase__ = 7 lowercase__ = [0, 0, 1, 2, 3, 3, 4, 4, 6] lowercase__ = [1, 3, 2, 0, 1, 4, 5, 6, 5] lowercase__ = [(u, v) for u, v in zip(source, target)] lowercase__ = create_graph(n_vertices, edges) assert [[5], [6], [4], [3, 2, 1, 0]] == tarjan(g)
716
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { "sayakpaul/vit-msn-base": "https://huggingface.co/sayakpaul/vit-msn-base/resolve/main/config.json", # See all ViT MSN models at https://huggingface.co/models?filter=vit_msn } class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = "vit_msn" def __init__(self , _lowercase=768 , _lowercase=12 , _lowercase=12 , _lowercase=3072 , _lowercase="gelu" , _lowercase=0.0 , _lowercase=0.0 , _lowercase=0.02 , _lowercase=1e-06 , _lowercase=224 , _lowercase=16 , _lowercase=3 , _lowercase=True , **_lowercase , ): '''simple docstring''' super().__init__(**_lowercase ) __a : int = hidden_size __a : str = num_hidden_layers __a : str = num_attention_heads __a : Optional[Any] = intermediate_size __a : Union[str, Any] = hidden_act __a : Tuple = hidden_dropout_prob __a : Any = attention_probs_dropout_prob __a : List[Any] = initializer_range __a : Any = layer_norm_eps __a : Dict = image_size __a : List[Any] = patch_size __a : Dict = num_channels __a : Optional[Any] = qkv_bias
63
0
"""simple docstring""" import argparse import tensorflow as tf import torch from transformers import BertConfig, BertForMaskedLM from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertPooler, BertSelfAttention, BertSelfOutput, ) from transformers.utils import logging logging.set_verbosity_info() def __magic_name__ ( _lowerCamelCase : str , _lowerCamelCase : str , _lowerCamelCase : str ): def get_masked_lm_array(_lowerCamelCase : str ): __a : List[str] = F'''masked_lm/{name}/.ATTRIBUTES/VARIABLE_VALUE''' __a : List[Any] = tf.train.load_variable(_lowerCamelCase , _lowerCamelCase ) if "kernel" in name: __a : Dict = array.transpose() return torch.from_numpy(_lowerCamelCase ) def get_encoder_array(_lowerCamelCase : str ): __a : Dict = F'''encoder/{name}/.ATTRIBUTES/VARIABLE_VALUE''' __a : str = tf.train.load_variable(_lowerCamelCase , _lowerCamelCase ) if "kernel" in name: __a : Any = array.transpose() return torch.from_numpy(_lowerCamelCase ) def get_encoder_layer_array(_lowerCamelCase : int , _lowerCamelCase : str ): __a : int = F'''encoder/_transformer_layers/{layer_index}/{name}/.ATTRIBUTES/VARIABLE_VALUE''' __a : Dict = tf.train.load_variable(_lowerCamelCase , _lowerCamelCase ) if "kernel" in name: __a : Tuple = array.transpose() return torch.from_numpy(_lowerCamelCase ) def get_encoder_attention_layer_array(_lowerCamelCase : int , _lowerCamelCase : str , _lowerCamelCase : Optional[int] ): __a : Dict = F'''encoder/_transformer_layers/{layer_index}/_attention_layer/{name}/.ATTRIBUTES/VARIABLE_VALUE''' __a : int = tf.train.load_variable(_lowerCamelCase , _lowerCamelCase ) __a : Dict = array.reshape(_lowerCamelCase ) if "kernel" in name: __a : List[Any] = array.transpose() return torch.from_numpy(_lowerCamelCase ) print(F'''Loading model based on config from {config_path}...''' ) __a : Tuple = BertConfig.from_json_file(_lowerCamelCase ) __a : Union[str, Any] = BertForMaskedLM(_lowerCamelCase ) # Layers for layer_index in range(0 , config.num_hidden_layers ): __a : BertLayer = model.bert.encoder.layer[layer_index] # Self-attention __a : BertSelfAttention = layer.attention.self __a : List[str] = get_encoder_attention_layer_array( _lowerCamelCase , """_query_dense/kernel""" , self_attn.query.weight.data.shape ) __a : Dict = get_encoder_attention_layer_array( _lowerCamelCase , """_query_dense/bias""" , self_attn.query.bias.data.shape ) __a : List[Any] = get_encoder_attention_layer_array( _lowerCamelCase , """_key_dense/kernel""" , self_attn.key.weight.data.shape ) __a : Optional[int] = get_encoder_attention_layer_array( _lowerCamelCase , """_key_dense/bias""" , self_attn.key.bias.data.shape ) __a : Optional[int] = get_encoder_attention_layer_array( _lowerCamelCase , """_value_dense/kernel""" , self_attn.value.weight.data.shape ) __a : Optional[int] = get_encoder_attention_layer_array( _lowerCamelCase , """_value_dense/bias""" , self_attn.value.bias.data.shape ) # Self-attention Output __a : BertSelfOutput = layer.attention.output __a : List[str] = get_encoder_attention_layer_array( _lowerCamelCase , """_output_dense/kernel""" , self_output.dense.weight.data.shape ) __a : List[Any] = get_encoder_attention_layer_array( _lowerCamelCase , """_output_dense/bias""" , self_output.dense.bias.data.shape ) __a : Optional[Any] = get_encoder_layer_array(_lowerCamelCase , """_attention_layer_norm/gamma""" ) __a : Tuple = get_encoder_layer_array(_lowerCamelCase , """_attention_layer_norm/beta""" ) # Intermediate __a : BertIntermediate = layer.intermediate __a : Union[str, Any] = get_encoder_layer_array(_lowerCamelCase , """_intermediate_dense/kernel""" ) __a : str = get_encoder_layer_array(_lowerCamelCase , """_intermediate_dense/bias""" ) # Output __a : BertOutput = layer.output __a : int = get_encoder_layer_array(_lowerCamelCase , """_output_dense/kernel""" ) __a : List[str] = get_encoder_layer_array(_lowerCamelCase , """_output_dense/bias""" ) __a : Tuple = get_encoder_layer_array(_lowerCamelCase , """_output_layer_norm/gamma""" ) __a : Optional[Any] = get_encoder_layer_array(_lowerCamelCase , """_output_layer_norm/beta""" ) # Embeddings __a : Tuple = get_encoder_array("""_position_embedding_layer/embeddings""" ) __a : Dict = get_encoder_array("""_type_embedding_layer/embeddings""" ) __a : List[str] = get_encoder_array("""_embedding_norm_layer/gamma""" ) __a : Dict = get_encoder_array("""_embedding_norm_layer/beta""" ) # LM Head __a : Union[str, Any] = model.cls.predictions.transform __a : str = get_masked_lm_array("""dense/kernel""" ) __a : str = get_masked_lm_array("""dense/bias""" ) __a : Union[str, Any] = get_masked_lm_array("""layer_norm/gamma""" ) __a : List[Any] = get_masked_lm_array("""layer_norm/beta""" ) __a : Union[str, Any] = get_masked_lm_array("""embedding_table""" ) # Pooling __a : Optional[Any] = BertPooler(config=_lowerCamelCase ) __a : BertPooler = get_encoder_array("""_pooler_layer/kernel""" ) __a : BertPooler = get_encoder_array("""_pooler_layer/bias""" ) # Export final model model.save_pretrained(_lowerCamelCase ) # Integration test - should load without any errors ;) __a : Tuple = BertForMaskedLM.from_pretrained(_lowerCamelCase ) print(new_model.eval() ) print("""Model conversion was done sucessfully!""" ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() parser.add_argument( "--tf_checkpoint_path", type=str, required=True, help="Path to the TensorFlow Token Dropping checkpoint path." ) parser.add_argument( "--bert_config_file", type=str, required=True, help="The config json file corresponding to the BERT model. This specifies the model architecture.", ) parser.add_argument( "--pytorch_dump_path", type=str, required=True, help="Path to the output PyTorch model.", ) lowercase__ = parser.parse_args() convert_checkpoint_to_pytorch(args.tf_checkpoint_path, args.bert_config_file, args.pytorch_dump_path)
717
"""simple docstring""" import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert_fast import BertTokenizerFast from .tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer, DPRReaderTokenizer lowercase__ = logging.get_logger(__name__) lowercase__ = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} lowercase__ = { "vocab_file": { "facebook/dpr-ctx_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-ctx_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-ctx_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-ctx_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/tokenizer.json" ), }, } lowercase__ = { "vocab_file": { "facebook/dpr-question_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-question_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-question_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-question_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/tokenizer.json" ), }, } lowercase__ = { "vocab_file": { "facebook/dpr-reader-single-nq-base": ( "https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-reader-multiset-base": ( "https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-reader-single-nq-base": ( "https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-reader-multiset-base": ( "https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/tokenizer.json" ), }, } lowercase__ = { "facebook/dpr-ctx_encoder-single-nq-base": 512, "facebook/dpr-ctx_encoder-multiset-base": 512, } lowercase__ = { "facebook/dpr-question_encoder-single-nq-base": 512, "facebook/dpr-question_encoder-multiset-base": 512, } lowercase__ = { "facebook/dpr-reader-single-nq-base": 512, "facebook/dpr-reader-multiset-base": 512, } lowercase__ = { "facebook/dpr-ctx_encoder-single-nq-base": {"do_lower_case": True}, "facebook/dpr-ctx_encoder-multiset-base": {"do_lower_case": True}, } lowercase__ = { "facebook/dpr-question_encoder-single-nq-base": {"do_lower_case": True}, "facebook/dpr-question_encoder-multiset-base": {"do_lower_case": True}, } lowercase__ = { "facebook/dpr-reader-single-nq-base": {"do_lower_case": True}, "facebook/dpr-reader-multiset-base": {"do_lower_case": True}, } class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = VOCAB_FILES_NAMES _lowerCAmelCase = CONTEXT_ENCODER_PRETRAINED_VOCAB_FILES_MAP _lowerCAmelCase = CONTEXT_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCAmelCase = CONTEXT_ENCODER_PRETRAINED_INIT_CONFIGURATION _lowerCAmelCase = DPRContextEncoderTokenizer class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = VOCAB_FILES_NAMES _lowerCAmelCase = QUESTION_ENCODER_PRETRAINED_VOCAB_FILES_MAP _lowerCAmelCase = QUESTION_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCAmelCase = QUESTION_ENCODER_PRETRAINED_INIT_CONFIGURATION _lowerCAmelCase = DPRQuestionEncoderTokenizer lowercase__ = collections.namedtuple( "DPRSpanPrediction", ["span_score", "relevance_score", "doc_id", "start_index", "end_index", "text"] ) lowercase__ = collections.namedtuple("DPRReaderOutput", ["start_logits", "end_logits", "relevance_logits"]) lowercase__ = R"\n Return a dictionary with the token ids of the input strings and other information to give to `.decode_best_spans`.\n It converts the strings of a question and different passages (title and text) in a sequence of IDs (integers),\n using the tokenizer and vocabulary. The resulting `input_ids` is a matrix of size `(n_passages, sequence_length)`\n with the format:\n\n [CLS] <question token ids> [SEP] <titles ids> [SEP] <texts ids>\n\n Args:\n questions (`str` or `List[str]`):\n The questions to be encoded. You can specify one question for many passages. In this case, the question\n will be duplicated like `[questions] * n_passages`. Otherwise you have to specify as many questions as in\n `titles` or `texts`.\n titles (`str` or `List[str]`):\n The passages titles to be encoded. This can be a string or a list of strings if there are several passages.\n texts (`str` or `List[str]`):\n The passages texts to be encoded. This can be a string or a list of strings if there are several passages.\n padding (`bool`, `str` or [`~utils.PaddingStrategy`], *optional*, defaults to `False`):\n Activates and controls padding. Accepts the following values:\n\n - `True` or `'longest'`: Pad to the longest sequence in the batch (or no padding if only a single sequence\n if provided).\n - `'max_length'`: Pad to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided.\n - `False` or `'do_not_pad'` (default): No padding (i.e., can output a batch with sequences of different\n lengths).\n truncation (`bool`, `str` or [`~tokenization_utils_base.TruncationStrategy`], *optional*, defaults to `False`):\n Activates and controls truncation. Accepts the following values:\n\n - `True` or `'longest_first'`: Truncate to a maximum length specified with the argument `max_length` or to\n the maximum acceptable input length for the model if that argument is not provided. This will truncate\n token by token, removing a token from the longest sequence in the pair if a pair of sequences (or a batch\n of pairs) is provided.\n - `'only_first'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the first\n sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `'only_second'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the\n second sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `False` or `'do_not_truncate'` (default): No truncation (i.e., can output batch with sequence lengths\n greater than the model maximum admissible input size).\n max_length (`int`, *optional*):\n Controls the maximum length to use by one of the truncation/padding parameters.\n\n If left unset or set to `None`, this will use the predefined model maximum length if a maximum length\n is required by one of the truncation/padding parameters. If the model has no specific maximum input\n length (like XLNet) truncation/padding to a maximum length will be deactivated.\n return_tensors (`str` or [`~utils.TensorType`], *optional*):\n If set, will return tensors instead of list of python integers. Acceptable values are:\n\n - `'tf'`: Return TensorFlow `tf.constant` objects.\n - `'pt'`: Return PyTorch `torch.Tensor` objects.\n - `'np'`: Return Numpy `np.ndarray` objects.\n return_attention_mask (`bool`, *optional*):\n Whether or not to return the attention mask. If not set, will return the attention mask according to the\n specific tokenizer's default, defined by the `return_outputs` attribute.\n\n [What are attention masks?](../glossary#attention-mask)\n\n Return:\n `Dict[str, List[List[int]]]`: A dictionary with the following keys:\n\n - `input_ids`: List of token ids to be fed to a model.\n - `attention_mask`: List of indices specifying which tokens should be attended to by the model.\n " @add_start_docstrings(__snake_case ) class SCREAMING_SNAKE_CASE__ : def __call__(self , _lowercase , _lowercase = None , _lowercase = None , _lowercase = False , _lowercase = False , _lowercase = None , _lowercase = None , _lowercase = None , **_lowercase , ): '''simple docstring''' if titles is None and texts is None: return super().__call__( _lowercase , padding=_lowercase , truncation=_lowercase , max_length=_lowercase , return_tensors=_lowercase , return_attention_mask=_lowercase , **_lowercase , ) elif titles is None or texts is None: __a : str = titles if texts is None else texts return super().__call__( _lowercase , _lowercase , padding=_lowercase , truncation=_lowercase , max_length=_lowercase , return_tensors=_lowercase , return_attention_mask=_lowercase , **_lowercase , ) __a : str = titles if not isinstance(_lowercase , _lowercase ) else [titles] __a : Optional[Any] = texts if not isinstance(_lowercase , _lowercase ) else [texts] __a : Tuple = len(_lowercase ) __a : Dict = questions if not isinstance(_lowercase , _lowercase ) else [questions] * n_passages assert len(_lowercase ) == len( _lowercase ), F'''There should be as many titles than texts but got {len(_lowercase )} titles and {len(_lowercase )} texts.''' __a : Optional[Any] = super().__call__(_lowercase , _lowercase , padding=_lowercase , truncation=_lowercase )["""input_ids"""] __a : str = super().__call__(_lowercase , add_special_tokens=_lowercase , padding=_lowercase , truncation=_lowercase )["""input_ids"""] __a : Union[str, Any] = { """input_ids""": [ (encoded_question_and_title + encoded_text)[:max_length] if max_length is not None and truncation else encoded_question_and_title + encoded_text for encoded_question_and_title, encoded_text in zip(_lowercase , _lowercase ) ] } if return_attention_mask is not False: __a : Optional[int] = [] for input_ids in encoded_inputs["input_ids"]: attention_mask.append([int(input_id != self.pad_token_id ) for input_id in input_ids] ) __a : str = attention_mask return self.pad(_lowercase , padding=_lowercase , max_length=_lowercase , return_tensors=_lowercase ) def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase = 16 , _lowercase = 64 , _lowercase = 4 , ): '''simple docstring''' __a : Union[str, Any] = reader_input["""input_ids"""] __a , __a , __a : Optional[int] = reader_output[:3] __a : int = len(_lowercase ) __a : Any = sorted(range(_lowercase ) , reverse=_lowercase , key=relevance_logits.__getitem__ ) __a : List[DPRReaderOutput] = [] for doc_id in sorted_docs: __a : Optional[int] = list(input_ids[doc_id] ) # assuming question & title information is at the beginning of the sequence __a : Dict = sequence_ids.index(self.sep_token_id , 2 ) + 1 # second sep id if sequence_ids[-1] == self.pad_token_id: __a : int = sequence_ids.index(self.pad_token_id ) else: __a : Optional[Any] = len(_lowercase ) __a : List[Any] = self._get_best_spans( start_logits=start_logits[doc_id][passage_offset:sequence_len] , end_logits=end_logits[doc_id][passage_offset:sequence_len] , max_answer_length=_lowercase , top_spans=_lowercase , ) for start_index, end_index in best_spans: start_index += passage_offset end_index += passage_offset nbest_spans_predictions.append( DPRSpanPrediction( span_score=start_logits[doc_id][start_index] + end_logits[doc_id][end_index] , relevance_score=relevance_logits[doc_id] , doc_id=_lowercase , start_index=_lowercase , end_index=_lowercase , text=self.decode(sequence_ids[start_index : end_index + 1] ) , ) ) if len(_lowercase ) >= num_spans: break return nbest_spans_predictions[:num_spans] def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase , _lowercase , ): '''simple docstring''' __a : Tuple = [] for start_index, start_score in enumerate(_lowercase ): for answer_length, end_score in enumerate(end_logits[start_index : start_index + max_answer_length] ): scores.append(((start_index, start_index + answer_length), start_score + end_score) ) __a : str = sorted(_lowercase , key=lambda _lowercase : x[1] , reverse=_lowercase ) __a : Union[str, Any] = [] for (start_index, end_index), score in scores: assert start_index <= end_index, F'''Wrong span indices: [{start_index}:{end_index}]''' __a : List[str] = end_index - start_index + 1 assert length <= max_answer_length, F'''Span is too long: {length} > {max_answer_length}''' if any( start_index <= prev_start_index <= prev_end_index <= end_index or prev_start_index <= start_index <= end_index <= prev_end_index for (prev_start_index, prev_end_index) in chosen_span_intervals ): continue chosen_span_intervals.append((start_index, end_index) ) if len(_lowercase ) == top_spans: break return chosen_span_intervals @add_end_docstrings(__snake_case ) class SCREAMING_SNAKE_CASE__ ( __snake_case , __snake_case ): _lowerCAmelCase = VOCAB_FILES_NAMES _lowerCAmelCase = READER_PRETRAINED_VOCAB_FILES_MAP _lowerCAmelCase = READER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCAmelCase = READER_PRETRAINED_INIT_CONFIGURATION _lowerCAmelCase = ["input_ids", "attention_mask"] _lowerCAmelCase = DPRReaderTokenizer
63
0
"""simple docstring""" from __future__ import annotations from math import pi # Define the Reduced Planck Constant ℏ (H bar), speed of light C, value of # Pi and the function lowercase__ = 1.054_571_817e-34 # unit of ℏ : J * s lowercase__ = 3e8 # unit of c : m * s^-1 def __magic_name__ ( _lowerCamelCase : float , _lowerCamelCase : float , _lowerCamelCase : float ): if (force, area, distance).count(0 ) != 1: raise ValueError("""One and only one argument must be 0""" ) if force < 0: raise ValueError("""Magnitude of force can not be negative""" ) if distance < 0: raise ValueError("""Distance can not be negative""" ) if area < 0: raise ValueError("""Area can not be negative""" ) if force == 0: __a : int = (REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 * area) / ( 2_4_0 * (distance) ** 4 ) return {"force": force} elif area == 0: __a : int = (2_4_0 * force * (distance) ** 4) / ( REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 ) return {"area": area} elif distance == 0: __a : Dict = ( (REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 * area) / (2_4_0 * force) ) ** (1 / 4) return {"distance": distance} raise ValueError("""One and only one argument must be 0""" ) # Run doctest if __name__ == "__main__": import doctest doctest.testmod()
718
"""simple docstring""" import os def __magic_name__ ( _lowerCamelCase : Dict ): __a : List[str] = len(grid[0] ) __a : int = len(_lowerCamelCase ) __a : Tuple = 0 __a : List[Any] = 0 __a : List[str] = 0 # Check vertically, horizontally, diagonally at the same time (only works # for nxn grid) for i in range(_lowerCamelCase ): for j in range(n_rows - 3 ): __a : List[Any] = grid[j][i] * grid[j + 1][i] * grid[j + 2][i] * grid[j + 3][i] __a : Tuple = grid[i][j] * grid[i][j + 1] * grid[i][j + 2] * grid[i][j + 3] # Left-to-right diagonal (\) product if i < n_columns - 3: __a : List[Any] = ( grid[i][j] * grid[i + 1][j + 1] * grid[i + 2][j + 2] * grid[i + 3][j + 3] ) # Right-to-left diagonal(/) product if i > 2: __a : List[Any] = ( grid[i][j] * grid[i - 1][j + 1] * grid[i - 2][j + 2] * grid[i - 3][j + 3] ) __a : str = max( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) if max_product > largest: __a : Optional[Any] = max_product return largest def __magic_name__ ( ): __a : Tuple = [] with open(os.path.dirname(_lowerCamelCase ) + """/grid.txt""" ) as file: for line in file: grid.append(line.strip("""\n""" ).split(""" """ ) ) __a : Tuple = [[int(_lowerCamelCase ) for i in grid[j]] for j in range(len(_lowerCamelCase ) )] return largest_product(_lowerCamelCase ) if __name__ == "__main__": print(solution())
63
0
"""simple docstring""" import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.activations import gelu_new, gelu_python, get_activation @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def lowerCAmelCase__(self ): '''simple docstring''' __a : List[str] = torch.tensor([-100, -1, -0.1, 0, 0.1, 1.0, 100] ) __a : Optional[Any] = get_activation("""gelu""" ) self.assertTrue(torch.allclose(gelu_python(_lowercase ) , torch_builtin(_lowercase ) ) ) self.assertFalse(torch.allclose(gelu_python(_lowercase ) , gelu_new(_lowercase ) ) ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Tuple = torch.tensor([-100, -1, -0.1, 0, 0.1, 1.0, 100] ) __a : Optional[Any] = get_activation("""gelu""" ) __a : Optional[int] = get_activation("""gelu_10""" ) __a : str = torch_builtin(_lowercase ) __a : int = geluaa(_lowercase ) __a : Optional[int] = torch.where(y_gelu_aa < 10.0 , 1 , 0 ) self.assertTrue(torch.max(_lowercase ).item() == 10.0 ) self.assertTrue(torch.allclose(y_gelu * clipped_mask , y_gelu_aa * clipped_mask ) ) def lowerCAmelCase__(self ): '''simple docstring''' get_activation("""gelu""" ) get_activation("""gelu_10""" ) get_activation("""gelu_fast""" ) get_activation("""gelu_new""" ) get_activation("""gelu_python""" ) get_activation("""gelu_pytorch_tanh""" ) get_activation("""linear""" ) get_activation("""mish""" ) get_activation("""quick_gelu""" ) get_activation("""relu""" ) get_activation("""sigmoid""" ) get_activation("""silu""" ) get_activation("""swish""" ) get_activation("""tanh""" ) with self.assertRaises(_lowercase ): get_activation("""bogus""" ) with self.assertRaises(_lowercase ): get_activation(_lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Dict = get_activation("""gelu""" ) __a : List[Any] = 1 __a : int = get_activation("""gelu""" ) self.assertEqual(acta.a , 1 ) with self.assertRaises(_lowercase ): __a : List[str] = acta.a
719
"""simple docstring""" from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import BaseOutput, is_torch_available, is_transformers_available @dataclass class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = 42 _lowerCAmelCase = 42 if is_transformers_available() and is_torch_available(): from .pipeline_semantic_stable_diffusion import SemanticStableDiffusionPipeline
63
0
"""simple docstring""" from __future__ import annotations lowercase__ = list[list[int]] # assigning initial values to the grid lowercase__ = [ [3, 0, 6, 5, 0, 8, 4, 0, 0], [5, 2, 0, 0, 0, 0, 0, 0, 0], [0, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] # a grid with no solution lowercase__ = [ [5, 0, 6, 5, 0, 8, 4, 0, 3], [5, 2, 0, 0, 0, 0, 0, 0, 2], [1, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] def __magic_name__ ( _lowerCamelCase : Matrix , _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : int ): for i in range(9 ): if grid[row][i] == n or grid[i][column] == n: return False for i in range(3 ): for j in range(3 ): if grid[(row - row % 3) + i][(column - column % 3) + j] == n: return False return True def __magic_name__ ( _lowerCamelCase : Matrix ): for i in range(9 ): for j in range(9 ): if grid[i][j] == 0: return i, j return None def __magic_name__ ( _lowerCamelCase : Matrix ): if location := find_empty_location(_lowerCamelCase ): __a : List[Any] = location else: # If the location is ``None``, then the grid is solved. return grid for digit in range(1 , 1_0 ): if is_safe(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): __a : int = digit if sudoku(_lowerCamelCase ) is not None: return grid __a : Optional[int] = 0 return None def __magic_name__ ( _lowerCamelCase : Matrix ): for row in grid: for cell in row: print(_lowerCamelCase , end=""" """ ) print() if __name__ == "__main__": # make a copy of grid so that you can compare with the unmodified grid for example_grid in (initial_grid, no_solution): print("\nExample grid:\n" + "=" * 20) print_solution(example_grid) print("\nExample grid solution:") lowercase__ = sudoku(example_grid) if solution is not None: print_solution(solution) else: print("Cannot find a solution.")
720
"""simple docstring""" import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TextClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. lowercase__ = {"LayoutLMv2Config", "LayoutLMv3Config"} @is_pipeline_test class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): _lowerCAmelCase = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING _lowerCAmelCase = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: _lowerCAmelCase = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: _lowerCAmelCase = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } @require_torch def lowerCAmelCase__(self ): '''simple docstring''' __a : int = pipeline( task="""text-classification""" , model="""hf-internal-testing/tiny-random-distilbert""" , framework="""pt""" ) __a : Tuple = text_classifier("""This is great !""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """LABEL_0""", """score""": 0.504}] ) __a : Optional[Any] = text_classifier("""This is great !""" , top_k=2 ) self.assertEqual( nested_simplify(_lowercase ) , [{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}] ) __a : int = text_classifier(["""This is great !""", """This is bad"""] , top_k=2 ) self.assertEqual( nested_simplify(_lowercase ) , [ [{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}], [{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}], ] , ) __a : List[str] = text_classifier("""This is great !""" , top_k=1 ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """LABEL_0""", """score""": 0.504}] ) # Legacy behavior __a : Optional[int] = text_classifier("""This is great !""" , return_all_scores=_lowercase ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """LABEL_0""", """score""": 0.504}] ) __a : Tuple = text_classifier("""This is great !""" , return_all_scores=_lowercase ) self.assertEqual( nested_simplify(_lowercase ) , [[{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}]] ) __a : Any = text_classifier(["""This is great !""", """Something else"""] , return_all_scores=_lowercase ) self.assertEqual( nested_simplify(_lowercase ) , [ [{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}], [{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}], ] , ) __a : Union[str, Any] = text_classifier(["""This is great !""", """Something else"""] , return_all_scores=_lowercase ) self.assertEqual( nested_simplify(_lowercase ) , [ {"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_0""", """score""": 0.504}, ] , ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' import torch __a : Any = pipeline( task="""text-classification""" , model="""hf-internal-testing/tiny-random-distilbert""" , framework="""pt""" , device=torch.device("""cpu""" ) , ) __a : Optional[int] = text_classifier("""This is great !""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """LABEL_0""", """score""": 0.504}] ) @require_tf def lowerCAmelCase__(self ): '''simple docstring''' __a : List[Any] = pipeline( task="""text-classification""" , model="""hf-internal-testing/tiny-random-distilbert""" , framework="""tf""" ) __a : List[str] = text_classifier("""This is great !""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """LABEL_0""", """score""": 0.504}] ) @slow @require_torch def lowerCAmelCase__(self ): '''simple docstring''' __a : Tuple = pipeline("""text-classification""" ) __a : Tuple = text_classifier("""This is great !""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """POSITIVE""", """score""": 1.0}] ) __a : Optional[int] = text_classifier("""This is bad !""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """NEGATIVE""", """score""": 1.0}] ) __a : Union[str, Any] = text_classifier("""Birds are a type of animal""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """POSITIVE""", """score""": 0.988}] ) @slow @require_tf def lowerCAmelCase__(self ): '''simple docstring''' __a : List[str] = pipeline("""text-classification""" , framework="""tf""" ) __a : str = text_classifier("""This is great !""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """POSITIVE""", """score""": 1.0}] ) __a : Tuple = text_classifier("""This is bad !""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """NEGATIVE""", """score""": 1.0}] ) __a : str = text_classifier("""Birds are a type of animal""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """POSITIVE""", """score""": 0.988}] ) def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase ): '''simple docstring''' __a : Dict = TextClassificationPipeline(model=_lowercase , tokenizer=_lowercase ) return text_classifier, ["HuggingFace is in", "This is another test"] def lowerCAmelCase__(self , _lowercase , _lowercase ): '''simple docstring''' __a : List[str] = text_classifier.model # Small inputs because BartTokenizer tiny has maximum position embeddings = 22 __a : Union[str, Any] = """HuggingFace is in""" __a : List[str] = text_classifier(_lowercase ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": ANY(_lowercase ), """score""": ANY(_lowercase )}] ) self.assertTrue(outputs[0]["""label"""] in model.config.idalabel.values() ) __a : Optional[int] = ["""HuggingFace is in """, """Paris is in France"""] __a : Dict = text_classifier(_lowercase ) self.assertEqual( nested_simplify(_lowercase ) , [{"""label""": ANY(_lowercase ), """score""": ANY(_lowercase )}, {"""label""": ANY(_lowercase ), """score""": ANY(_lowercase )}] , ) self.assertTrue(outputs[0]["""label"""] in model.config.idalabel.values() ) self.assertTrue(outputs[1]["""label"""] in model.config.idalabel.values() ) # Forcing to get all results with `top_k=None` # This is NOT the legacy format __a : Dict = text_classifier(_lowercase , top_k=_lowercase ) __a : Dict = len(model.config.idalabel.values() ) self.assertEqual( nested_simplify(_lowercase ) , [[{"""label""": ANY(_lowercase ), """score""": ANY(_lowercase )}] * N, [{"""label""": ANY(_lowercase ), """score""": ANY(_lowercase )}] * N] , ) __a : Dict = {"""text""": """HuggingFace is in """, """text_pair""": """Paris is in France"""} __a : Any = text_classifier(_lowercase ) self.assertEqual( nested_simplify(_lowercase ) , {"""label""": ANY(_lowercase ), """score""": ANY(_lowercase )} , ) self.assertTrue(outputs["""label"""] in model.config.idalabel.values() ) # This might be used a text pair, but tokenizer + pipe interaction # makes it hard to understand that it's not using the pair properly # https://github.com/huggingface/transformers/issues/17305 # We disabled this usage instead as it was outputting wrong outputs. __a : Dict = [["""HuggingFace is in """, """Paris is in France"""]] with self.assertRaises(_lowercase ): text_classifier(_lowercase ) # This used to be valid for doing text pairs # We're keeping it working because of backward compatibility __a : Optional[int] = text_classifier([[["""HuggingFace is in """, """Paris is in France"""]]] ) self.assertEqual( nested_simplify(_lowercase ) , [{"""label""": ANY(_lowercase ), """score""": ANY(_lowercase )}] , ) self.assertTrue(outputs[0]["""label"""] in model.config.idalabel.values() )
63
0
"""simple docstring""" from manim import * class SCREAMING_SNAKE_CASE__ ( __snake_case ): def lowerCAmelCase__(self ): '''simple docstring''' __a : List[str] = Rectangle(height=0.5 , width=0.5 ) __a : Union[str, Any] = Rectangle(height=0.25 , width=0.25 ) __a : Dict = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) __a : Dict = [mem.copy() for i in range(6 )] __a : str = [mem.copy() for i in range(6 )] __a : Tuple = VGroup(*_lowercase ).arrange(_lowercase , buff=0 ) __a : List[Any] = VGroup(*_lowercase ).arrange(_lowercase , buff=0 ) __a : List[Any] = VGroup(_lowercase , _lowercase ).arrange(_lowercase , buff=0 ) __a : Union[str, Any] = Text("""CPU""" , font_size=24 ) __a : Dict = Group(_lowercase , _lowercase ).arrange(_lowercase , buff=0.5 , aligned_edge=_lowercase ) cpu.move_to([-2.5, -0.5, 0] ) self.add(_lowercase ) __a : Optional[Any] = [mem.copy() for i in range(4 )] __a : Dict = VGroup(*_lowercase ).arrange(_lowercase , buff=0 ) __a : List[str] = Text("""GPU""" , font_size=24 ) __a : Any = Group(_lowercase , _lowercase ).arrange(_lowercase , buff=0.5 , aligned_edge=_lowercase ) gpu.move_to([-1, -1, 0] ) self.add(_lowercase ) __a : List[Any] = [mem.copy() for i in range(6 )] __a : Any = VGroup(*_lowercase ).arrange(_lowercase , buff=0 ) __a : Optional[Any] = Text("""Model""" , font_size=24 ) __a : Any = Group(_lowercase , _lowercase ).arrange(_lowercase , buff=0.5 , aligned_edge=_lowercase ) model.move_to([3, -1.0, 0] ) self.add(_lowercase ) __a : Tuple = [] __a : Tuple = [] __a : Optional[int] = [] for i, rect in enumerate(_lowercase ): rect.set_stroke(_lowercase ) __a : str = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(_lowercase , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=_lowercase ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(model_cpu_arr[0] , direction=_lowercase , buff=0.0 ) else: cpu_target.next_to(model_cpu_arr[i - 1] , direction=_lowercase , buff=0.0 ) self.add(_lowercase ) model_cpu_arr.append(_lowercase ) self.add(*_lowercase , *_lowercase , *_lowercase ) __a : Optional[Any] = [mem.copy() for i in range(6 )] __a : Union[str, Any] = VGroup(*_lowercase ).arrange(_lowercase , buff=0 ) __a : Any = Text("""Loaded Checkpoint""" , font_size=24 ) __a : str = Group(_lowercase , _lowercase ).arrange(_lowercase , buff=0.5 , aligned_edge=_lowercase ) checkpoint.move_to([3, 0.5, 0] ) self.add(_lowercase ) __a : Dict = [] __a : int = [] for i, rect in enumerate(_lowercase ): __a : List[str] = fill.copy().set_fill(_lowercase , opacity=0.7 ) target.move_to(_lowercase ) ckpt_arr.append(_lowercase ) __a : Union[str, Any] = target.copy() if i < 5: cpu_target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.move_to(cpu_right_col_base[i - 5] ) ckpt_cpu_arr.append(_lowercase ) self.add(*_lowercase , *_lowercase ) __a : List[str] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) __a : List[Any] = MarkupText( F'''<b>Key:</b>\n\n<span fgcolor=\'{YELLOW}\'>●</span> Empty Model''' , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(_lowercase , _lowercase ) __a : str = MarkupText( F'''<span fgcolor=\'{BLUE}\'>●</span> Checkpoint''' , font_size=18 , ) blue_text.next_to(_lowercase , DOWN * 2.4 , aligned_edge=key_text.get_left() ) self.add(_lowercase ) __a : Optional[int] = MarkupText( F'''Based on the passed in configuration, weights are stored in\na variety of np.memmaps on disk or to a particular device.''' , font_size=24 , ) step_a.move_to([2, 2, 0] ) __a : List[Any] = [meta_mem.copy() for i in range(6 )] __a : Optional[int] = [meta_mem.copy() for i in range(6 )] __a : List[Any] = VGroup(*_lowercase ).arrange(_lowercase , buff=0 ) __a : List[str] = VGroup(*_lowercase ).arrange(_lowercase , buff=0 ) __a : Tuple = VGroup(_lowercase , _lowercase ).arrange(_lowercase , buff=0 ) __a : Dict = Text("""Disk""" , font_size=24 ) __a : Dict = Group(_lowercase , _lowercase ).arrange(_lowercase , buff=0.5 , aligned_edge=_lowercase ) disk.move_to([-4.0, -1.25, 0] ) self.play(Write(_lowercase , run_time=3 ) , Write(_lowercase , run_time=1 ) , Create(_lowercase , run_time=1 ) ) __a : Optional[Any] = [] for i, rect in enumerate(_lowercase ): __a : List[str] = rect.copy() target.generate_target() target.target.move_to(disk_left_col_base[i] ).scale(0.5 ) animations.append(MoveToTarget(_lowercase , run_time=1.5 ) ) self.play(*_lowercase ) self.play(FadeOut(_lowercase ) ) __a : List[str] = MarkupText(F'''Then, the checkpoint is removed from memory\nthrough garbage collection.''' , font_size=24 ) step_a.move_to([2, 2, 0] ) self.play(Write(_lowercase , run_time=3 ) ) self.play( FadeOut(_lowercase , _lowercase , *_lowercase , *_lowercase ) , ) self.wait()
721
"""simple docstring""" import unittest from knapsack import knapsack as k class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def lowerCAmelCase__(self ): '''simple docstring''' __a : str = 0 __a : Optional[Any] = [0] __a : int = [0] __a : str = len(_lowercase ) self.assertEqual(k.knapsack(_lowercase , _lowercase , _lowercase , _lowercase ) , 0 ) __a : int = [60] __a : Union[str, Any] = [10] __a : Tuple = len(_lowercase ) self.assertEqual(k.knapsack(_lowercase , _lowercase , _lowercase , _lowercase ) , 0 ) def lowerCAmelCase__(self ): '''simple docstring''' __a : int = 3 __a : str = [1, 2, 3] __a : Optional[Any] = [3, 2, 1] __a : int = len(_lowercase ) self.assertEqual(k.knapsack(_lowercase , _lowercase , _lowercase , _lowercase ) , 5 ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Dict = 50 __a : Tuple = [60, 100, 120] __a : List[str] = [10, 20, 30] __a : Union[str, Any] = len(_lowercase ) self.assertEqual(k.knapsack(_lowercase , _lowercase , _lowercase , _lowercase ) , 220 ) if __name__ == "__main__": unittest.main()
63
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { "facebook/timesformer": "https://huggingface.co/facebook/timesformer/resolve/main/config.json", } class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = "timesformer" def __init__(self , _lowercase=224 , _lowercase=16 , _lowercase=3 , _lowercase=8 , _lowercase=768 , _lowercase=12 , _lowercase=12 , _lowercase=3072 , _lowercase="gelu" , _lowercase=0.0 , _lowercase=0.0 , _lowercase=0.02 , _lowercase=1e-6 , _lowercase=True , _lowercase="divided_space_time" , _lowercase=0 , **_lowercase , ): '''simple docstring''' super().__init__(**_lowercase ) __a : Union[str, Any] = image_size __a : Optional[Any] = patch_size __a : Union[str, Any] = num_channels __a : Optional[int] = num_frames __a : Optional[int] = hidden_size __a : Any = num_hidden_layers __a : str = num_attention_heads __a : Optional[Any] = intermediate_size __a : Dict = hidden_act __a : List[Any] = hidden_dropout_prob __a : str = attention_probs_dropout_prob __a : Any = initializer_range __a : Tuple = layer_norm_eps __a : Optional[Any] = qkv_bias __a : Union[str, Any] = attention_type __a : Optional[Any] = drop_path_rate
700
"""simple docstring""" from manim import * class SCREAMING_SNAKE_CASE__ ( __snake_case ): def lowerCAmelCase__(self ): '''simple docstring''' __a : List[str] = Rectangle(height=0.5 , width=0.5 ) __a : Union[str, Any] = Rectangle(height=0.25 , width=0.25 ) __a : Dict = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) __a : Dict = [mem.copy() for i in range(6 )] __a : str = [mem.copy() for i in range(6 )] __a : Tuple = VGroup(*_lowercase ).arrange(_lowercase , buff=0 ) __a : List[Any] = VGroup(*_lowercase ).arrange(_lowercase , buff=0 ) __a : List[Any] = VGroup(_lowercase , _lowercase ).arrange(_lowercase , buff=0 ) __a : Union[str, Any] = Text("""CPU""" , font_size=24 ) __a : Dict = Group(_lowercase , _lowercase ).arrange(_lowercase , buff=0.5 , aligned_edge=_lowercase ) cpu.move_to([-2.5, -0.5, 0] ) self.add(_lowercase ) __a : Optional[Any] = [mem.copy() for i in range(4 )] __a : Dict = VGroup(*_lowercase ).arrange(_lowercase , buff=0 ) __a : List[str] = Text("""GPU""" , font_size=24 ) __a : Any = Group(_lowercase , _lowercase ).arrange(_lowercase , buff=0.5 , aligned_edge=_lowercase ) gpu.move_to([-1, -1, 0] ) self.add(_lowercase ) __a : List[Any] = [mem.copy() for i in range(6 )] __a : Any = VGroup(*_lowercase ).arrange(_lowercase , buff=0 ) __a : Optional[Any] = Text("""Model""" , font_size=24 ) __a : Any = Group(_lowercase , _lowercase ).arrange(_lowercase , buff=0.5 , aligned_edge=_lowercase ) model.move_to([3, -1.0, 0] ) self.add(_lowercase ) __a : Tuple = [] __a : Tuple = [] __a : Optional[int] = [] for i, rect in enumerate(_lowercase ): rect.set_stroke(_lowercase ) __a : str = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(_lowercase , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=_lowercase ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(model_cpu_arr[0] , direction=_lowercase , buff=0.0 ) else: cpu_target.next_to(model_cpu_arr[i - 1] , direction=_lowercase , buff=0.0 ) self.add(_lowercase ) model_cpu_arr.append(_lowercase ) self.add(*_lowercase , *_lowercase , *_lowercase ) __a : Optional[Any] = [mem.copy() for i in range(6 )] __a : Union[str, Any] = VGroup(*_lowercase ).arrange(_lowercase , buff=0 ) __a : Any = Text("""Loaded Checkpoint""" , font_size=24 ) __a : str = Group(_lowercase , _lowercase ).arrange(_lowercase , buff=0.5 , aligned_edge=_lowercase ) checkpoint.move_to([3, 0.5, 0] ) self.add(_lowercase ) __a : Dict = [] __a : int = [] for i, rect in enumerate(_lowercase ): __a : List[str] = fill.copy().set_fill(_lowercase , opacity=0.7 ) target.move_to(_lowercase ) ckpt_arr.append(_lowercase ) __a : Union[str, Any] = target.copy() if i < 5: cpu_target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.move_to(cpu_right_col_base[i - 5] ) ckpt_cpu_arr.append(_lowercase ) self.add(*_lowercase , *_lowercase ) __a : List[str] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) __a : List[Any] = MarkupText( F'''<b>Key:</b>\n\n<span fgcolor=\'{YELLOW}\'>●</span> Empty Model''' , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(_lowercase , _lowercase ) __a : str = MarkupText( F'''<span fgcolor=\'{BLUE}\'>●</span> Checkpoint''' , font_size=18 , ) blue_text.next_to(_lowercase , DOWN * 2.4 , aligned_edge=key_text.get_left() ) self.add(_lowercase ) __a : Optional[int] = MarkupText( F'''Based on the passed in configuration, weights are stored in\na variety of np.memmaps on disk or to a particular device.''' , font_size=24 , ) step_a.move_to([2, 2, 0] ) __a : List[Any] = [meta_mem.copy() for i in range(6 )] __a : Optional[int] = [meta_mem.copy() for i in range(6 )] __a : List[Any] = VGroup(*_lowercase ).arrange(_lowercase , buff=0 ) __a : List[str] = VGroup(*_lowercase ).arrange(_lowercase , buff=0 ) __a : Tuple = VGroup(_lowercase , _lowercase ).arrange(_lowercase , buff=0 ) __a : Dict = Text("""Disk""" , font_size=24 ) __a : Dict = Group(_lowercase , _lowercase ).arrange(_lowercase , buff=0.5 , aligned_edge=_lowercase ) disk.move_to([-4.0, -1.25, 0] ) self.play(Write(_lowercase , run_time=3 ) , Write(_lowercase , run_time=1 ) , Create(_lowercase , run_time=1 ) ) __a : Optional[Any] = [] for i, rect in enumerate(_lowercase ): __a : List[str] = rect.copy() target.generate_target() target.target.move_to(disk_left_col_base[i] ).scale(0.5 ) animations.append(MoveToTarget(_lowercase , run_time=1.5 ) ) self.play(*_lowercase ) self.play(FadeOut(_lowercase ) ) __a : List[str] = MarkupText(F'''Then, the checkpoint is removed from memory\nthrough garbage collection.''' , font_size=24 ) step_a.move_to([2, 2, 0] ) self.play(Write(_lowercase , run_time=3 ) ) self.play( FadeOut(_lowercase , _lowercase , *_lowercase , *_lowercase ) , ) self.wait()
63
0
"""simple docstring""" from unittest import TestCase from datasets import Dataset from minhash_deduplication import deduplicate_dataset, make_duplicate_clusters def __magic_name__ ( ): __a : Dict = { """repo_name""": ["""test_repo1""", """test_repo2""", """test_repo3"""], """path""": ["""test_1.py""", """test_2.py""", """unit_test.py"""], """content""": ["""a """ * 2_0, """a """ * 3_0, """b """ * 7], } __a : Optional[Any] = Dataset.from_dict(_lowerCamelCase ) return dataset class SCREAMING_SNAKE_CASE__ ( __snake_case ): def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[int] = get_dataset() __a : List[Any] = make_duplicate_clusters(_lowercase , 0.85 ) self.assertEqual(len(duplicate_clusters[0] ) , 2 ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Tuple = get_dataset() __a : Optional[Any] = deduplicate_dataset(_lowercase ) self.assertEqual(len(_lowercase ) , 2 ) print(_lowercase ) self.assertEqual(duplicate_clusters[0][0]["""copies"""] , 2 ) self.assertEqual(duplicate_clusters[0][0]["""is_extreme"""] , _lowercase )
701
"""simple docstring""" def __magic_name__ ( _lowerCamelCase : int , _lowerCamelCase : float , _lowerCamelCase : float ): return round(float(moles / volume ) * nfactor ) def __magic_name__ ( _lowerCamelCase : float , _lowerCamelCase : float , _lowerCamelCase : float ): return round(float((moles * 0.08_21 * temperature) / (volume) ) ) def __magic_name__ ( _lowerCamelCase : float , _lowerCamelCase : float , _lowerCamelCase : float ): return round(float((moles * 0.08_21 * temperature) / (pressure) ) ) def __magic_name__ ( _lowerCamelCase : float , _lowerCamelCase : float , _lowerCamelCase : float ): return round(float((pressure * volume) / (0.08_21 * moles) ) ) if __name__ == "__main__": import doctest doctest.testmod()
63
0
from typing import List, Optional, Union import numpy as np import PIL import torch from PIL import Image from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) lowercase__ = logging.get_logger(__name__) # pylint: disable=invalid-name lowercase__ = "\n Examples:\n ```py\n >>> from diffusers import KandinskyV22Img2ImgPipeline, KandinskyV22PriorPipeline\n >>> from diffusers.utils import load_image\n >>> import torch\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\n ... \"kandinsky-community/kandinsky-2-2-prior\", torch_dtype=torch.float16\n ... )\n >>> pipe_prior.to(\"cuda\")\n\n >>> prompt = \"A red cartoon frog, 4k\"\n >>> image_emb, zero_image_emb = pipe_prior(prompt, return_dict=False)\n\n >>> pipe = KandinskyV22Img2ImgPipeline.from_pretrained(\n ... \"kandinsky-community/kandinsky-2-2-decoder\", torch_dtype=torch.float16\n ... )\n >>> pipe.to(\"cuda\")\n\n >>> init_image = load_image(\n ... \"https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main\"\n ... \"/kandinsky/frog.png\"\n ... )\n\n >>> image = pipe(\n ... image=init_image,\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=100,\n ... strength=0.2,\n ... ).images\n\n >>> image[0].save(\"red_frog.png\")\n ```\n" def __magic_name__ ( _lowerCamelCase : Any , _lowerCamelCase : List[str] , _lowerCamelCase : List[Any]=8 ): __a : List[str] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 __a : Any = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor def __magic_name__ ( _lowerCamelCase : int , _lowerCamelCase : Optional[Any]=5_1_2 , _lowerCamelCase : str=5_1_2 ): __a : Dict = pil_image.resize((w, h) , resample=Image.BICUBIC , reducing_gap=1 ) __a : Any = np.array(pil_image.convert("""RGB""" ) ) __a : Any = arr.astype(np.floataa ) / 1_2_7.5 - 1 __a : Any = np.transpose(_lowerCamelCase , [2, 0, 1] ) __a : Dict = torch.from_numpy(_lowerCamelCase ).unsqueeze(0 ) return image class SCREAMING_SNAKE_CASE__ ( __snake_case ): def __init__(self , _lowercase , _lowercase , _lowercase , ): '''simple docstring''' super().__init__() self.register_modules( unet=_lowercase , scheduler=_lowercase , movq=_lowercase , ) __a : List[str] = 2 ** (len(self.movq.config.block_out_channels ) - 1) def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase ): '''simple docstring''' __a : Optional[Any] = min(int(num_inference_steps * strength ) , _lowercase ) __a : str = max(num_inference_steps - init_timestep , 0 ) __a : List[Any] = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase=None ): '''simple docstring''' if not isinstance(_lowercase , (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( F'''`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(_lowercase )}''' ) __a : Optional[Any] = image.to(device=_lowercase , dtype=_lowercase ) __a : Optional[int] = batch_size * num_images_per_prompt if image.shape[1] == 4: __a : int = image else: if isinstance(_lowercase , _lowercase ) and len(_lowercase ) != batch_size: raise ValueError( F'''You have passed a list of generators of length {len(_lowercase )}, but requested an effective batch''' F''' size of {batch_size}. Make sure the batch size matches the length of the generators.''' ) elif isinstance(_lowercase , _lowercase ): __a : int = [ self.movq.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(_lowercase ) ] __a : Optional[Any] = torch.cat(_lowercase , dim=0 ) else: __a : Optional[Any] = self.movq.encode(_lowercase ).latent_dist.sample(_lowercase ) __a : Union[str, Any] = self.movq.config.scaling_factor * init_latents __a : Dict = torch.cat([init_latents] , dim=0 ) __a : Dict = init_latents.shape __a : Optional[int] = randn_tensor(_lowercase , generator=_lowercase , device=_lowercase , dtype=_lowercase ) # get latents __a : Any = self.scheduler.add_noise(_lowercase , _lowercase , _lowercase ) __a : Any = init_latents return latents def lowerCAmelCase__(self , _lowercase=0 ): '''simple docstring''' if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("""Please install accelerate via `pip install accelerate`""" ) __a : Any = torch.device(F'''cuda:{gpu_id}''' ) __a : List[Any] = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(_lowercase , _lowercase ) def lowerCAmelCase__(self , _lowercase=0 ): '''simple docstring''' if is_accelerate_available() and is_accelerate_version(""">=""" , """0.17.0.dev0""" ): from accelerate import cpu_offload_with_hook else: raise ImportError("""`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.""" ) __a : List[Any] = torch.device(F'''cuda:{gpu_id}''' ) if self.device.type != "cpu": self.to("""cpu""" , silence_dtype_warnings=_lowercase ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) __a : Tuple = None for cpu_offloaded_model in [self.unet, self.movq]: __a : Optional[int] = cpu_offload_with_hook(_lowercase , _lowercase , prev_module_hook=_lowercase ) # We'll offload the last model manually. __a : Tuple = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def lowerCAmelCase__(self ): '''simple docstring''' if not hasattr(self.unet , """_hf_hook""" ): return self.device for module in self.unet.modules(): if ( hasattr(_lowercase , """_hf_hook""" ) and hasattr(module._hf_hook , """execution_device""" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(_lowercase ) def __call__(self , _lowercase , _lowercase , _lowercase , _lowercase = 512 , _lowercase = 512 , _lowercase = 100 , _lowercase = 4.0 , _lowercase = 0.3 , _lowercase = 1 , _lowercase = None , _lowercase = "pil" , _lowercase = True , ): '''simple docstring''' __a : List[Any] = self._execution_device __a : List[str] = guidance_scale > 1.0 if isinstance(_lowercase , _lowercase ): __a : Tuple = torch.cat(_lowercase , dim=0 ) __a : Union[str, Any] = image_embeds.shape[0] if isinstance(_lowercase , _lowercase ): __a : Optional[int] = torch.cat(_lowercase , dim=0 ) if do_classifier_free_guidance: __a : int = image_embeds.repeat_interleave(_lowercase , dim=0 ) __a : Union[str, Any] = negative_image_embeds.repeat_interleave(_lowercase , dim=0 ) __a : Union[str, Any] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=_lowercase ) if not isinstance(_lowercase , _lowercase ): __a : List[str] = [image] if not all(isinstance(_lowercase , (PIL.Image.Image, torch.Tensor) ) for i in image ): raise ValueError( F'''Input is in incorrect format: {[type(_lowercase ) for i in image]}. Currently, we only support PIL image and pytorch tensor''' ) __a : str = torch.cat([prepare_image(_lowercase , _lowercase , _lowercase ) for i in image] , dim=0 ) __a : Optional[int] = image.to(dtype=image_embeds.dtype , device=_lowercase ) __a : Optional[Any] = self.movq.encode(_lowercase )["""latents"""] __a : List[Any] = latents.repeat_interleave(_lowercase , dim=0 ) self.scheduler.set_timesteps(_lowercase , device=_lowercase ) __a : Optional[int] = self.get_timesteps(_lowercase , _lowercase , _lowercase ) __a : Optional[int] = timesteps[:1].repeat(batch_size * num_images_per_prompt ) __a : int = downscale_height_and_width(_lowercase , _lowercase , self.movq_scale_factor ) __a : List[str] = self.prepare_latents( _lowercase , _lowercase , _lowercase , _lowercase , image_embeds.dtype , _lowercase , _lowercase ) for i, t in enumerate(self.progress_bar(_lowercase ) ): # expand the latents if we are doing classifier free guidance __a : Dict = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents __a : Tuple = {"""image_embeds""": image_embeds} __a : str = self.unet( sample=_lowercase , timestep=_lowercase , encoder_hidden_states=_lowercase , added_cond_kwargs=_lowercase , return_dict=_lowercase , )[0] if do_classifier_free_guidance: __a : Union[str, Any] = noise_pred.split(latents.shape[1] , dim=1 ) __a : int = noise_pred.chunk(2 ) __a : int = variance_pred.chunk(2 ) __a : List[Any] = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) __a : Optional[Any] = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , """variance_type""" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): __a : Optional[int] = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 __a : Optional[int] = self.scheduler.step( _lowercase , _lowercase , _lowercase , generator=_lowercase , )[0] # post-processing __a : int = self.movq.decode(_lowercase , force_not_quantize=_lowercase )["""sample"""] if output_type not in ["pt", "np", "pil"]: raise ValueError(F'''Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}''' ) if output_type in ["np", "pil"]: __a : List[Any] = image * 0.5 + 0.5 __a : Union[str, Any] = image.clamp(0 , 1 ) __a : List[str] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": __a : Optional[int] = self.numpy_to_pil(_lowercase ) if not return_dict: return (image,) return ImagePipelineOutput(images=_lowercase )
702
"""simple docstring""" def __magic_name__ ( _lowerCamelCase : list[int] ): if not nums: # Makes sure that the list is not empty raise ValueError("""List is empty""" ) __a : Any = sum(_lowerCamelCase ) / len(_lowerCamelCase ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(_lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
63
0
"""simple docstring""" import inspect import unittest from typing import List import numpy as np from transformers import EfficientFormerConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, ) from transformers.models.efficientformer.modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_vision_available(): from PIL import Image from transformers import EfficientFormerImageProcessor class SCREAMING_SNAKE_CASE__ : def __init__(self , _lowercase , _lowercase = 13 , _lowercase = 64 , _lowercase = 2 , _lowercase = 3 , _lowercase = 3 , _lowercase = True , _lowercase = True , _lowercase = 128 , _lowercase=[16, 32, 64, 128] , _lowercase = 7 , _lowercase = 4 , _lowercase = 37 , _lowercase = "gelu" , _lowercase = 0.1 , _lowercase = 0.1 , _lowercase = 10 , _lowercase = 0.02 , _lowercase = 2 , _lowercase = 1 , _lowercase = 128 , _lowercase = [2, 2, 2, 2] , _lowercase = 2 , _lowercase = 2 , ): '''simple docstring''' __a : str = parent __a : List[Any] = batch_size __a : int = image_size __a : Tuple = patch_size __a : str = num_channels __a : Union[str, Any] = is_training __a : List[Any] = use_labels __a : int = hidden_size __a : Optional[Any] = num_hidden_layers __a : List[Any] = num_attention_heads __a : Dict = intermediate_size __a : str = hidden_act __a : Dict = hidden_dropout_prob __a : str = attention_probs_dropout_prob __a : Optional[int] = type_sequence_label_size __a : Dict = initializer_range __a : Dict = encoder_stride __a : int = num_attention_outputs __a : List[Any] = embed_dim __a : Optional[Any] = embed_dim + 1 __a : Optional[Any] = resolution __a : Optional[Any] = depths __a : Union[str, Any] = hidden_sizes __a : List[str] = dim __a : Any = mlp_expansion_ratio def lowerCAmelCase__(self ): '''simple docstring''' __a : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __a : str = None if self.use_labels: __a : List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __a : List[str] = self.get_config() return config, pixel_values, labels def lowerCAmelCase__(self ): '''simple docstring''' return EfficientFormerConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowercase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , resolution=self.resolution , depths=self.depths , hidden_sizes=self.hidden_sizes , dim=self.dim , mlp_expansion_ratio=self.mlp_expansion_ratio , ) def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase ): '''simple docstring''' __a : Optional[Any] = TFEfficientFormerModel(config=_lowercase ) __a : List[Any] = model(_lowercase , training=_lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase ): '''simple docstring''' __a : Optional[Any] = self.type_sequence_label_size __a : Any = TFEfficientFormerForImageClassification(_lowercase ) __a : Union[str, Any] = model(_lowercase , labels=_lowercase , training=_lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __a : Optional[Any] = 1 __a : int = TFEfficientFormerForImageClassification(_lowercase ) __a : List[str] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __a : str = model(_lowercase , labels=_lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Any = self.prepare_config_and_inputs() __a : Tuple = config_and_inputs __a : Tuple = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class SCREAMING_SNAKE_CASE__ ( __snake_case , __snake_case , unittest.TestCase ): _lowerCAmelCase = ( ( TFEfficientFormerModel, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerForImageClassification, ) if is_tf_available() else () ) _lowerCAmelCase = ( { "feature-extraction": TFEfficientFormerModel, "image-classification": ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, ), } if is_tf_available() else {} ) _lowerCAmelCase = False _lowerCAmelCase = False _lowerCAmelCase = False _lowerCAmelCase = False _lowerCAmelCase = False def lowerCAmelCase__(self ): '''simple docstring''' __a : Tuple = TFEfficientFormerModelTester(self ) __a : Any = ConfigTester( self , config_class=_lowercase , has_text_modality=_lowercase , hidden_size=37 ) def lowerCAmelCase__(self ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="""EfficientFormer does not use inputs_embeds""" ) def lowerCAmelCase__(self ): '''simple docstring''' pass @unittest.skip(reason="""EfficientFormer does not support input and output embeddings""" ) def lowerCAmelCase__(self ): '''simple docstring''' pass def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a : Dict = model_class(_lowercase ) __a : Optional[Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __a : Optional[Any] = [*signature.parameters.keys()] __a : Union[str, Any] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' def check_hidden_states_output(_lowercase , _lowercase , _lowercase ): __a : Tuple = model_class(_lowercase ) __a : int = model(**self._prepare_for_class(_lowercase , _lowercase ) , training=_lowercase ) __a : Tuple = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __a : str = getattr( self.model_tester , """expected_num_hidden_layers""" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(_lowercase ) , _lowercase ) if hasattr(self.model_tester , """encoder_seq_length""" ): __a : Any = self.model_tester.encoder_seq_length if hasattr(self.model_tester , """chunk_length""" ) and self.model_tester.chunk_length > 1: __a : int = seq_length * self.model_tester.chunk_length else: __a : Any = self.model_tester.seq_length self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) if config.is_encoder_decoder: __a : Optional[int] = outputs.decoder_hidden_states self.asseretIsInstance(_lowercase , (list, tuple) ) self.assertEqual(len(_lowercase ) , _lowercase ) __a : Any = getattr(self.model_tester , """seq_length""" , _lowercase ) __a : List[Any] = getattr(self.model_tester , """decoder_seq_length""" , _lowercase ) self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [decoder_seq_length, self.model_tester.hidden_size] , ) __a : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a : Dict = True check_hidden_states_output(_lowercase , _lowercase , _lowercase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __a : int = True check_hidden_states_output(_lowercase , _lowercase , _lowercase ) def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase=False ): '''simple docstring''' __a : Any = super()._prepare_for_class(_lowercase , _lowercase , return_labels=_lowercase ) if return_labels: if model_class.__name__ == "TFEfficientFormerForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowercase ) @unittest.skip(reason="""EfficientFormer does not implement masked image modeling yet""" ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowercase ) @slow def lowerCAmelCase__(self ): '''simple docstring''' for model_name in TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a : Union[str, Any] = TFEfficientFormerModel.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() __a : int = True __a : Optional[int] = getattr(self.model_tester , """seq_length""" , _lowercase ) __a : Dict = getattr(self.model_tester , """encoder_seq_length""" , _lowercase ) __a : Dict = getattr(self.model_tester , """key_length""" , _lowercase ) __a : int = getattr(self.model_tester , """chunk_length""" , _lowercase ) if chunk_length is not None and hasattr(self.model_tester , """num_hashes""" ): __a : List[str] = encoder_seq_length * self.model_tester.num_hashes for model_class in self.all_model_classes: __a : List[Any] = True __a : Tuple = False __a : List[Any] = True __a : int = model_class(_lowercase ) __a : List[Any] = model(**self._prepare_for_class(_lowercase , _lowercase ) , training=_lowercase ) __a : Dict = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(_lowercase ) , self.model_tester.num_attention_outputs ) # check that output_attentions also work using config del inputs_dict["output_attentions"] __a : Optional[Any] = True __a : List[str] = model_class(_lowercase ) __a : Dict = model(**self._prepare_for_class(_lowercase , _lowercase ) , training=_lowercase ) __a : int = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(_lowercase ) , self.model_tester.num_attention_outputs ) if chunk_length is not None: self.assertListEqual( list(attentions[0].shape[-4:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, chunk_length, encoder_key_length] , ) else: self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, encoder_key_length] , ) def lowerCAmelCase__(self ): '''simple docstring''' __a : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # Prepare our model __a : Dict = model_class(_lowercase ) # These are maximally general inputs for the model, with multiple None dimensions # Hopefully this will catch any conditionals that fail for flexible shapes __a : Optional[Any] = { key: tf.keras.Input(shape=val.shape[1:] , dtype=val.dtype , name=_lowercase ) for key, val in model.input_signature.items() if key in model.dummy_inputs } __a : Optional[Any] = model(_lowercase ) self.assertTrue(outputs_dict is not None ) def __magic_name__ ( ): __a : int = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @cached_property def lowerCAmelCase__(self ): '''simple docstring''' return ( EfficientFormerImageProcessor.from_pretrained("""snap-research/efficientformer-l1-300""" ) if is_vision_available() else None ) @slow def lowerCAmelCase__(self ): '''simple docstring''' __a : str = TFEfficientFormerForImageClassification.from_pretrained("""snap-research/efficientformer-l1-300""" ) __a : Optional[Any] = self.default_image_processor __a : List[str] = prepare_img() __a : int = image_processor(images=_lowercase , return_tensors="""tf""" ) # forward pass __a : Optional[Any] = model(**_lowercase , training=_lowercase ) # verify the logits __a : str = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowercase ) __a : Dict = tf.constant([-0.0555, 0.4825, -0.0852] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , _lowercase , atol=1e-4 ) ) @slow def lowerCAmelCase__(self ): '''simple docstring''' __a : Any = TFEfficientFormerForImageClassificationWithTeacher.from_pretrained( """snap-research/efficientformer-l1-300""" ) __a : Any = self.default_image_processor __a : str = prepare_img() __a : str = image_processor(images=_lowercase , return_tensors="""tf""" ) # forward pass __a : List[Any] = model(**_lowercase , training=_lowercase ) # verify the logits __a : int = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowercase ) __a : List[str] = tf.constant([-0.1312, 0.4353, -1.0499] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , _lowercase , atol=1e-4 ) )
703
"""simple docstring""" import math import sys import cva import numpy as np def __magic_name__ ( _lowerCamelCase : np.ndarray , _lowerCamelCase : float ): # For applying gaussian function for each element in matrix. __a : int = math.sqrt(_lowerCamelCase ) __a : Any = 1 / (sigma * math.sqrt(2 * math.pi )) return cons * np.exp(-((img / sigma) ** 2) * 0.5 ) def __magic_name__ ( _lowerCamelCase : np.ndarray , _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : int ): __a : Any = kernel_size // 2 return img[x - half : x + half + 1, y - half : y + half + 1] def __magic_name__ ( _lowerCamelCase : int , _lowerCamelCase : float ): # Creates a gaussian kernel of given dimension. __a : int = np.zeros((kernel_size, kernel_size) ) for i in range(0 , _lowerCamelCase ): for j in range(0 , _lowerCamelCase ): __a : Any = math.sqrt( abs(i - kernel_size // 2 ) ** 2 + abs(j - kernel_size // 2 ) ** 2 ) return vec_gaussian(_lowerCamelCase , _lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : np.ndarray , _lowerCamelCase : float , _lowerCamelCase : float , _lowerCamelCase : int , ): __a : Tuple = np.zeros(img.shape ) __a : Optional[int] = get_gauss_kernel(_lowerCamelCase , _lowerCamelCase ) __a , __a : int = img.shape for i in range(kernel_size // 2 , size_x - kernel_size // 2 ): for j in range(kernel_size // 2 , size_y - kernel_size // 2 ): __a : List[str] = get_slice(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) __a : Any = img_s - img_s[kernel_size // 2, kernel_size // 2] __a : Optional[Any] = vec_gaussian(_lowerCamelCase , _lowerCamelCase ) __a : Optional[Any] = np.multiply(_lowerCamelCase , _lowerCamelCase ) __a : Any = np.multiply(_lowerCamelCase , _lowerCamelCase ) __a : Tuple = np.sum(_lowerCamelCase ) / np.sum(_lowerCamelCase ) __a : Optional[Any] = val return imga def __magic_name__ ( _lowerCamelCase : list ): __a : Optional[Any] = args[1] if args[1:] else """../image_data/lena.jpg""" __a : Union[str, Any] = float(args[2] ) if args[2:] else 1.0 __a : Optional[int] = float(args[3] ) if args[3:] else 1.0 if args[4:]: __a : Any = int(args[4] ) __a : Any = kernel_size + abs(kernel_size % 2 - 1 ) else: __a : Optional[int] = 5 return filename, spatial_variance, intensity_variance, kernel_size if __name__ == "__main__": lowercase__ , lowercase__ , lowercase__ , lowercase__ = parse_args(sys.argv) lowercase__ = cva.imread(filename, 0) cva.imshow("input image", img) lowercase__ = img / 255 lowercase__ = out.astype("float32") lowercase__ = bilateral_filter(out, spatial_variance, intensity_variance, kernel_size) lowercase__ = out * 255 lowercase__ = np.uinta(out) cva.imshow("output image", out) cva.waitKey(0) cva.destroyAllWindows()
63
0
"""simple docstring""" import sys from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be checked before tokenizers lowercase__ = "python tqdm regex requests packaging filelock numpy tokenizers".split() if sys.version_info < (3, 7): pkgs_to_check_at_runtime.append("dataclasses") if sys.version_info < (3, 8): pkgs_to_check_at_runtime.append("importlib_metadata") for pkg in pkgs_to_check_at_runtime: if pkg in deps: if pkg == "tokenizers": # must be loaded here, or else tqdm check may fail from .utils import is_tokenizers_available if not is_tokenizers_available(): continue # not required, check version only if installed require_version_core(deps[pkg]) else: raise ValueError(f'can\'t find {pkg} in {deps.keys()}, check dependency_versions_table.py') def __magic_name__ ( _lowerCamelCase : Optional[int] , _lowerCamelCase : str=None ): require_version(deps[pkg] , _lowerCamelCase )
704
"""simple docstring""" from unittest import TestCase from datasets import Dataset from minhash_deduplication import deduplicate_dataset, make_duplicate_clusters def __magic_name__ ( ): __a : Dict = { """repo_name""": ["""test_repo1""", """test_repo2""", """test_repo3"""], """path""": ["""test_1.py""", """test_2.py""", """unit_test.py"""], """content""": ["""a """ * 2_0, """a """ * 3_0, """b """ * 7], } __a : Optional[Any] = Dataset.from_dict(_lowerCamelCase ) return dataset class SCREAMING_SNAKE_CASE__ ( __snake_case ): def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[int] = get_dataset() __a : List[Any] = make_duplicate_clusters(_lowercase , 0.85 ) self.assertEqual(len(duplicate_clusters[0] ) , 2 ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Tuple = get_dataset() __a , __a : Optional[Any] = deduplicate_dataset(_lowercase ) self.assertEqual(len(_lowercase ) , 2 ) print(_lowercase ) self.assertEqual(duplicate_clusters[0][0]["""copies"""] , 2 ) self.assertEqual(duplicate_clusters[0][0]["""is_extreme"""] , _lowercase )
63
0
"""simple docstring""" import inspect import logging import os import random import shutil import tempfile import unittest import pytest import torch from torch import nn from torch.utils.data import DataLoader, TensorDataset from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_cuda from accelerate.utils import ProjectConfiguration, set_seed lowercase__ = logging.getLogger(__name__) def __magic_name__ ( _lowerCamelCase : Tuple=2 , _lowerCamelCase : Union[str, Any]=3 , _lowerCamelCase : Optional[int]=1_6 , _lowerCamelCase : int = 1_0 , _lowerCamelCase : int = 2 ): def get_dataset(_lowerCamelCase : str ): __a : Tuple = torch.randn(batch_size * n_batches , 1 ) return TensorDataset(_lowerCamelCase , a * x + b + 0.1 * torch.randn(batch_size * n_batches , 1 ) ) __a : List[str] = get_dataset(_lowerCamelCase ) __a : int = get_dataset(_lowerCamelCase ) __a : Tuple = DataLoader(_lowerCamelCase , shuffle=_lowerCamelCase , batch_size=_lowerCamelCase , num_workers=4 ) __a : str = DataLoader(_lowerCamelCase , shuffle=_lowerCamelCase , batch_size=_lowerCamelCase , num_workers=4 ) return (train_dataloader, valid_dataloader) def __magic_name__ ( _lowerCamelCase : Tuple , _lowerCamelCase : Optional[int] , _lowerCamelCase : str , _lowerCamelCase : Tuple , _lowerCamelCase : Union[str, Any] , _lowerCamelCase : Optional[Any]=None ): __a : Dict = [] for epoch in range(_lowerCamelCase ): # Train quickly model.train() for batch in dataloader: __a : Tuple = batch __a : Union[str, Any] = model(_lowerCamelCase ) __a : int = torch.nn.functional.mse_loss(_lowerCamelCase , _lowerCamelCase ) accelerator.backward(_lowerCamelCase ) optimizer.step() optimizer.zero_grad() rands.append(random.random() ) # Introduce some randomness if scheduler is not None: scheduler.step() return rands class SCREAMING_SNAKE_CASE__ ( nn.Module ): def __init__(self ): '''simple docstring''' super().__init__() __a : int = nn.Parameter(torch.randn(1 ) ) __a : Any = nn.Parameter(torch.randn(1 ) ) def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' return x * self.a + self.b class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def lowerCAmelCase__(self ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) __a : Optional[int] = DummyModel() __a : List[str] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) __a : Optional[int] = dummy_dataloaders() __a : List[str] = ProjectConfiguration(total_limit=1 , project_dir=_lowercase , automatic_checkpoint_naming=_lowercase ) # Train baseline __a : str = Accelerator(project_config=_lowercase ) __a : Optional[int] = accelerator.prepare( _lowercase , _lowercase , _lowercase , _lowercase ) # Save initial accelerator.save_state() # Save second state accelerator.save_state() self.assertEqual(len(os.listdir(accelerator.project_dir ) ) , 1 ) def lowerCAmelCase__(self ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) __a : Optional[int] = DummyModel() __a : List[str] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) __a : Dict = dummy_dataloaders() # Train baseline __a : Dict = Accelerator() __a : Dict = accelerator.prepare( _lowercase , _lowercase , _lowercase , _lowercase ) # Save initial __a : Tuple = os.path.join(_lowercase , """initial""" ) accelerator.save_state(_lowercase ) (__a) : Tuple = model.a.item(), model.b.item() __a : Optional[int] = optimizer.state_dict() __a : Tuple = train(3 , _lowercase , _lowercase , _lowercase , _lowercase ) (__a) : List[Any] = model.a.item(), model.b.item() __a : List[str] = optimizer.state_dict() # Train partially set_seed(42 ) __a : Tuple = DummyModel() __a : Any = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) __a : int = dummy_dataloaders() __a : Optional[Any] = Accelerator() __a : List[Any] = accelerator.prepare( _lowercase , _lowercase , _lowercase , _lowercase ) accelerator.load_state(_lowercase ) (__a) : Union[str, Any] = model.a.item(), model.b.item() __a : List[Any] = optimizer.state_dict() self.assertEqual(_lowercase , _lowercase ) self.assertEqual(_lowercase , _lowercase ) self.assertEqual(_lowercase , _lowercase ) __a : Optional[Any] = train(2 , _lowercase , _lowercase , _lowercase , _lowercase ) # Save everything __a : Optional[Any] = os.path.join(_lowercase , """checkpoint""" ) accelerator.save_state(_lowercase ) # Load everything back in and make sure all states work accelerator.load_state(_lowercase ) test_rands += train(1 , _lowercase , _lowercase , _lowercase , _lowercase ) (__a) : str = model.a.item(), model.b.item() __a : int = optimizer.state_dict() self.assertEqual(_lowercase , _lowercase ) self.assertEqual(_lowercase , _lowercase ) self.assertEqual(_lowercase , _lowercase ) self.assertEqual(_lowercase , _lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) __a : List[Any] = DummyModel() __a : Any = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) __a : List[str] = dummy_dataloaders() __a : Optional[Any] = ProjectConfiguration(automatic_checkpoint_naming=_lowercase ) # Train baseline __a : List[str] = Accelerator(project_dir=_lowercase , project_config=_lowercase ) __a : Union[str, Any] = accelerator.prepare( _lowercase , _lowercase , _lowercase , _lowercase ) # Save initial accelerator.save_state() (__a) : Dict = model.a.item(), model.b.item() __a : Optional[Any] = optimizer.state_dict() __a : str = train(3 , _lowercase , _lowercase , _lowercase , _lowercase ) (__a) : Union[str, Any] = model.a.item(), model.b.item() __a : int = optimizer.state_dict() # Train partially set_seed(42 ) __a : Union[str, Any] = DummyModel() __a : List[Any] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) __a : List[Any] = dummy_dataloaders() __a : int = ProjectConfiguration(iteration=1 , automatic_checkpoint_naming=_lowercase ) __a : str = Accelerator(project_dir=_lowercase , project_config=_lowercase ) __a : Dict = accelerator.prepare( _lowercase , _lowercase , _lowercase , _lowercase ) accelerator.load_state(os.path.join(_lowercase , """checkpoints""" , """checkpoint_0""" ) ) (__a) : Optional[Any] = model.a.item(), model.b.item() __a : Union[str, Any] = optimizer.state_dict() self.assertEqual(_lowercase , _lowercase ) self.assertEqual(_lowercase , _lowercase ) self.assertEqual(_lowercase , _lowercase ) __a : Dict = train(2 , _lowercase , _lowercase , _lowercase , _lowercase ) # Save everything accelerator.save_state() # Load everything back in and make sure all states work accelerator.load_state(os.path.join(_lowercase , """checkpoints""" , """checkpoint_1""" ) ) test_rands += train(1 , _lowercase , _lowercase , _lowercase , _lowercase ) (__a) : int = model.a.item(), model.b.item() __a : str = optimizer.state_dict() self.assertEqual(_lowercase , _lowercase ) self.assertEqual(_lowercase , _lowercase ) self.assertEqual(_lowercase , _lowercase ) self.assertEqual(_lowercase , _lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' __a : str = torch.tensor([1, 2, 3] ) __a : Any = torch.tensor([2, 3, 4] ) __a : List[str] = DummyModel() __a : Optional[Any] = torch.optim.Adam(net.parameters() ) __a : List[str] = Accelerator() with self.assertRaises(_lowercase ) as ve: accelerator.register_for_checkpointing(_lowercase , _lowercase , _lowercase , _lowercase ) __a : List[Any] = str(ve.exception ) self.assertTrue("""Item at index 0""" in message ) self.assertTrue("""Item at index 1""" in message ) self.assertFalse("""Item at index 2""" in message ) self.assertFalse("""Item at index 3""" in message ) def lowerCAmelCase__(self ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) __a : int = DummyModel() __a : int = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) __a : Tuple = torch.optim.lr_scheduler.StepLR(_lowercase , step_size=1 , gamma=0.99 ) __a : Optional[Any] = dummy_dataloaders() __a : Optional[int] = ProjectConfiguration(automatic_checkpoint_naming=_lowercase ) # Train baseline __a : List[Any] = Accelerator(project_dir=_lowercase , project_config=_lowercase ) __a : Union[str, Any] = accelerator.prepare( _lowercase , _lowercase , _lowercase , _lowercase , _lowercase ) # Save initial accelerator.save_state() __a : Optional[int] = scheduler.state_dict() train(3 , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase ) self.assertNotEqual(_lowercase , scheduler.state_dict() ) # Load everything back in and make sure all states work accelerator.load_state(os.path.join(_lowercase , """checkpoints""" , """checkpoint_0""" ) ) self.assertEqual(_lowercase , scheduler.state_dict() ) def lowerCAmelCase__(self ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) __a : Dict = DummyModel() __a : Dict = ProjectConfiguration(automatic_checkpoint_naming=_lowercase , total_limit=2 ) # Train baseline __a : Tuple = Accelerator(project_dir=_lowercase , project_config=_lowercase ) __a : str = accelerator.prepare(_lowercase ) # Save 3 states: for _ in range(11 ): accelerator.save_state() self.assertTrue(not os.path.exists(os.path.join(_lowercase , """checkpoints""" , """checkpoint_0""" ) ) ) self.assertTrue(os.path.exists(os.path.join(_lowercase , """checkpoints""" , """checkpoint_9""" ) ) ) self.assertTrue(os.path.exists(os.path.join(_lowercase , """checkpoints""" , """checkpoint_10""" ) ) ) @require_cuda def lowerCAmelCase__(self ): '''simple docstring''' __a : List[Any] = ["""torchrun""", F'''--nproc_per_node={torch.cuda.device_count()}''', inspect.getfile(self.__class__ )] execute_subprocess_async(_lowercase , env=os.environ.copy() ) if __name__ == "__main__": lowercase__ = "/tmp/accelerate/state_checkpointing" lowercase__ = DummyModel() lowercase__ = torch.optim.Adam(params=model.parameters(), lr=1e-3) lowercase__ = torch.optim.lr_scheduler.StepLR(optimizer, step_size=1, gamma=0.99) lowercase__ , lowercase__ = dummy_dataloaders() lowercase__ = ProjectConfiguration(automatic_checkpoint_naming=True) # Train baseline lowercase__ = Accelerator(project_dir=savedir, project_config=project_config, mixed_precision="no") if accelerator.process_index == 0: if os.path.exists(savedir): shutil.rmtree(savedir) os.makedirs(savedir) lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ = accelerator.prepare( model, optimizer, train_dataloader, valid_dataloader, scheduler ) lowercase__ , lowercase__ = accelerator.prepare(model, optimizer) train(3, model, train_dataloader, optimizer, accelerator, scheduler) # Check that the intial optimizer is loaded on the GPU for group in optimizer.param_groups: lowercase__ = group["params"][0].device break assert param_device.type == accelerator.device.type lowercase__ = model.cpu() accelerator.wait_for_everyone() accelerator.save_state() accelerator.wait_for_everyone() # Check CPU state accelerator.load_state(os.path.join(savedir, "checkpoints", "checkpoint_0"), map_location="cpu") for group in optimizer.param_groups: lowercase__ = group["params"][0].device break assert ( param_device.type == torch.device("cpu").type ), f"Loaded optimizer states did not match, expected to be loaded on the CPU but got {param_device}" # Check device state model.to(accelerator.device) accelerator.load_state(os.path.join(savedir, "checkpoints", "checkpoint_0"), map_location="on_device") for group in optimizer.param_groups: lowercase__ = group["params"][0].device break assert ( param_device.type == accelerator.device.type ), f"Loaded optimizer states did not match, expected to be loaded on {accelerator.device} but got {param_device}" # Check error with pytest.raises(TypeError, match="Unsupported optimizer map location passed"): accelerator.load_state(os.path.join(savedir, "checkpoints", "checkpoint_0"), map_location="invalid") accelerator.wait_for_everyone() if accelerator.process_index == 0: shutil.rmtree(savedir) accelerator.wait_for_everyone()
705
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tensorflow_text_available, is_torch_available lowercase__ = { "configuration_ernie": ["ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP", "ErnieConfig", "ErnieOnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ "ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST", "ErnieForCausalLM", "ErnieForMaskedLM", "ErnieForMultipleChoice", "ErnieForNextSentencePrediction", "ErnieForPreTraining", "ErnieForQuestionAnswering", "ErnieForSequenceClassification", "ErnieForTokenClassification", "ErnieModel", "ErniePreTrainedModel", ] if TYPE_CHECKING: from .configuration_ernie import ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP, ErnieConfig, ErnieOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ernie import ( ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST, ErnieForCausalLM, ErnieForMaskedLM, ErnieForMultipleChoice, ErnieForNextSentencePrediction, ErnieForPreTraining, ErnieForQuestionAnswering, ErnieForSequenceClassification, ErnieForTokenClassification, ErnieModel, ErniePreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
63
0
"""simple docstring""" import os import pytest import yaml from datasets.features.features import Features, Value from datasets.info import DatasetInfo, DatasetInfosDict @pytest.mark.parametrize( """files""" , [ ["""full:README.md""", """dataset_infos.json"""], ["""empty:README.md""", """dataset_infos.json"""], ["""dataset_infos.json"""], ["""full:README.md"""], ] , ) def __magic_name__ ( _lowerCamelCase : Optional[Any] , _lowerCamelCase : Any ): __a : Dict = tmp_path_factory.mktemp("""dset_infos_dir""" ) if "full:README.md" in files: with open(dataset_infos_dir / """README.md""" , """w""" ) as f: f.write("""---\ndataset_info:\n dataset_size: 42\n---""" ) if "empty:README.md" in files: with open(dataset_infos_dir / """README.md""" , """w""" ) as f: f.write("""""" ) # we want to support dataset_infos.json for backward compatibility if "dataset_infos.json" in files: with open(dataset_infos_dir / """dataset_infos.json""" , """w""" ) as f: f.write("""{\"default\": {\"dataset_size\": 42}}""" ) __a : List[str] = DatasetInfosDict.from_directory(_lowerCamelCase ) assert dataset_infos assert dataset_infos["default"].dataset_size == 4_2 @pytest.mark.parametrize( """dataset_info""" , [ DatasetInfo(), DatasetInfo( description="""foo""" , features=Features({"""a""": Value("""int32""" )} ) , builder_name="""builder""" , config_name="""config""" , version="""1.0.0""" , splits=[{"""name""": """train"""}] , download_size=4_2 , ), ] , ) def __magic_name__ ( _lowerCamelCase : List[Any] , _lowerCamelCase : DatasetInfo ): __a : str = str(_lowerCamelCase ) dataset_info.write_to_directory(_lowerCamelCase ) __a : List[str] = DatasetInfo.from_directory(_lowerCamelCase ) assert dataset_info == reloaded assert os.path.exists(os.path.join(_lowerCamelCase , """dataset_info.json""" ) ) def __magic_name__ ( ): __a : Dict = DatasetInfo( description="""foo""" , citation="""bar""" , homepage="""https://foo.bar""" , license="""CC0""" , features=Features({"""a""": Value("""int32""" )} ) , post_processed={} , supervised_keys=() , task_templates=[] , builder_name="""builder""" , config_name="""config""" , version="""1.0.0""" , splits=[{"""name""": """train""", """num_examples""": 4_2}] , download_checksums={} , download_size=1_3_3_7 , post_processing_size=4_4_2 , dataset_size=1_2_3_4 , size_in_bytes=1_3_3_7 + 4_4_2 + 1_2_3_4 , ) __a : List[str] = dataset_info._to_yaml_dict() assert sorted(_lowerCamelCase ) == sorted(DatasetInfo._INCLUDED_INFO_IN_YAML ) for key in DatasetInfo._INCLUDED_INFO_IN_YAML: assert key in dataset_info_yaml_dict assert isinstance(dataset_info_yaml_dict[key] , (list, dict, int, str) ) __a : Optional[Any] = yaml.safe_dump(_lowerCamelCase ) __a : List[Any] = yaml.safe_load(_lowerCamelCase ) assert dataset_info_yaml_dict == reloaded def __magic_name__ ( ): __a : Union[str, Any] = DatasetInfo() __a : List[str] = dataset_info._to_yaml_dict() assert dataset_info_yaml_dict == {} @pytest.mark.parametrize( """dataset_infos_dict""" , [ DatasetInfosDict(), DatasetInfosDict({"""default""": DatasetInfo()} ), DatasetInfosDict({"""my_config_name""": DatasetInfo()} ), DatasetInfosDict( { """default""": DatasetInfo( description="""foo""" , features=Features({"""a""": Value("""int32""" )} ) , builder_name="""builder""" , config_name="""config""" , version="""1.0.0""" , splits=[{"""name""": """train"""}] , download_size=4_2 , ) } ), DatasetInfosDict( { """v1""": DatasetInfo(dataset_size=4_2 ), """v2""": DatasetInfo(dataset_size=1_3_3_7 ), } ), ] , ) def __magic_name__ ( _lowerCamelCase : Tuple , _lowerCamelCase : DatasetInfosDict ): __a : List[str] = str(_lowerCamelCase ) dataset_infos_dict.write_to_directory(_lowerCamelCase ) __a : Union[str, Any] = DatasetInfosDict.from_directory(_lowerCamelCase ) # the config_name of the dataset_infos_dict take over the attribute for config_name, dataset_info in dataset_infos_dict.items(): __a : Optional[int] = config_name # the yaml representation doesn't include fields like description or citation # so we just test that we can recover what we can from the yaml __a : str = DatasetInfo._from_yaml_dict(dataset_info._to_yaml_dict() ) assert dataset_infos_dict == reloaded if dataset_infos_dict: assert os.path.exists(os.path.join(_lowerCamelCase , """README.md""" ) )
706
"""simple docstring""" import math from enum import Enum from typing import Optional, Union from torch.optim import Optimizer from torch.optim.lr_scheduler import LambdaLR from .utils import logging lowercase__ = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = "linear" _lowerCAmelCase = "cosine" _lowerCAmelCase = "cosine_with_restarts" _lowerCAmelCase = "polynomial" _lowerCAmelCase = "constant" _lowerCAmelCase = "constant_with_warmup" _lowerCAmelCase = "piecewise_constant" def __magic_name__ ( _lowerCamelCase : Optimizer , _lowerCamelCase : int = -1 ): return LambdaLR(_lowerCamelCase , lambda _lowerCamelCase : 1 , last_epoch=_lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : Optimizer , _lowerCamelCase : int , _lowerCamelCase : int = -1 ): def lr_lambda(_lowerCamelCase : int ): if current_step < num_warmup_steps: return float(_lowerCamelCase ) / float(max(1.0 , _lowerCamelCase ) ) return 1.0 return LambdaLR(_lowerCamelCase , _lowerCamelCase , last_epoch=_lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : Optimizer , _lowerCamelCase : str , _lowerCamelCase : int = -1 ): __a : Optional[int] = {} __a : Any = step_rules.split(""",""" ) for rule_str in rule_list[:-1]: __a , __a : int = rule_str.split(""":""" ) __a : Optional[int] = int(_lowerCamelCase ) __a : str = float(_lowerCamelCase ) __a : int = value __a : Dict = float(rule_list[-1] ) def create_rules_function(_lowerCamelCase : str , _lowerCamelCase : Tuple ): def rule_func(_lowerCamelCase : int ) -> float: __a : Optional[Any] = sorted(rules_dict.keys() ) for i, sorted_step in enumerate(_lowerCamelCase ): if steps < sorted_step: return rules_dict[sorted_steps[i]] return last_lr_multiple return rule_func __a : Optional[int] = create_rules_function(_lowerCamelCase , _lowerCamelCase ) return LambdaLR(_lowerCamelCase , _lowerCamelCase , last_epoch=_lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : Optional[Any] , _lowerCamelCase : Any , _lowerCamelCase : List[str] , _lowerCamelCase : str=-1 ): def lr_lambda(_lowerCamelCase : int ): if current_step < num_warmup_steps: return float(_lowerCamelCase ) / float(max(1 , _lowerCamelCase ) ) return max( 0.0 , float(num_training_steps - current_step ) / float(max(1 , num_training_steps - num_warmup_steps ) ) ) return LambdaLR(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : Optimizer , _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : float = 0.5 , _lowerCamelCase : int = -1 ): def lr_lambda(_lowerCamelCase : Any ): if current_step < num_warmup_steps: return float(_lowerCamelCase ) / float(max(1 , _lowerCamelCase ) ) __a : Dict = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * float(_lowerCamelCase ) * 2.0 * progress )) ) return LambdaLR(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : Optimizer , _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : int = 1 , _lowerCamelCase : int = -1 ): def lr_lambda(_lowerCamelCase : Optional[int] ): if current_step < num_warmup_steps: return float(_lowerCamelCase ) / float(max(1 , _lowerCamelCase ) ) __a : Dict = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) if progress >= 1.0: return 0.0 return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * ((float(_lowerCamelCase ) * progress) % 1.0) )) ) return LambdaLR(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def __magic_name__ ( _lowerCamelCase : Any , _lowerCamelCase : Any , _lowerCamelCase : Optional[Any] , _lowerCamelCase : Optional[Any]=1E-7 , _lowerCamelCase : Optional[int]=1.0 , _lowerCamelCase : Optional[int]=-1 ): __a : Union[str, Any] = optimizer.defaults["""lr"""] if not (lr_init > lr_end): raise ValueError(F'''lr_end ({lr_end}) must be be smaller than initial lr ({lr_init})''' ) def lr_lambda(_lowerCamelCase : int ): if current_step < num_warmup_steps: return float(_lowerCamelCase ) / float(max(1 , _lowerCamelCase ) ) elif current_step > num_training_steps: return lr_end / lr_init # as LambdaLR multiplies by lr_init else: __a : Tuple = lr_init - lr_end __a : int = num_training_steps - num_warmup_steps __a : Optional[int] = 1 - (current_step - num_warmup_steps) / decay_steps __a : List[str] = lr_range * pct_remaining**power + lr_end return decay / lr_init # as LambdaLR multiplies by lr_init return LambdaLR(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) lowercase__ = { SchedulerType.LINEAR: get_linear_schedule_with_warmup, SchedulerType.COSINE: get_cosine_schedule_with_warmup, SchedulerType.COSINE_WITH_RESTARTS: get_cosine_with_hard_restarts_schedule_with_warmup, SchedulerType.POLYNOMIAL: get_polynomial_decay_schedule_with_warmup, SchedulerType.CONSTANT: get_constant_schedule, SchedulerType.CONSTANT_WITH_WARMUP: get_constant_schedule_with_warmup, SchedulerType.PIECEWISE_CONSTANT: get_piecewise_constant_schedule, } def __magic_name__ ( _lowerCamelCase : Union[str, SchedulerType] , _lowerCamelCase : Optimizer , _lowerCamelCase : Optional[str] = None , _lowerCamelCase : Optional[int] = None , _lowerCamelCase : Optional[int] = None , _lowerCamelCase : int = 1 , _lowerCamelCase : float = 1.0 , _lowerCamelCase : int = -1 , ): __a : int = SchedulerType(_lowerCamelCase ) __a : Optional[int] = TYPE_TO_SCHEDULER_FUNCTION[name] if name == SchedulerType.CONSTANT: return schedule_func(_lowerCamelCase , last_epoch=_lowerCamelCase ) if name == SchedulerType.PIECEWISE_CONSTANT: return schedule_func(_lowerCamelCase , step_rules=_lowerCamelCase , last_epoch=_lowerCamelCase ) # All other schedulers require `num_warmup_steps` if num_warmup_steps is None: raise ValueError(F'''{name} requires `num_warmup_steps`, please provide that argument.''' ) if name == SchedulerType.CONSTANT_WITH_WARMUP: return schedule_func(_lowerCamelCase , num_warmup_steps=_lowerCamelCase , last_epoch=_lowerCamelCase ) # All other schedulers require `num_training_steps` if num_training_steps is None: raise ValueError(F'''{name} requires `num_training_steps`, please provide that argument.''' ) if name == SchedulerType.COSINE_WITH_RESTARTS: return schedule_func( _lowerCamelCase , num_warmup_steps=_lowerCamelCase , num_training_steps=_lowerCamelCase , num_cycles=_lowerCamelCase , last_epoch=_lowerCamelCase , ) if name == SchedulerType.POLYNOMIAL: return schedule_func( _lowerCamelCase , num_warmup_steps=_lowerCamelCase , num_training_steps=_lowerCamelCase , power=_lowerCamelCase , last_epoch=_lowerCamelCase , ) return schedule_func( _lowerCamelCase , num_warmup_steps=_lowerCamelCase , num_training_steps=_lowerCamelCase , last_epoch=_lowerCamelCase )
63
0
"""simple docstring""" from __future__ import annotations lowercase__ = "Muhammad Umer Farooq" lowercase__ = "MIT" lowercase__ = "1.0.0" lowercase__ = "Muhammad Umer Farooq" lowercase__ = "contact@muhammadumerfarooq.me" lowercase__ = "Alpha" import re from html.parser import HTMLParser from urllib import parse import requests class SCREAMING_SNAKE_CASE__ ( __snake_case ): def __init__(self , _lowercase ): '''simple docstring''' super().__init__() __a : list[str] = [] __a : Optional[Any] = domain def lowerCAmelCase__(self , _lowercase , _lowercase ): '''simple docstring''' if tag == "a": # Check the list of defined attributes. for name, value in attrs: # If href is defined, and not empty nor # print it. if name == "href" and value != "#" and value != "": # If not already in urls. if value not in self.urls: __a : Any = parse.urljoin(self.domain , _lowercase ) self.urls.append(_lowercase ) def __magic_name__ ( _lowerCamelCase : str ): return ".".join(get_sub_domain_name(_lowerCamelCase ).split(""".""" )[-2:] ) def __magic_name__ ( _lowerCamelCase : str ): return parse.urlparse(_lowerCamelCase ).netloc def __magic_name__ ( _lowerCamelCase : str = "https://github.com" ): __a : Dict = get_domain_name(_lowerCamelCase ) # Initialize the parser __a : Tuple = Parser(_lowerCamelCase ) try: # Open URL __a : List[Any] = requests.get(_lowerCamelCase ) # pass the raw HTML to the parser to get links parser.feed(r.text ) # Get links and loop through __a : Union[str, Any] = set() for link in parser.urls: # open URL. # read = requests.get(link) try: __a : List[str] = requests.get(_lowerCamelCase ) # Get the valid email. __a : Union[str, Any] = re.findall("""[a-zA-Z0-9]+@""" + domain , read.text ) # If not in list then append it. for email in emails: valid_emails.add(_lowerCamelCase ) except ValueError: pass except ValueError: raise SystemExit(1 ) # Finally return a sorted list of email addresses with no duplicates. return sorted(_lowerCamelCase ) if __name__ == "__main__": lowercase__ = emails_from_url("https://github.com") print(f'{len(emails)} emails found:') print("\n".join(sorted(emails)))
707
"""simple docstring""" import importlib import torch import yaml from omegaconf import OmegaConf from taming.models.vqgan import VQModel def __magic_name__ ( _lowerCamelCase : int , _lowerCamelCase : Optional[Any]=False ): __a : Dict = OmegaConf.load(_lowerCamelCase ) if display: print(yaml.dump(OmegaConf.to_container(_lowerCamelCase ) ) ) return config def __magic_name__ ( _lowerCamelCase : str , _lowerCamelCase : Union[str, Any]=None , _lowerCamelCase : int=None ): if conf_path is None: __a : str = """./model_checkpoints/vqgan_only.yaml""" __a : List[Any] = load_config(_lowerCamelCase , display=_lowerCamelCase ) __a : Dict = VQModel(**config.model.params ) if ckpt_path is None: __a : List[Any] = """./model_checkpoints/vqgan_only.pt""" __a : Tuple = torch.load(_lowerCamelCase , map_location=_lowerCamelCase ) if ".ckpt" in ckpt_path: __a : List[str] = sd["""state_dict"""] model.load_state_dict(_lowerCamelCase , strict=_lowerCamelCase ) model.to(_lowerCamelCase ) del sd return model def __magic_name__ ( _lowerCamelCase : Optional[int] , _lowerCamelCase : List[str] ): __a , __a , __a : Tuple = model.encode(_lowerCamelCase ) print(F'''VQGAN --- {model.__class__.__name__}: latent shape: {z.shape[2:]}''' ) __a : Union[str, Any] = model.decode(_lowerCamelCase ) return xrec def __magic_name__ ( _lowerCamelCase : Optional[int] , _lowerCamelCase : Union[str, Any]=False ): __a , __a : Optional[Any] = string.rsplit(""".""" , 1 ) if reload: __a : Optional[Any] = importlib.import_module(_lowerCamelCase ) importlib.reload(_lowerCamelCase ) return getattr(importlib.import_module(_lowerCamelCase , package=_lowerCamelCase ) , cls ) def __magic_name__ ( _lowerCamelCase : Any ): if "target" not in config: raise KeyError("""Expected key `target` to instantiate.""" ) return get_obj_from_str(config["""target"""] )(**config.get("""params""" , {} ) ) def __magic_name__ ( _lowerCamelCase : Any , _lowerCamelCase : Dict , _lowerCamelCase : int=True , _lowerCamelCase : int=True ): __a : Union[str, Any] = instantiate_from_config(_lowerCamelCase ) if sd is not None: model.load_state_dict(_lowerCamelCase ) if gpu: model.cuda() if eval_mode: model.eval() return {"model": model} def __magic_name__ ( _lowerCamelCase : Any , _lowerCamelCase : Union[str, Any] , _lowerCamelCase : str , _lowerCamelCase : int ): # load the specified checkpoint if ckpt: __a : List[str] = torch.load(_lowerCamelCase , map_location="""cpu""" ) __a : Any = pl_sd["""global_step"""] print(F'''loaded model from global step {global_step}.''' ) else: __a : List[Any] = {"""state_dict""": None} __a : Any = None __a : Union[str, Any] = load_model_from_config(config.model , pl_sd["""state_dict"""] , gpu=_lowerCamelCase , eval_mode=_lowerCamelCase )["""model"""] return model, global_step
63
0
"""simple docstring""" from __future__ import annotations lowercase__ = [True] * 1000001 lowercase__ = 2 while i * i <= 1000000: if seive[i]: for j in range(i * i, 1000001, i): lowercase__ = False i += 1 def __magic_name__ ( _lowerCamelCase : int ): return seive[n] def __magic_name__ ( _lowerCamelCase : int ): return any(digit in """02468""" for digit in str(_lowerCamelCase ) ) def __magic_name__ ( _lowerCamelCase : int = 1_0_0_0_0_0_0 ): __a : str = [2] # result already includes the number 2. for num in range(3 , limit + 1 , 2 ): if is_prime(_lowerCamelCase ) and not contains_an_even_digit(_lowerCamelCase ): __a : str = str(_lowerCamelCase ) __a : str = [int(str_num[j:] + str_num[:j] ) for j in range(len(_lowerCamelCase ) )] if all(is_prime(_lowerCamelCase ) for i in list_nums ): result.append(_lowerCamelCase ) return result def __magic_name__ ( ): return len(find_circular_primes() ) if __name__ == "__main__": print(f'{len(find_circular_primes()) = }')
708
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) lowercase__ = { "configuration_llama": ["LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP", "LlamaConfig"], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = ["LlamaTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = ["LlamaTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ "LlamaForCausalLM", "LlamaModel", "LlamaPreTrainedModel", "LlamaForSequenceClassification", ] if TYPE_CHECKING: from .configuration_llama import LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP, LlamaConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama import LlamaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama_fast import LlamaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_llama import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaPreTrainedModel else: import sys lowercase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
63
0
"""simple docstring""" def __magic_name__ ( _lowerCamelCase : int , _lowerCamelCase : float , _lowerCamelCase : float ): return round(float(moles / volume ) * nfactor ) def __magic_name__ ( _lowerCamelCase : float , _lowerCamelCase : float , _lowerCamelCase : float ): return round(float((moles * 0.08_21 * temperature) / (volume) ) ) def __magic_name__ ( _lowerCamelCase : float , _lowerCamelCase : float , _lowerCamelCase : float ): return round(float((moles * 0.08_21 * temperature) / (pressure) ) ) def __magic_name__ ( _lowerCamelCase : float , _lowerCamelCase : float , _lowerCamelCase : float ): return round(float((pressure * volume) / (0.08_21 * moles) ) ) if __name__ == "__main__": import doctest doctest.testmod()
709
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { "microsoft/unispeech-large-1500h-cv": ( "https://huggingface.co/microsoft/unispeech-large-1500h-cv/resolve/main/config.json" ), # See all UniSpeech models at https://huggingface.co/models?filter=unispeech } class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = "unispeech" def __init__(self , _lowercase=32 , _lowercase=768 , _lowercase=12 , _lowercase=12 , _lowercase=3072 , _lowercase="gelu" , _lowercase=0.1 , _lowercase=0.1 , _lowercase=0.1 , _lowercase=0.0 , _lowercase=0.0 , _lowercase=0.1 , _lowercase=0.1 , _lowercase=0.02 , _lowercase=1e-5 , _lowercase="group" , _lowercase="gelu" , _lowercase=(512, 512, 512, 512, 512, 512, 512) , _lowercase=(5, 2, 2, 2, 2, 2, 2) , _lowercase=(10, 3, 3, 3, 3, 2, 2) , _lowercase=False , _lowercase=128 , _lowercase=16 , _lowercase=False , _lowercase=True , _lowercase=0.05 , _lowercase=10 , _lowercase=2 , _lowercase=0.0 , _lowercase=10 , _lowercase=0 , _lowercase=320 , _lowercase=2 , _lowercase=0.1 , _lowercase=100 , _lowercase=256 , _lowercase=256 , _lowercase=0.1 , _lowercase="mean" , _lowercase=False , _lowercase=False , _lowercase=256 , _lowercase=80 , _lowercase=0 , _lowercase=1 , _lowercase=2 , _lowercase=0.5 , **_lowercase , ): '''simple docstring''' super().__init__(**_lowercase , pad_token_id=_lowercase , bos_token_id=_lowercase , eos_token_id=_lowercase ) __a : Union[str, Any] = hidden_size __a : Any = feat_extract_norm __a : Union[str, Any] = feat_extract_activation __a : Tuple = list(_lowercase ) __a : Dict = list(_lowercase ) __a : List[Any] = list(_lowercase ) __a : List[Any] = conv_bias __a : Optional[Any] = num_conv_pos_embeddings __a : Union[str, Any] = num_conv_pos_embedding_groups __a : Dict = len(self.conv_dim ) __a : Dict = num_hidden_layers __a : Union[str, Any] = intermediate_size __a : List[str] = hidden_act __a : int = num_attention_heads __a : int = hidden_dropout __a : Any = attention_dropout __a : List[Any] = activation_dropout __a : List[Any] = feat_proj_dropout __a : Union[str, Any] = final_dropout __a : str = layerdrop __a : Dict = layer_norm_eps __a : Dict = initializer_range __a : Union[str, Any] = num_ctc_classes __a : List[Any] = vocab_size __a : Any = do_stable_layer_norm __a : List[str] = use_weighted_layer_sum __a : List[str] = classifier_proj_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==""" """ `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =""" F''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' F''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __a : Dict = apply_spec_augment __a : Union[str, Any] = mask_time_prob __a : List[str] = mask_time_length __a : Dict = mask_time_min_masks __a : List[Any] = mask_feature_prob __a : Tuple = mask_feature_length __a : int = mask_feature_min_masks # parameters for pretraining with codevector quantized representations __a : List[Any] = num_codevectors_per_group __a : Union[str, Any] = num_codevector_groups __a : List[Any] = contrastive_logits_temperature __a : Any = feat_quantizer_dropout __a : Optional[int] = num_negatives __a : List[str] = codevector_dim __a : List[Any] = proj_codevector_dim __a : Tuple = diversity_loss_weight # ctc loss __a : Any = ctc_loss_reduction __a : List[str] = ctc_zero_infinity # pretraining loss __a : Tuple = replace_prob @property def lowerCAmelCase__(self ): '''simple docstring''' return functools.reduce(operator.mul , self.conv_stride , 1 )
63
0
"""simple docstring""" from itertools import product from cva import COLOR_BGR2GRAY, cvtColor, imread, imshow, waitKey from numpy import dot, exp, mgrid, pi, ravel, square, uinta, zeros def __magic_name__ ( _lowerCamelCase : List[Any] , _lowerCamelCase : List[str] ): __a : Dict = k_size // 2 __a : str = mgrid[0 - center : k_size - center, 0 - center : k_size - center] __a : Optional[Any] = 1 / (2 * pi * sigma) * exp(-(square(_lowerCamelCase ) + square(_lowerCamelCase )) / (2 * square(_lowerCamelCase )) ) return g def __magic_name__ ( _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : Union[str, Any] ): __a : Optional[Any] = image.shape[0], image.shape[1] # dst image height and width __a : int = height - k_size + 1 __a : Dict = width - k_size + 1 # im2col, turn the k_size*k_size pixels into a row and np.vstack all rows __a : Any = zeros((dst_height * dst_width, k_size * k_size) ) __a : List[str] = 0 for i, j in product(range(_lowerCamelCase ) , range(_lowerCamelCase ) ): __a : Optional[int] = ravel(image[i : i + k_size, j : j + k_size] ) __a : Union[str, Any] = window row += 1 # turn the kernel into shape(k*k, 1) __a : List[Any] = gen_gaussian_kernel(_lowerCamelCase , _lowerCamelCase ) __a : List[str] = ravel(_lowerCamelCase ) # reshape and get the dst image __a : Optional[Any] = dot(_lowerCamelCase , _lowerCamelCase ).reshape(_lowerCamelCase , _lowerCamelCase ).astype(_lowerCamelCase ) return dst if __name__ == "__main__": # read original image lowercase__ = imread(R"../image_data/lena.jpg") # turn image in gray scale value lowercase__ = cvtColor(img, COLOR_BGR2GRAY) # get values with two different mask size lowercase__ = gaussian_filter(gray, 3, sigma=1) lowercase__ = gaussian_filter(gray, 5, sigma=0.8) # show result images imshow("gaussian filter with 3x3 mask", gaussianaxa) imshow("gaussian filter with 5x5 mask", gaussianaxa) waitKey()
710
"""simple docstring""" import inspect import unittest from typing import List import numpy as np from transformers import EfficientFormerConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, ) from transformers.models.efficientformer.modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_vision_available(): from PIL import Image from transformers import EfficientFormerImageProcessor class SCREAMING_SNAKE_CASE__ : def __init__(self , _lowercase , _lowercase = 13 , _lowercase = 64 , _lowercase = 2 , _lowercase = 3 , _lowercase = 3 , _lowercase = True , _lowercase = True , _lowercase = 128 , _lowercase=[16, 32, 64, 128] , _lowercase = 7 , _lowercase = 4 , _lowercase = 37 , _lowercase = "gelu" , _lowercase = 0.1 , _lowercase = 0.1 , _lowercase = 10 , _lowercase = 0.02 , _lowercase = 2 , _lowercase = 1 , _lowercase = 128 , _lowercase = [2, 2, 2, 2] , _lowercase = 2 , _lowercase = 2 , ): '''simple docstring''' __a : str = parent __a : List[Any] = batch_size __a : int = image_size __a : Tuple = patch_size __a : str = num_channels __a : Union[str, Any] = is_training __a : List[Any] = use_labels __a : int = hidden_size __a : Optional[Any] = num_hidden_layers __a : List[Any] = num_attention_heads __a : Dict = intermediate_size __a : str = hidden_act __a : Dict = hidden_dropout_prob __a : str = attention_probs_dropout_prob __a : Optional[int] = type_sequence_label_size __a : Dict = initializer_range __a : Dict = encoder_stride __a : int = num_attention_outputs __a : List[Any] = embed_dim __a : Optional[Any] = embed_dim + 1 __a : Optional[Any] = resolution __a : Optional[Any] = depths __a : Union[str, Any] = hidden_sizes __a : List[str] = dim __a : Any = mlp_expansion_ratio def lowerCAmelCase__(self ): '''simple docstring''' __a : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __a : str = None if self.use_labels: __a : List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __a : List[str] = self.get_config() return config, pixel_values, labels def lowerCAmelCase__(self ): '''simple docstring''' return EfficientFormerConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowercase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , resolution=self.resolution , depths=self.depths , hidden_sizes=self.hidden_sizes , dim=self.dim , mlp_expansion_ratio=self.mlp_expansion_ratio , ) def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase ): '''simple docstring''' __a : Optional[Any] = TFEfficientFormerModel(config=_lowercase ) __a : List[Any] = model(_lowercase , training=_lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase ): '''simple docstring''' __a : Optional[Any] = self.type_sequence_label_size __a : Any = TFEfficientFormerForImageClassification(_lowercase ) __a : Union[str, Any] = model(_lowercase , labels=_lowercase , training=_lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __a : Optional[Any] = 1 __a : int = TFEfficientFormerForImageClassification(_lowercase ) __a : List[str] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __a : str = model(_lowercase , labels=_lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Any = self.prepare_config_and_inputs() __a , __a , __a : Tuple = config_and_inputs __a : Tuple = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class SCREAMING_SNAKE_CASE__ ( __snake_case , __snake_case , unittest.TestCase ): _lowerCAmelCase = ( ( TFEfficientFormerModel, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerForImageClassification, ) if is_tf_available() else () ) _lowerCAmelCase = ( { "feature-extraction": TFEfficientFormerModel, "image-classification": ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, ), } if is_tf_available() else {} ) _lowerCAmelCase = False _lowerCAmelCase = False _lowerCAmelCase = False _lowerCAmelCase = False _lowerCAmelCase = False def lowerCAmelCase__(self ): '''simple docstring''' __a : Tuple = TFEfficientFormerModelTester(self ) __a : Any = ConfigTester( self , config_class=_lowercase , has_text_modality=_lowercase , hidden_size=37 ) def lowerCAmelCase__(self ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="""EfficientFormer does not use inputs_embeds""" ) def lowerCAmelCase__(self ): '''simple docstring''' pass @unittest.skip(reason="""EfficientFormer does not support input and output embeddings""" ) def lowerCAmelCase__(self ): '''simple docstring''' pass def lowerCAmelCase__(self ): '''simple docstring''' __a , __a : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a : Dict = model_class(_lowercase ) __a : Optional[Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __a : Optional[Any] = [*signature.parameters.keys()] __a : Union[str, Any] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' def check_hidden_states_output(_lowercase , _lowercase , _lowercase ): __a : Tuple = model_class(_lowercase ) __a : int = model(**self._prepare_for_class(_lowercase , _lowercase ) , training=_lowercase ) __a : Tuple = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __a : str = getattr( self.model_tester , """expected_num_hidden_layers""" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(_lowercase ) , _lowercase ) if hasattr(self.model_tester , """encoder_seq_length""" ): __a : Any = self.model_tester.encoder_seq_length if hasattr(self.model_tester , """chunk_length""" ) and self.model_tester.chunk_length > 1: __a : int = seq_length * self.model_tester.chunk_length else: __a : Any = self.model_tester.seq_length self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) if config.is_encoder_decoder: __a : Optional[int] = outputs.decoder_hidden_states self.asseretIsInstance(_lowercase , (list, tuple) ) self.assertEqual(len(_lowercase ) , _lowercase ) __a : Any = getattr(self.model_tester , """seq_length""" , _lowercase ) __a : List[Any] = getattr(self.model_tester , """decoder_seq_length""" , _lowercase ) self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [decoder_seq_length, self.model_tester.hidden_size] , ) __a , __a : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a : Dict = True check_hidden_states_output(_lowercase , _lowercase , _lowercase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __a : int = True check_hidden_states_output(_lowercase , _lowercase , _lowercase ) def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase=False ): '''simple docstring''' __a : Any = super()._prepare_for_class(_lowercase , _lowercase , return_labels=_lowercase ) if return_labels: if model_class.__name__ == "TFEfficientFormerForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowercase ) @unittest.skip(reason="""EfficientFormer does not implement masked image modeling yet""" ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowercase ) @slow def lowerCAmelCase__(self ): '''simple docstring''' for model_name in TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a : Union[str, Any] = TFEfficientFormerModel.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' __a , __a : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() __a : int = True __a : Optional[int] = getattr(self.model_tester , """seq_length""" , _lowercase ) __a : Dict = getattr(self.model_tester , """encoder_seq_length""" , _lowercase ) __a : Dict = getattr(self.model_tester , """key_length""" , _lowercase ) __a : int = getattr(self.model_tester , """chunk_length""" , _lowercase ) if chunk_length is not None and hasattr(self.model_tester , """num_hashes""" ): __a : List[str] = encoder_seq_length * self.model_tester.num_hashes for model_class in self.all_model_classes: __a : List[Any] = True __a : Tuple = False __a : List[Any] = True __a : int = model_class(_lowercase ) __a : List[Any] = model(**self._prepare_for_class(_lowercase , _lowercase ) , training=_lowercase ) __a : Dict = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(_lowercase ) , self.model_tester.num_attention_outputs ) # check that output_attentions also work using config del inputs_dict["output_attentions"] __a : Optional[Any] = True __a : List[str] = model_class(_lowercase ) __a : Dict = model(**self._prepare_for_class(_lowercase , _lowercase ) , training=_lowercase ) __a : int = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(_lowercase ) , self.model_tester.num_attention_outputs ) if chunk_length is not None: self.assertListEqual( list(attentions[0].shape[-4:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, chunk_length, encoder_key_length] , ) else: self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, encoder_key_length] , ) def lowerCAmelCase__(self ): '''simple docstring''' __a , __a : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # Prepare our model __a : Dict = model_class(_lowercase ) # These are maximally general inputs for the model, with multiple None dimensions # Hopefully this will catch any conditionals that fail for flexible shapes __a : Optional[Any] = { key: tf.keras.Input(shape=val.shape[1:] , dtype=val.dtype , name=_lowercase ) for key, val in model.input_signature.items() if key in model.dummy_inputs } __a : Optional[Any] = model(_lowercase ) self.assertTrue(outputs_dict is not None ) def __magic_name__ ( ): __a : int = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @cached_property def lowerCAmelCase__(self ): '''simple docstring''' return ( EfficientFormerImageProcessor.from_pretrained("""snap-research/efficientformer-l1-300""" ) if is_vision_available() else None ) @slow def lowerCAmelCase__(self ): '''simple docstring''' __a : str = TFEfficientFormerForImageClassification.from_pretrained("""snap-research/efficientformer-l1-300""" ) __a : Optional[Any] = self.default_image_processor __a : List[str] = prepare_img() __a : int = image_processor(images=_lowercase , return_tensors="""tf""" ) # forward pass __a : Optional[Any] = model(**_lowercase , training=_lowercase ) # verify the logits __a : str = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowercase ) __a : Dict = tf.constant([-0.0555, 0.4825, -0.0852] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , _lowercase , atol=1e-4 ) ) @slow def lowerCAmelCase__(self ): '''simple docstring''' __a : Any = TFEfficientFormerForImageClassificationWithTeacher.from_pretrained( """snap-research/efficientformer-l1-300""" ) __a : Any = self.default_image_processor __a : str = prepare_img() __a : str = image_processor(images=_lowercase , return_tensors="""tf""" ) # forward pass __a : List[Any] = model(**_lowercase , training=_lowercase ) # verify the logits __a : int = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowercase ) __a : List[str] = tf.constant([-0.1312, 0.4353, -1.0499] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , _lowercase , atol=1e-4 ) )
63
0
"""simple docstring""" import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, LDMTextToImagePipeline, UNetaDConditionModel from diffusers.utils.testing_utils import ( enable_full_determinism, load_numpy, nightly, require_torch_gpu, slow, torch_device, ) from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class SCREAMING_SNAKE_CASE__ ( __snake_case , unittest.TestCase ): _lowerCAmelCase = LDMTextToImagePipeline _lowerCAmelCase = TEXT_TO_IMAGE_PARAMS - { "negative_prompt", "negative_prompt_embeds", "cross_attention_kwargs", "prompt_embeds", } _lowerCAmelCase = PipelineTesterMixin.required_optional_params - { "num_images_per_prompt", "callback", "callback_steps", } _lowerCAmelCase = TEXT_TO_IMAGE_BATCH_PARAMS _lowerCAmelCase = False def lowerCAmelCase__(self ): '''simple docstring''' torch.manual_seed(0 ) __a : Tuple = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) __a : int = DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , clip_sample=_lowercase , set_alpha_to_one=_lowercase , ) torch.manual_seed(0 ) __a : List[str] = AutoencoderKL( block_out_channels=(32, 64) , in_channels=3 , out_channels=3 , down_block_types=("""DownEncoderBlock2D""", """DownEncoderBlock2D""") , up_block_types=("""UpDecoderBlock2D""", """UpDecoderBlock2D""") , latent_channels=4 , ) torch.manual_seed(0 ) __a : Union[str, Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) __a : Optional[Any] = CLIPTextModel(_lowercase ) __a : Tuple = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) __a : List[str] = { """unet""": unet, """scheduler""": scheduler, """vqvae""": vae, """bert""": text_encoder, """tokenizer""": tokenizer, } return components def lowerCAmelCase__(self , _lowercase , _lowercase=0 ): '''simple docstring''' if str(_lowercase ).startswith("""mps""" ): __a : Union[str, Any] = torch.manual_seed(_lowercase ) else: __a : Any = torch.Generator(device=_lowercase ).manual_seed(_lowercase ) __a : Dict = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", } return inputs def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[int] = """cpu""" # ensure determinism for the device-dependent torch.Generator __a : List[str] = self.get_dummy_components() __a : Union[str, Any] = LDMTextToImagePipeline(**_lowercase ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) __a : Dict = self.get_dummy_inputs(_lowercase ) __a : List[Any] = pipe(**_lowercase ).images __a : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 16, 16, 3) __a : Union[str, Any] = np.array([0.6101, 0.6156, 0.5622, 0.4895, 0.6661, 0.3804, 0.5748, 0.6136, 0.5014] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 @slow @require_torch_gpu class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def lowerCAmelCase__(self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCAmelCase__(self , _lowercase , _lowercase=torch.floataa , _lowercase=0 ): '''simple docstring''' __a : Union[str, Any] = torch.manual_seed(_lowercase ) __a : int = np.random.RandomState(_lowercase ).standard_normal((1, 4, 32, 32) ) __a : Optional[int] = torch.from_numpy(_lowercase ).to(device=_lowercase , dtype=_lowercase ) __a : Any = { """prompt""": """A painting of a squirrel eating a burger""", """latents""": latents, """generator""": generator, """num_inference_steps""": 3, """guidance_scale""": 6.0, """output_type""": """numpy""", } return inputs def lowerCAmelCase__(self ): '''simple docstring''' __a : Union[str, Any] = LDMTextToImagePipeline.from_pretrained("""CompVis/ldm-text2im-large-256""" ).to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) __a : Optional[Any] = self.get_inputs(_lowercase ) __a : Any = pipe(**_lowercase ).images __a : Optional[int] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 256, 256, 3) __a : Union[str, Any] = np.array([0.5_1825, 0.5_2850, 0.5_2543, 0.5_4258, 0.5_2304, 0.5_2569, 0.5_4363, 0.5_5276, 0.5_6878] ) __a : int = np.abs(expected_slice - image_slice ).max() assert max_diff < 1e-3 @nightly @require_torch_gpu class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def lowerCAmelCase__(self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCAmelCase__(self , _lowercase , _lowercase=torch.floataa , _lowercase=0 ): '''simple docstring''' __a : List[str] = torch.manual_seed(_lowercase ) __a : str = np.random.RandomState(_lowercase ).standard_normal((1, 4, 32, 32) ) __a : List[Any] = torch.from_numpy(_lowercase ).to(device=_lowercase , dtype=_lowercase ) __a : int = { """prompt""": """A painting of a squirrel eating a burger""", """latents""": latents, """generator""": generator, """num_inference_steps""": 50, """guidance_scale""": 6.0, """output_type""": """numpy""", } return inputs def lowerCAmelCase__(self ): '''simple docstring''' __a : List[Any] = LDMTextToImagePipeline.from_pretrained("""CompVis/ldm-text2im-large-256""" ).to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) __a : Tuple = self.get_inputs(_lowercase ) __a : str = pipe(**_lowercase ).images[0] __a : Any = load_numpy( """https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/ldm_text2img/ldm_large_256_ddim.npy""" ) __a : List[str] = np.abs(expected_image - image ).max() assert max_diff < 1e-3
711
"""simple docstring""" from typing import Callable, Dict, Optional, Tuple import torch from torch import nn from torch.distributions import ( AffineTransform, Distribution, Independent, NegativeBinomial, Normal, StudentT, TransformedDistribution, ) class SCREAMING_SNAKE_CASE__ ( __snake_case ): def __init__(self , _lowercase , _lowercase=None , _lowercase=None , _lowercase=0 ): '''simple docstring''' __a : Any = 1.0 if scale is None else scale __a : str = 0.0 if loc is None else loc super().__init__(_lowercase , [AffineTransform(loc=self.loc , scale=self.scale , event_dim=_lowercase )] ) @property def lowerCAmelCase__(self ): '''simple docstring''' return self.base_dist.mean * self.scale + self.loc @property def lowerCAmelCase__(self ): '''simple docstring''' return self.base_dist.variance * self.scale**2 @property def lowerCAmelCase__(self ): '''simple docstring''' return self.variance.sqrt() class SCREAMING_SNAKE_CASE__ ( nn.Module ): def __init__(self , _lowercase , _lowercase , _lowercase , **_lowercase ): '''simple docstring''' super().__init__(**_lowercase ) __a : str = args_dim __a : List[Any] = nn.ModuleList([nn.Linear(_lowercase , _lowercase ) for dim in args_dim.values()] ) __a : Dict = domain_map def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' __a : List[Any] = [proj(_lowercase ) for proj in self.proj] return self.domain_map(*_lowercase ) class SCREAMING_SNAKE_CASE__ ( nn.Module ): def __init__(self , _lowercase ): '''simple docstring''' super().__init__() __a : Optional[int] = function def lowerCAmelCase__(self , _lowercase , *_lowercase ): '''simple docstring''' return self.function(_lowercase , *_lowercase ) class SCREAMING_SNAKE_CASE__ : _lowerCAmelCase = 42 _lowerCAmelCase = 42 _lowerCAmelCase = 42 def __init__(self , _lowercase = 1 ): '''simple docstring''' __a : Optional[int] = dim __a : str = {k: dim * self.args_dim[k] for k in self.args_dim} def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' if self.dim == 1: return self.distribution_class(*_lowercase ) else: return Independent(self.distribution_class(*_lowercase ) , 1 ) def lowerCAmelCase__(self , _lowercase , _lowercase = None , _lowercase = None , ): '''simple docstring''' __a : Tuple = self._base_distribution(_lowercase ) if loc is None and scale is None: return distr else: return AffineTransformed(_lowercase , loc=_lowercase , scale=_lowercase , event_dim=self.event_dim ) @property def lowerCAmelCase__(self ): '''simple docstring''' return () if self.dim == 1 else (self.dim,) @property def lowerCAmelCase__(self ): '''simple docstring''' return len(self.event_shape ) @property def lowerCAmelCase__(self ): '''simple docstring''' return 0.0 def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' return ParameterProjection( in_features=_lowercase , args_dim=self.args_dim , domain_map=LambdaLayer(self.domain_map ) , ) def lowerCAmelCase__(self , *_lowercase ): '''simple docstring''' raise NotImplementedError() @staticmethod def lowerCAmelCase__(_lowercase ): '''simple docstring''' return (x + torch.sqrt(torch.square(_lowercase ) + 4.0 )) / 2.0 class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = {"df": 1, "loc": 1, "scale": 1} _lowerCAmelCase = StudentT @classmethod def lowerCAmelCase__(cls , _lowercase , _lowercase , _lowercase ): '''simple docstring''' __a : int = cls.squareplus(_lowercase ).clamp_min(torch.finfo(scale.dtype ).eps ) __a : Optional[Any] = 2.0 + cls.squareplus(_lowercase ) return df.squeeze(-1 ), loc.squeeze(-1 ), scale.squeeze(-1 ) class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = {"loc": 1, "scale": 1} _lowerCAmelCase = Normal @classmethod def lowerCAmelCase__(cls , _lowercase , _lowercase ): '''simple docstring''' __a : str = cls.squareplus(_lowercase ).clamp_min(torch.finfo(scale.dtype ).eps ) return loc.squeeze(-1 ), scale.squeeze(-1 ) class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = {"total_count": 1, "logits": 1} _lowerCAmelCase = NegativeBinomial @classmethod def lowerCAmelCase__(cls , _lowercase , _lowercase ): '''simple docstring''' __a : Union[str, Any] = cls.squareplus(_lowercase ) return total_count.squeeze(-1 ), logits.squeeze(-1 ) def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' __a , __a : Optional[Any] = distr_args if self.dim == 1: return self.distribution_class(total_count=_lowercase , logits=_lowercase ) else: return Independent(self.distribution_class(total_count=_lowercase , logits=_lowercase ) , 1 ) def lowerCAmelCase__(self , _lowercase , _lowercase = None , _lowercase = None ): '''simple docstring''' __a , __a : List[Any] = distr_args if scale is not None: # See scaling property of Gamma. logits += scale.log() return self._base_distribution((total_count, logits) )
63
0
"""simple docstring""" import json import os from functools import lru_cache from typing import List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = {"vocab_file": "vocab.json", "merges_file": "merges.txt"} # See all BART models at https://huggingface.co/models?filter=bart lowercase__ = { "vocab_file": { "facebook/bart-base": "https://huggingface.co/facebook/bart-base/resolve/main/vocab.json", "facebook/bart-large": "https://huggingface.co/facebook/bart-large/resolve/main/vocab.json", "facebook/bart-large-mnli": "https://huggingface.co/facebook/bart-large-mnli/resolve/main/vocab.json", "facebook/bart-large-cnn": "https://huggingface.co/facebook/bart-large-cnn/resolve/main/vocab.json", "facebook/bart-large-xsum": "https://huggingface.co/facebook/bart-large-xsum/resolve/main/vocab.json", "yjernite/bart_eli5": "https://huggingface.co/yjernite/bart_eli5/resolve/main/vocab.json", }, "merges_file": { "facebook/bart-base": "https://huggingface.co/facebook/bart-base/resolve/main/merges.txt", "facebook/bart-large": "https://huggingface.co/facebook/bart-large/resolve/main/merges.txt", "facebook/bart-large-mnli": "https://huggingface.co/facebook/bart-large-mnli/resolve/main/merges.txt", "facebook/bart-large-cnn": "https://huggingface.co/facebook/bart-large-cnn/resolve/main/merges.txt", "facebook/bart-large-xsum": "https://huggingface.co/facebook/bart-large-xsum/resolve/main/merges.txt", "yjernite/bart_eli5": "https://huggingface.co/yjernite/bart_eli5/resolve/main/merges.txt", }, } lowercase__ = { "facebook/bart-base": 1024, "facebook/bart-large": 1024, "facebook/bart-large-mnli": 1024, "facebook/bart-large-cnn": 1024, "facebook/bart-large-xsum": 1024, "yjernite/bart_eli5": 1024, } @lru_cache() def __magic_name__ ( ): __a : Optional[int] = ( list(range(ord("""!""" ) , ord("""~""" ) + 1 ) ) + list(range(ord("""¡""" ) , ord("""¬""" ) + 1 ) ) + list(range(ord("""®""" ) , ord("""ÿ""" ) + 1 ) ) ) __a : List[str] = bs[:] __a : Optional[int] = 0 for b in range(2**8 ): if b not in bs: bs.append(_lowerCamelCase ) cs.append(2**8 + n ) n += 1 __a : Tuple = [chr(_lowerCamelCase ) for n in cs] return dict(zip(_lowerCamelCase , _lowerCamelCase ) ) def __magic_name__ ( _lowerCamelCase : Dict ): __a : Tuple = set() __a : Union[str, Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __a : Union[str, Any] = char return pairs class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = VOCAB_FILES_NAMES _lowerCAmelCase = PRETRAINED_VOCAB_FILES_MAP _lowerCAmelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCAmelCase = ["input_ids", "attention_mask"] def __init__(self , _lowercase , _lowercase , _lowercase="replace" , _lowercase="<s>" , _lowercase="</s>" , _lowercase="</s>" , _lowercase="<s>" , _lowercase="<unk>" , _lowercase="<pad>" , _lowercase="<mask>" , _lowercase=False , **_lowercase , ): '''simple docstring''' __a : Optional[Any] = AddedToken(_lowercase , lstrip=_lowercase , rstrip=_lowercase ) if isinstance(_lowercase , _lowercase ) else bos_token __a : str = AddedToken(_lowercase , lstrip=_lowercase , rstrip=_lowercase ) if isinstance(_lowercase , _lowercase ) else eos_token __a : str = AddedToken(_lowercase , lstrip=_lowercase , rstrip=_lowercase ) if isinstance(_lowercase , _lowercase ) else sep_token __a : str = AddedToken(_lowercase , lstrip=_lowercase , rstrip=_lowercase ) if isinstance(_lowercase , _lowercase ) else cls_token __a : int = AddedToken(_lowercase , lstrip=_lowercase , rstrip=_lowercase ) if isinstance(_lowercase , _lowercase ) else unk_token __a : Any = AddedToken(_lowercase , lstrip=_lowercase , rstrip=_lowercase ) if isinstance(_lowercase , _lowercase ) else pad_token # Mask token behave like a normal word, i.e. include the space before it __a : str = AddedToken(_lowercase , lstrip=_lowercase , rstrip=_lowercase ) if isinstance(_lowercase , _lowercase ) else mask_token super().__init__( errors=_lowercase , bos_token=_lowercase , eos_token=_lowercase , unk_token=_lowercase , sep_token=_lowercase , cls_token=_lowercase , pad_token=_lowercase , mask_token=_lowercase , add_prefix_space=_lowercase , **_lowercase , ) with open(_lowercase , encoding="""utf-8""" ) as vocab_handle: __a : Optional[int] = json.load(_lowercase ) __a : List[str] = {v: k for k, v in self.encoder.items()} __a : Optional[int] = errors # how to handle errors in decoding __a : int = bytes_to_unicode() __a : str = {v: k for k, v in self.byte_encoder.items()} with open(_lowercase , encoding="""utf-8""" ) as merges_handle: __a : int = merges_handle.read().split("""\n""" )[1:-1] __a : Dict = [tuple(merge.split() ) for merge in bpe_merges] __a : Dict = dict(zip(_lowercase , range(len(_lowercase ) ) ) ) __a : Dict = {} __a : List[str] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions __a : Union[str, Any] = re.compile(r"""'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+""" ) @property def lowerCAmelCase__(self ): '''simple docstring''' return len(self.encoder ) def lowerCAmelCase__(self ): '''simple docstring''' return dict(self.encoder , **self.added_tokens_encoder ) def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' if token in self.cache: return self.cache[token] __a : int = tuple(_lowercase ) __a : str = get_pairs(_lowercase ) if not pairs: return token while True: __a : Any = min(_lowercase , key=lambda _lowercase : self.bpe_ranks.get(_lowercase , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break __a : Tuple = bigram __a : int = [] __a : Union[str, Any] = 0 while i < len(_lowercase ): try: __a : Union[str, Any] = word.index(_lowercase , _lowercase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) __a : List[str] = j if word[i] == first and i < len(_lowercase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __a : Any = tuple(_lowercase ) __a : List[Any] = new_word if len(_lowercase ) == 1: break else: __a : Optional[int] = get_pairs(_lowercase ) __a : List[Any] = """ """.join(_lowercase ) __a : Union[str, Any] = word return word def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' __a : str = [] for token in re.findall(self.pat , _lowercase ): __a : Union[str, Any] = """""".join( self.byte_encoder[b] for b in token.encode("""utf-8""" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(_lowercase ).split(""" """ ) ) return bpe_tokens def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' return self.encoder.get(_lowercase , self.encoder.get(self.unk_token ) ) def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' return self.decoder.get(_lowercase ) def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' __a : Optional[int] = """""".join(_lowercase ) __a : Any = bytearray([self.byte_decoder[c] for c in text] ).decode("""utf-8""" , errors=self.errors ) return text def lowerCAmelCase__(self , _lowercase , _lowercase = None ): '''simple docstring''' if not os.path.isdir(_lowercase ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return __a : Optional[Any] = os.path.join( _lowercase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) __a : Union[str, Any] = os.path.join( _lowercase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(_lowercase , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_lowercase , ensure_ascii=_lowercase ) + """\n""" ) __a : List[Any] = 0 with open(_lowercase , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _lowercase : kv[1] ): if index != token_index: logger.warning( F'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' """ Please check that the tokenizer is not corrupted!""" ) __a : Tuple = token_index writer.write(""" """.join(_lowercase ) + """\n""" ) index += 1 return vocab_file, merge_file def lowerCAmelCase__(self , _lowercase , _lowercase = None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] __a : Optional[Any] = [self.cls_token_id] __a : Tuple = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def lowerCAmelCase__(self , _lowercase , _lowercase = None , _lowercase = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_lowercase , token_ids_a=_lowercase , already_has_special_tokens=_lowercase ) if token_ids_a is None: return [1] + ([0] * len(_lowercase )) + [1] return [1] + ([0] * len(_lowercase )) + [1, 1] + ([0] * len(_lowercase )) + [1] def lowerCAmelCase__(self , _lowercase , _lowercase = None ): '''simple docstring''' __a : Union[str, Any] = [self.sep_token_id] __a : Dict = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowerCAmelCase__(self , _lowercase , _lowercase=False , **_lowercase ): '''simple docstring''' __a : Any = kwargs.pop("""add_prefix_space""" , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(_lowercase ) > 0 and not text[0].isspace()): __a : Optional[int] = """ """ + text return (text, kwargs)
712
"""simple docstring""" import unittest import numpy as np import torch from torch import nn from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import KandinskyVaaPriorPipeline, PriorTransformer, UnCLIPScheduler from diffusers.utils import torch_device from diffusers.utils.testing_utils import enable_full_determinism, skip_mps from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class SCREAMING_SNAKE_CASE__ ( __snake_case , unittest.TestCase ): _lowerCAmelCase = KandinskyVaaPriorPipeline _lowerCAmelCase = ["prompt"] _lowerCAmelCase = ["prompt", "negative_prompt"] _lowerCAmelCase = [ "num_images_per_prompt", "generator", "num_inference_steps", "latents", "negative_prompt", "guidance_scale", "output_type", "return_dict", ] _lowerCAmelCase = False @property def lowerCAmelCase__(self ): '''simple docstring''' return 32 @property def lowerCAmelCase__(self ): '''simple docstring''' return 32 @property def lowerCAmelCase__(self ): '''simple docstring''' return self.time_input_dim @property def lowerCAmelCase__(self ): '''simple docstring''' return self.time_input_dim * 4 @property def lowerCAmelCase__(self ): '''simple docstring''' return 100 @property def lowerCAmelCase__(self ): '''simple docstring''' __a : List[str] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) return tokenizer @property def lowerCAmelCase__(self ): '''simple docstring''' torch.manual_seed(0 ) __a : str = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModelWithProjection(_lowercase ) @property def lowerCAmelCase__(self ): '''simple docstring''' torch.manual_seed(0 ) __a : Dict = { """num_attention_heads""": 2, """attention_head_dim""": 12, """embedding_dim""": self.text_embedder_hidden_size, """num_layers""": 1, } __a : Tuple = PriorTransformer(**_lowercase ) # clip_std and clip_mean is initialized to be 0 so PriorTransformer.post_process_latents will always return 0 - set clip_std to be 1 so it won't return 0 __a : int = nn.Parameter(torch.ones(model.clip_std.shape ) ) return model @property def lowerCAmelCase__(self ): '''simple docstring''' torch.manual_seed(0 ) __a : List[str] = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , image_size=224 , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_channels=3 , num_hidden_layers=5 , patch_size=14 , ) __a : Optional[Any] = CLIPVisionModelWithProjection(_lowercase ) return model @property def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[Any] = CLIPImageProcessor( crop_size=224 , do_center_crop=_lowercase , do_normalize=_lowercase , do_resize=_lowercase , image_mean=[0.4814_5466, 0.457_8275, 0.4082_1073] , image_std=[0.2686_2954, 0.2613_0258, 0.2757_7711] , resample=3 , size=224 , ) return image_processor def lowerCAmelCase__(self ): '''simple docstring''' __a : Union[str, Any] = self.dummy_prior __a : int = self.dummy_image_encoder __a : Any = self.dummy_text_encoder __a : int = self.dummy_tokenizer __a : Optional[Any] = self.dummy_image_processor __a : List[Any] = UnCLIPScheduler( variance_type="""fixed_small_log""" , prediction_type="""sample""" , num_train_timesteps=1000 , clip_sample=_lowercase , clip_sample_range=10.0 , ) __a : List[Any] = { """prior""": prior, """image_encoder""": image_encoder, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """scheduler""": scheduler, """image_processor""": image_processor, } return components def lowerCAmelCase__(self , _lowercase , _lowercase=0 ): '''simple docstring''' if str(_lowercase ).startswith("""mps""" ): __a : Dict = torch.manual_seed(_lowercase ) else: __a : Union[str, Any] = torch.Generator(device=_lowercase ).manual_seed(_lowercase ) __a : Union[str, Any] = { """prompt""": """horse""", """generator""": generator, """guidance_scale""": 4.0, """num_inference_steps""": 2, """output_type""": """np""", } return inputs def lowerCAmelCase__(self ): '''simple docstring''' __a : Union[str, Any] = """cpu""" __a : Union[str, Any] = self.get_dummy_components() __a : Dict = self.pipeline_class(**_lowercase ) __a : Tuple = pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) __a : Optional[int] = pipe(**self.get_dummy_inputs(_lowercase ) ) __a : str = output.image_embeds __a : Any = pipe( **self.get_dummy_inputs(_lowercase ) , return_dict=_lowercase , )[0] __a : List[Any] = image[0, -10:] __a : List[Any] = image_from_tuple[0, -10:] assert image.shape == (1, 32) __a : Optional[Any] = np.array( [-0.0532, 1.7120, 0.3656, -1.0852, -0.8946, -1.1756, 0.4348, 0.2482, 0.5146, -0.1156] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @skip_mps def lowerCAmelCase__(self ): '''simple docstring''' __a : Any = torch_device == """cpu""" __a : Any = True __a : Any = False self._test_inference_batch_single_identical( test_max_difference=_lowercase , relax_max_difference=_lowercase , test_mean_pixel_difference=_lowercase , ) @skip_mps def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[int] = torch_device == """cpu""" __a : Union[str, Any] = False self._test_attention_slicing_forward_pass( test_max_difference=_lowercase , test_mean_pixel_difference=_lowercase , )
63
0
"""simple docstring""" def __magic_name__ ( _lowerCamelCase : int = 1_0**1_2 ): __a : List[str] = 1 __a : Any = 0 __a : Any = 1 __a : List[Any] = 1 while numerator <= 2 * min_total - 1: prev_numerator += 2 * numerator numerator += 2 * prev_numerator prev_denominator += 2 * denominator denominator += 2 * prev_denominator return (denominator + 1) // 2 if __name__ == "__main__": print(f'{solution() = }')
713
"""simple docstring""" import json import os import unittest from transformers import BatchEncoding, LEDTokenizer, LEDTokenizerFast from transformers.models.led.tokenization_led import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class SCREAMING_SNAKE_CASE__ ( __snake_case , unittest.TestCase ): _lowerCAmelCase = LEDTokenizer _lowerCAmelCase = LEDTokenizerFast _lowerCAmelCase = True def lowerCAmelCase__(self ): '''simple docstring''' super().setUp() __a : str = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", ] __a : int = dict(zip(_lowercase , range(len(_lowercase ) ) ) ) __a : Optional[int] = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] __a : List[Any] = {"""unk_token""": """<unk>"""} __a : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) __a : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(_lowercase ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(_lowercase ) ) def lowerCAmelCase__(self , **_lowercase ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **_lowercase ) def lowerCAmelCase__(self , **_lowercase ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **_lowercase ) def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' return "lower newer", "lower newer" @cached_property def lowerCAmelCase__(self ): '''simple docstring''' return LEDTokenizer.from_pretrained("""allenai/led-base-16384""" ) @cached_property def lowerCAmelCase__(self ): '''simple docstring''' return LEDTokenizerFast.from_pretrained("""allenai/led-base-16384""" ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' __a : Any = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] __a : List[str] = [0, 250, 251, 17818, 13, 39186, 1938, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __a : Optional[int] = tokenizer(_lowercase , max_length=len(_lowercase ) , padding=_lowercase , return_tensors="""pt""" ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual((2, 9) , batch.input_ids.shape ) self.assertEqual((2, 9) , batch.attention_mask.shape ) __a : Dict = batch.input_ids.tolist()[0] self.assertListEqual(_lowercase , _lowercase ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' __a : Union[str, Any] = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __a : Tuple = tokenizer(_lowercase , padding=_lowercase , return_tensors="""pt""" ) self.assertIn("""input_ids""" , _lowercase ) self.assertIn("""attention_mask""" , _lowercase ) self.assertNotIn("""labels""" , _lowercase ) self.assertNotIn("""decoder_attention_mask""" , _lowercase ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[Any] = [ """Summary of the text.""", """Another summary.""", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __a : Dict = tokenizer(text_target=_lowercase , max_length=32 , padding="""max_length""" , return_tensors="""pt""" ) self.assertEqual(32 , targets["""input_ids"""].shape[1] ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __a : Optional[int] = tokenizer( ["""I am a small frog""" * 1024, """I am a small frog"""] , padding=_lowercase , truncation=_lowercase , return_tensors="""pt""" ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual(batch.input_ids.shape , (2, 5122) ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' __a : Tuple = ["""A long paragraph for summarization."""] __a : Dict = [ """Summary of the text.""", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __a : int = tokenizer(_lowercase , return_tensors="""pt""" ) __a : Dict = tokenizer(text_target=_lowercase , return_tensors="""pt""" ) __a : List[str] = inputs["""input_ids"""] __a : List[Any] = targets["""input_ids"""] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item() ) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item() ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: __a : Optional[Any] = ["""Summary of the text.""", """Another summary."""] __a : List[Any] = [[0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, -1, -1]] __a : Union[str, Any] = tokenizer(_lowercase , padding=_lowercase ) __a : Tuple = [[0] * len(_lowercase ) for x in encoded_output["""input_ids"""]] __a : Union[str, Any] = tokenizer.pad(_lowercase ) self.assertSequenceEqual(outputs["""global_attention_mask"""] , _lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' pass def lowerCAmelCase__(self ): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): __a : Dict = self.rust_tokenizer_class.from_pretrained(_lowercase , **_lowercase ) __a : Union[str, Any] = self.tokenizer_class.from_pretrained(_lowercase , **_lowercase ) __a : Union[str, Any] = """A, <mask> AllenNLP sentence.""" __a : Dict = tokenizer_r.encode_plus(_lowercase , add_special_tokens=_lowercase , return_token_type_ids=_lowercase ) __a : Tuple = tokenizer_p.encode_plus(_lowercase , add_special_tokens=_lowercase , return_token_type_ids=_lowercase ) self.assertEqual(sum(tokens_r["""token_type_ids"""] ) , sum(tokens_p["""token_type_ids"""] ) ) self.assertEqual( sum(tokens_r["""attention_mask"""] ) / len(tokens_r["""attention_mask"""] ) , sum(tokens_p["""attention_mask"""] ) / len(tokens_p["""attention_mask"""] ) , ) __a : Tuple = tokenizer_r.convert_ids_to_tokens(tokens_r["""input_ids"""] ) __a : Any = tokenizer_p.convert_ids_to_tokens(tokens_p["""input_ids"""] ) self.assertSequenceEqual(tokens_p["""input_ids"""] , [0, 250, 6, 50264, 3823, 487, 21992, 3645, 4, 2] ) self.assertSequenceEqual(tokens_r["""input_ids"""] , [0, 250, 6, 50264, 3823, 487, 21992, 3645, 4, 2] ) self.assertSequenceEqual( _lowercase , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] ) self.assertSequenceEqual( _lowercase , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] )
63
0
"""simple docstring""" import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DeformableDetrImageProcessor class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def __init__(self , _lowercase , _lowercase=7 , _lowercase=3 , _lowercase=30 , _lowercase=400 , _lowercase=True , _lowercase=None , _lowercase=True , _lowercase=[0.5, 0.5, 0.5] , _lowercase=[0.5, 0.5, 0.5] , _lowercase=True , _lowercase=1 / 255 , _lowercase=True , ): '''simple docstring''' __a : List[Any] = size if size is not None else {"""shortest_edge""": 18, """longest_edge""": 1333} __a : Any = parent __a : Union[str, Any] = batch_size __a : List[str] = num_channels __a : str = min_resolution __a : List[Any] = max_resolution __a : Optional[int] = do_resize __a : List[str] = size __a : Any = do_normalize __a : Any = image_mean __a : List[Any] = image_std __a : str = do_rescale __a : Optional[int] = rescale_factor __a : int = do_pad def lowerCAmelCase__(self ): '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def lowerCAmelCase__(self , _lowercase , _lowercase=False ): '''simple docstring''' if not batched: __a : Union[str, Any] = image_inputs[0] if isinstance(_lowercase , Image.Image ): __a : Optional[Any] = image.size else: __a : Tuple = image.shape[1], image.shape[2] if w < h: __a : List[Any] = int(self.size["""shortest_edge"""] * h / w ) __a : Union[str, Any] = self.size["""shortest_edge"""] elif w > h: __a : str = self.size["""shortest_edge"""] __a : Dict = int(self.size["""shortest_edge"""] * w / h ) else: __a : Tuple = self.size["""shortest_edge"""] __a : Dict = self.size["""shortest_edge"""] else: __a : Tuple = [] for image in image_inputs: __a : Tuple = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __a : List[Any] = max(_lowercase , key=lambda _lowercase : item[0] )[0] __a : str = max(_lowercase , key=lambda _lowercase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class SCREAMING_SNAKE_CASE__ ( __snake_case , unittest.TestCase ): _lowerCAmelCase = DeformableDetrImageProcessor if is_vision_available() else None def lowerCAmelCase__(self ): '''simple docstring''' __a : int = DeformableDetrImageProcessingTester(self ) @property def lowerCAmelCase__(self ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[int] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowercase , """image_mean""" ) ) self.assertTrue(hasattr(_lowercase , """image_std""" ) ) self.assertTrue(hasattr(_lowercase , """do_normalize""" ) ) self.assertTrue(hasattr(_lowercase , """do_resize""" ) ) self.assertTrue(hasattr(_lowercase , """do_rescale""" ) ) self.assertTrue(hasattr(_lowercase , """do_pad""" ) ) self.assertTrue(hasattr(_lowercase , """size""" ) ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Tuple = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""shortest_edge""": 18, """longest_edge""": 1333} ) self.assertEqual(image_processor.do_pad , _lowercase ) __a : List[str] = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=_lowercase ) self.assertEqual(image_processor.size , {"""shortest_edge""": 42, """longest_edge""": 84} ) self.assertEqual(image_processor.do_pad , _lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' pass def lowerCAmelCase__(self ): '''simple docstring''' __a : int = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __a : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , Image.Image ) # Test not batched input __a : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values __a : Optional[Any] = self.image_processor_tester.get_expected_values(_lowercase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __a : Tuple = self.image_processor_tester.get_expected_values(_lowercase , batched=_lowercase ) __a : List[Any] = image_processing(_lowercase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def lowerCAmelCase__(self ): '''simple docstring''' __a : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __a : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowercase , numpify=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , np.ndarray ) # Test not batched input __a : List[str] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values __a : str = self.image_processor_tester.get_expected_values(_lowercase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __a : List[Any] = image_processing(_lowercase , return_tensors="""pt""" ).pixel_values __a : Any = self.image_processor_tester.get_expected_values(_lowercase , batched=_lowercase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def lowerCAmelCase__(self ): '''simple docstring''' __a : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __a : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowercase , torchify=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , torch.Tensor ) # Test not batched input __a : List[str] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values __a : Dict = self.image_processor_tester.get_expected_values(_lowercase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __a : Tuple = image_processing(_lowercase , return_tensors="""pt""" ).pixel_values __a : Optional[Any] = self.image_processor_tester.get_expected_values(_lowercase , batched=_lowercase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[int] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) with open("""./tests/fixtures/tests_samples/COCO/coco_annotations.txt""" , """r""" ) as f: __a : str = json.loads(f.read() ) __a : Union[str, Any] = {"""image_id""": 39769, """annotations""": target} # encode them __a : List[Any] = DeformableDetrImageProcessor() __a : List[str] = image_processing(images=_lowercase , annotations=_lowercase , return_tensors="""pt""" ) # verify pixel values __a : Dict = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["""pixel_values"""].shape , _lowercase ) __a : Optional[int] = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["""pixel_values"""][0, 0, 0, :3] , _lowercase , atol=1e-4 ) ) # verify area __a : List[str] = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""area"""] , _lowercase ) ) # verify boxes __a : Union[str, Any] = torch.Size([6, 4] ) self.assertEqual(encoding["""labels"""][0]["""boxes"""].shape , _lowercase ) __a : Tuple = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""boxes"""][0] , _lowercase , atol=1e-3 ) ) # verify image_id __a : Optional[int] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""image_id"""] , _lowercase ) ) # verify is_crowd __a : str = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""iscrowd"""] , _lowercase ) ) # verify class_labels __a : Any = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""class_labels"""] , _lowercase ) ) # verify orig_size __a : Dict = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""orig_size"""] , _lowercase ) ) # verify size __a : Optional[Any] = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""size"""] , _lowercase ) ) @slow def lowerCAmelCase__(self ): '''simple docstring''' __a : int = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) with open("""./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt""" , """r""" ) as f: __a : List[Any] = json.loads(f.read() ) __a : int = {"""file_name""": """000000039769.png""", """image_id""": 39769, """segments_info""": target} __a : Optional[int] = pathlib.Path("""./tests/fixtures/tests_samples/COCO/coco_panoptic""" ) # encode them __a : Optional[int] = DeformableDetrImageProcessor(format="""coco_panoptic""" ) __a : List[Any] = image_processing(images=_lowercase , annotations=_lowercase , masks_path=_lowercase , return_tensors="""pt""" ) # verify pixel values __a : Optional[int] = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["""pixel_values"""].shape , _lowercase ) __a : str = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["""pixel_values"""][0, 0, 0, :3] , _lowercase , atol=1e-4 ) ) # verify area __a : List[Any] = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""area"""] , _lowercase ) ) # verify boxes __a : Optional[Any] = torch.Size([6, 4] ) self.assertEqual(encoding["""labels"""][0]["""boxes"""].shape , _lowercase ) __a : Optional[int] = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""boxes"""][0] , _lowercase , atol=1e-3 ) ) # verify image_id __a : Optional[int] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""image_id"""] , _lowercase ) ) # verify is_crowd __a : str = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""iscrowd"""] , _lowercase ) ) # verify class_labels __a : Any = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""class_labels"""] , _lowercase ) ) # verify masks __a : str = 822873 self.assertEqual(encoding["""labels"""][0]["""masks"""].sum().item() , _lowercase ) # verify orig_size __a : Tuple = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""orig_size"""] , _lowercase ) ) # verify size __a : Tuple = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""size"""] , _lowercase ) )
714
"""simple docstring""" import argparse from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_controlnet_from_original_ckpt if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() parser.add_argument( "--checkpoint_path", default=None, type=str, required=True, help="Path to the checkpoint to convert." ) parser.add_argument( "--original_config_file", type=str, required=True, help="The YAML config file corresponding to the original architecture.", ) parser.add_argument( "--num_in_channels", default=None, type=int, help="The number of input channels. If `None` number of input channels will be automatically inferred.", ) parser.add_argument( "--image_size", default=512, type=int, help=( "The image size that the model was trained on. Use 512 for Stable Diffusion v1.X and Stable Siffusion v2" " Base. Use 768 for Stable Diffusion v2." ), ) parser.add_argument( "--extract_ema", action="store_true", help=( "Only relevant for checkpoints that have both EMA and non-EMA weights. Whether to extract the EMA weights" " or not. Defaults to `False`. Add `--extract_ema` to extract the EMA weights. EMA weights usually yield" " higher quality images for inference. Non-EMA weights are usually better to continue fine-tuning." ), ) parser.add_argument( "--upcast_attention", action="store_true", help=( "Whether the attention computation should always be upcasted. This is necessary when running stable" " diffusion 2.1." ), ) parser.add_argument( "--from_safetensors", action="store_true", help="If `--checkpoint_path` is in `safetensors` format, load checkpoint with safetensors instead of PyTorch.", ) parser.add_argument( "--to_safetensors", action="store_true", help="Whether to store pipeline in safetensors format or not.", ) parser.add_argument("--dump_path", default=None, type=str, required=True, help="Path to the output model.") parser.add_argument("--device", type=str, help="Device to use (e.g. cpu, cuda:0, cuda:1, etc.)") def __magic_name__ ( _lowerCamelCase : Optional[Any] ): if string == "True": return True elif string == "False": return False else: raise ValueError(F'''could not parse string as bool {string}''' ) parser.add_argument( "--use_linear_projection", help="Override for use linear projection", required=False, type=parse_bool ) parser.add_argument("--cross_attention_dim", help="Override for cross attention_dim", required=False, type=int) lowercase__ = parser.parse_args() lowercase__ = download_controlnet_from_original_ckpt( checkpoint_path=args.checkpoint_path, original_config_file=args.original_config_file, image_size=args.image_size, extract_ema=args.extract_ema, num_in_channels=args.num_in_channels, upcast_attention=args.upcast_attention, from_safetensors=args.from_safetensors, device=args.device, use_linear_projection=args.use_linear_projection, cross_attention_dim=args.cross_attention_dim, ) controlnet.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
63
0
"""simple docstring""" from __future__ import annotations def __magic_name__ ( _lowerCamelCase : list[int] ): # This function is recursive __a : str = len(_lowerCamelCase ) # If the array contains only one element, we return it (it's the stop condition of # recursion) if array_length <= 1: return array # Else __a : Dict = array[0] __a : Any = False __a : Any = 1 __a : list[int] = [] while not is_found and i < array_length: if array[i] < pivot: __a : str = True __a : str = [element for element in array[i:] if element >= array[i]] __a : Optional[int] = longest_subsequence(_lowerCamelCase ) if len(_lowerCamelCase ) > len(_lowerCamelCase ): __a : Dict = temp_array else: i += 1 __a : int = [element for element in array[1:] if element >= pivot] __a : List[str] = [pivot, *longest_subsequence(_lowerCamelCase )] if len(_lowerCamelCase ) > len(_lowerCamelCase ): return temp_array else: return longest_subseq if __name__ == "__main__": import doctest doctest.testmod()
715
"""simple docstring""" import torch from diffusers import DiffusionPipeline class SCREAMING_SNAKE_CASE__ ( __snake_case ): def __init__(self , _lowercase , _lowercase ): '''simple docstring''' super().__init__() self.register_modules(unet=_lowercase , scheduler=_lowercase ) def __call__(self ): '''simple docstring''' __a : Dict = torch.randn( (1, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , ) __a : Optional[Any] = 1 __a : List[str] = self.unet(_lowercase , _lowercase ).sample __a : Union[str, Any] = self.scheduler.step(_lowercase , _lowercase , _lowercase ).prev_sample __a : Optional[int] = scheduler_output - scheduler_output + torch.ones_like(_lowercase ) return result
63
0
"""simple docstring""" import warnings from ...utils import logging from .image_processing_poolformer import PoolFormerImageProcessor lowercase__ = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( __snake_case ): def __init__(self , *_lowercase , **_lowercase ): '''simple docstring''' warnings.warn( """The class PoolFormerFeatureExtractor is deprecated and will be removed in version 5 of Transformers.""" """ Please use PoolFormerImageProcessor instead.""" , _lowercase , ) super().__init__(*_lowercase , **_lowercase )
716
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { "sayakpaul/vit-msn-base": "https://huggingface.co/sayakpaul/vit-msn-base/resolve/main/config.json", # See all ViT MSN models at https://huggingface.co/models?filter=vit_msn } class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = "vit_msn" def __init__(self , _lowercase=768 , _lowercase=12 , _lowercase=12 , _lowercase=3072 , _lowercase="gelu" , _lowercase=0.0 , _lowercase=0.0 , _lowercase=0.02 , _lowercase=1e-06 , _lowercase=224 , _lowercase=16 , _lowercase=3 , _lowercase=True , **_lowercase , ): '''simple docstring''' super().__init__(**_lowercase ) __a : int = hidden_size __a : str = num_hidden_layers __a : str = num_attention_heads __a : Optional[Any] = intermediate_size __a : Union[str, Any] = hidden_act __a : Tuple = hidden_dropout_prob __a : Any = attention_probs_dropout_prob __a : List[Any] = initializer_range __a : Any = layer_norm_eps __a : Dict = image_size __a : List[Any] = patch_size __a : Dict = num_channels __a : Optional[Any] = qkv_bias
63
0
"""simple docstring""" import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv("TEST_SAGEMAKER" , "False" ) ) is not True , reason="Skipping test because should only be run when releasing minor transformers version" , ) @pytest.mark.usefixtures("sm_env" ) @parameterized_class( [ { "framework": "pytorch", "script": "run_glue_model_parallelism.py", "model_name_or_path": "roberta-large", "instance_type": "ml.p3dn.24xlarge", "results": {"train_runtime": 1_6_0_0, "eval_accuracy": 0.3, "eval_loss": 1.2}, }, { "framework": "pytorch", "script": "run_glue.py", "model_name_or_path": "roberta-large", "instance_type": "ml.p3dn.24xlarge", "results": {"train_runtime": 1_6_0_0, "eval_accuracy": 0.3, "eval_loss": 1.2}, }, ] ) class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def lowerCAmelCase__(self ): '''simple docstring''' if self.framework == "pytorch": subprocess.run( F'''cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py'''.split() , encoding="""utf-8""" , check=_lowercase , ) assert hasattr(self , """env""" ) def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' __a : Optional[Any] = { """enabled""": True, """processes_per_host""": 8, } __a : Union[str, Any] = { """enabled""": True, """parameters""": { """microbatches""": 4, """placement_strategy""": """spread""", """pipeline""": """interleaved""", """optimize""": """speed""", """partitions""": 4, """ddp""": True, }, } __a : Optional[Any] = {"""smdistributed""": {"""modelparallel""": smp_options}, """mpi""": mpi_options} __a : Tuple = """trainer""" if self.script == """run_glue.py""" else """smtrainer""" # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=F'''{self.env.base_job_name}-{instance_count}-smp-{name_extension}''' , instance_count=_lowercase , instance_type=self.instance_type , debugger_hook_config=_lowercase , hyperparameters={ **self.env.hyperparameters, """model_name_or_path""": self.model_name_or_path, """max_steps""": 500, } , metric_definitions=self.env.metric_definitions , distribution=_lowercase , py_version="""py36""" , ) def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' TrainingJobAnalytics(_lowercase ).export_csv(F'''{self.env.test_path}/{job_name}_metrics.csv''' ) @parameterized.expand([(1,)] ) def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' __a : Any = self.create_estimator(_lowercase ) # run training estimator.fit() # result dataframe __a : Tuple = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis __a : List[Any] = list(result_metrics_df[result_metrics_df.metric_name == """eval_accuracy"""]["""value"""] ) __a : List[str] = list(result_metrics_df[result_metrics_df.metric_name == """eval_loss"""]["""value"""] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping __a : Dict = ( Session().describe_training_job(estimator.latest_training_job.name ).get("""TrainingTimeInSeconds""" , 999999 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["""eval_accuracy"""] for t in eval_accuracy ) assert all(t <= self.results["""eval_loss"""] for t in eval_loss ) # dump tests result into json file to share in PR with open(F'''{estimator.latest_training_job.name}.json''' , """w""" ) as outfile: json.dump({"""train_time""": train_runtime, """eval_accuracy""": eval_accuracy, """eval_loss""": eval_loss} , _lowercase )
717
"""simple docstring""" import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert_fast import BertTokenizerFast from .tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer, DPRReaderTokenizer lowercase__ = logging.get_logger(__name__) lowercase__ = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} lowercase__ = { "vocab_file": { "facebook/dpr-ctx_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-ctx_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-ctx_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-ctx_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/tokenizer.json" ), }, } lowercase__ = { "vocab_file": { "facebook/dpr-question_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-question_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-question_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-question_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/tokenizer.json" ), }, } lowercase__ = { "vocab_file": { "facebook/dpr-reader-single-nq-base": ( "https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-reader-multiset-base": ( "https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-reader-single-nq-base": ( "https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-reader-multiset-base": ( "https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/tokenizer.json" ), }, } lowercase__ = { "facebook/dpr-ctx_encoder-single-nq-base": 512, "facebook/dpr-ctx_encoder-multiset-base": 512, } lowercase__ = { "facebook/dpr-question_encoder-single-nq-base": 512, "facebook/dpr-question_encoder-multiset-base": 512, } lowercase__ = { "facebook/dpr-reader-single-nq-base": 512, "facebook/dpr-reader-multiset-base": 512, } lowercase__ = { "facebook/dpr-ctx_encoder-single-nq-base": {"do_lower_case": True}, "facebook/dpr-ctx_encoder-multiset-base": {"do_lower_case": True}, } lowercase__ = { "facebook/dpr-question_encoder-single-nq-base": {"do_lower_case": True}, "facebook/dpr-question_encoder-multiset-base": {"do_lower_case": True}, } lowercase__ = { "facebook/dpr-reader-single-nq-base": {"do_lower_case": True}, "facebook/dpr-reader-multiset-base": {"do_lower_case": True}, } class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = VOCAB_FILES_NAMES _lowerCAmelCase = CONTEXT_ENCODER_PRETRAINED_VOCAB_FILES_MAP _lowerCAmelCase = CONTEXT_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCAmelCase = CONTEXT_ENCODER_PRETRAINED_INIT_CONFIGURATION _lowerCAmelCase = DPRContextEncoderTokenizer class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = VOCAB_FILES_NAMES _lowerCAmelCase = QUESTION_ENCODER_PRETRAINED_VOCAB_FILES_MAP _lowerCAmelCase = QUESTION_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCAmelCase = QUESTION_ENCODER_PRETRAINED_INIT_CONFIGURATION _lowerCAmelCase = DPRQuestionEncoderTokenizer lowercase__ = collections.namedtuple( "DPRSpanPrediction", ["span_score", "relevance_score", "doc_id", "start_index", "end_index", "text"] ) lowercase__ = collections.namedtuple("DPRReaderOutput", ["start_logits", "end_logits", "relevance_logits"]) lowercase__ = R"\n Return a dictionary with the token ids of the input strings and other information to give to `.decode_best_spans`.\n It converts the strings of a question and different passages (title and text) in a sequence of IDs (integers),\n using the tokenizer and vocabulary. The resulting `input_ids` is a matrix of size `(n_passages, sequence_length)`\n with the format:\n\n [CLS] <question token ids> [SEP] <titles ids> [SEP] <texts ids>\n\n Args:\n questions (`str` or `List[str]`):\n The questions to be encoded. You can specify one question for many passages. In this case, the question\n will be duplicated like `[questions] * n_passages`. Otherwise you have to specify as many questions as in\n `titles` or `texts`.\n titles (`str` or `List[str]`):\n The passages titles to be encoded. This can be a string or a list of strings if there are several passages.\n texts (`str` or `List[str]`):\n The passages texts to be encoded. This can be a string or a list of strings if there are several passages.\n padding (`bool`, `str` or [`~utils.PaddingStrategy`], *optional*, defaults to `False`):\n Activates and controls padding. Accepts the following values:\n\n - `True` or `'longest'`: Pad to the longest sequence in the batch (or no padding if only a single sequence\n if provided).\n - `'max_length'`: Pad to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided.\n - `False` or `'do_not_pad'` (default): No padding (i.e., can output a batch with sequences of different\n lengths).\n truncation (`bool`, `str` or [`~tokenization_utils_base.TruncationStrategy`], *optional*, defaults to `False`):\n Activates and controls truncation. Accepts the following values:\n\n - `True` or `'longest_first'`: Truncate to a maximum length specified with the argument `max_length` or to\n the maximum acceptable input length for the model if that argument is not provided. This will truncate\n token by token, removing a token from the longest sequence in the pair if a pair of sequences (or a batch\n of pairs) is provided.\n - `'only_first'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the first\n sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `'only_second'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the\n second sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `False` or `'do_not_truncate'` (default): No truncation (i.e., can output batch with sequence lengths\n greater than the model maximum admissible input size).\n max_length (`int`, *optional*):\n Controls the maximum length to use by one of the truncation/padding parameters.\n\n If left unset or set to `None`, this will use the predefined model maximum length if a maximum length\n is required by one of the truncation/padding parameters. If the model has no specific maximum input\n length (like XLNet) truncation/padding to a maximum length will be deactivated.\n return_tensors (`str` or [`~utils.TensorType`], *optional*):\n If set, will return tensors instead of list of python integers. Acceptable values are:\n\n - `'tf'`: Return TensorFlow `tf.constant` objects.\n - `'pt'`: Return PyTorch `torch.Tensor` objects.\n - `'np'`: Return Numpy `np.ndarray` objects.\n return_attention_mask (`bool`, *optional*):\n Whether or not to return the attention mask. If not set, will return the attention mask according to the\n specific tokenizer's default, defined by the `return_outputs` attribute.\n\n [What are attention masks?](../glossary#attention-mask)\n\n Return:\n `Dict[str, List[List[int]]]`: A dictionary with the following keys:\n\n - `input_ids`: List of token ids to be fed to a model.\n - `attention_mask`: List of indices specifying which tokens should be attended to by the model.\n " @add_start_docstrings(__snake_case ) class SCREAMING_SNAKE_CASE__ : def __call__(self , _lowercase , _lowercase = None , _lowercase = None , _lowercase = False , _lowercase = False , _lowercase = None , _lowercase = None , _lowercase = None , **_lowercase , ): '''simple docstring''' if titles is None and texts is None: return super().__call__( _lowercase , padding=_lowercase , truncation=_lowercase , max_length=_lowercase , return_tensors=_lowercase , return_attention_mask=_lowercase , **_lowercase , ) elif titles is None or texts is None: __a : str = titles if texts is None else texts return super().__call__( _lowercase , _lowercase , padding=_lowercase , truncation=_lowercase , max_length=_lowercase , return_tensors=_lowercase , return_attention_mask=_lowercase , **_lowercase , ) __a : str = titles if not isinstance(_lowercase , _lowercase ) else [titles] __a : Optional[Any] = texts if not isinstance(_lowercase , _lowercase ) else [texts] __a : Tuple = len(_lowercase ) __a : Dict = questions if not isinstance(_lowercase , _lowercase ) else [questions] * n_passages assert len(_lowercase ) == len( _lowercase ), F'''There should be as many titles than texts but got {len(_lowercase )} titles and {len(_lowercase )} texts.''' __a : Optional[Any] = super().__call__(_lowercase , _lowercase , padding=_lowercase , truncation=_lowercase )["""input_ids"""] __a : str = super().__call__(_lowercase , add_special_tokens=_lowercase , padding=_lowercase , truncation=_lowercase )["""input_ids"""] __a : Union[str, Any] = { """input_ids""": [ (encoded_question_and_title + encoded_text)[:max_length] if max_length is not None and truncation else encoded_question_and_title + encoded_text for encoded_question_and_title, encoded_text in zip(_lowercase , _lowercase ) ] } if return_attention_mask is not False: __a : Optional[int] = [] for input_ids in encoded_inputs["input_ids"]: attention_mask.append([int(input_id != self.pad_token_id ) for input_id in input_ids] ) __a : str = attention_mask return self.pad(_lowercase , padding=_lowercase , max_length=_lowercase , return_tensors=_lowercase ) def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase = 16 , _lowercase = 64 , _lowercase = 4 , ): '''simple docstring''' __a : Union[str, Any] = reader_input["""input_ids"""] __a , __a , __a : Optional[int] = reader_output[:3] __a : int = len(_lowercase ) __a : Any = sorted(range(_lowercase ) , reverse=_lowercase , key=relevance_logits.__getitem__ ) __a : List[DPRReaderOutput] = [] for doc_id in sorted_docs: __a : Optional[int] = list(input_ids[doc_id] ) # assuming question & title information is at the beginning of the sequence __a : Dict = sequence_ids.index(self.sep_token_id , 2 ) + 1 # second sep id if sequence_ids[-1] == self.pad_token_id: __a : int = sequence_ids.index(self.pad_token_id ) else: __a : Optional[Any] = len(_lowercase ) __a : List[Any] = self._get_best_spans( start_logits=start_logits[doc_id][passage_offset:sequence_len] , end_logits=end_logits[doc_id][passage_offset:sequence_len] , max_answer_length=_lowercase , top_spans=_lowercase , ) for start_index, end_index in best_spans: start_index += passage_offset end_index += passage_offset nbest_spans_predictions.append( DPRSpanPrediction( span_score=start_logits[doc_id][start_index] + end_logits[doc_id][end_index] , relevance_score=relevance_logits[doc_id] , doc_id=_lowercase , start_index=_lowercase , end_index=_lowercase , text=self.decode(sequence_ids[start_index : end_index + 1] ) , ) ) if len(_lowercase ) >= num_spans: break return nbest_spans_predictions[:num_spans] def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase , _lowercase , ): '''simple docstring''' __a : Tuple = [] for start_index, start_score in enumerate(_lowercase ): for answer_length, end_score in enumerate(end_logits[start_index : start_index + max_answer_length] ): scores.append(((start_index, start_index + answer_length), start_score + end_score) ) __a : str = sorted(_lowercase , key=lambda _lowercase : x[1] , reverse=_lowercase ) __a : Union[str, Any] = [] for (start_index, end_index), score in scores: assert start_index <= end_index, F'''Wrong span indices: [{start_index}:{end_index}]''' __a : List[str] = end_index - start_index + 1 assert length <= max_answer_length, F'''Span is too long: {length} > {max_answer_length}''' if any( start_index <= prev_start_index <= prev_end_index <= end_index or prev_start_index <= start_index <= end_index <= prev_end_index for (prev_start_index, prev_end_index) in chosen_span_intervals ): continue chosen_span_intervals.append((start_index, end_index) ) if len(_lowercase ) == top_spans: break return chosen_span_intervals @add_end_docstrings(__snake_case ) class SCREAMING_SNAKE_CASE__ ( __snake_case , __snake_case ): _lowerCAmelCase = VOCAB_FILES_NAMES _lowerCAmelCase = READER_PRETRAINED_VOCAB_FILES_MAP _lowerCAmelCase = READER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCAmelCase = READER_PRETRAINED_INIT_CONFIGURATION _lowerCAmelCase = ["input_ids", "attention_mask"] _lowerCAmelCase = DPRReaderTokenizer
63
0
"""simple docstring""" import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import DetrConfig, MaskFormerConfig, SwinConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskFormerForInstanceSegmentation, MaskFormerModel if is_vision_available(): from transformers import MaskFormerImageProcessor if is_vision_available(): from PIL import Image class SCREAMING_SNAKE_CASE__ : def __init__(self , _lowercase , _lowercase=2 , _lowercase=True , _lowercase=False , _lowercase=10 , _lowercase=3 , _lowercase=32 * 4 , _lowercase=32 * 6 , _lowercase=4 , _lowercase=32 , ): '''simple docstring''' __a : Union[str, Any] = parent __a : Union[str, Any] = batch_size __a : List[str] = is_training __a : Optional[Any] = use_auxiliary_loss __a : List[str] = num_queries __a : Tuple = num_channels __a : Any = min_size __a : int = max_size __a : str = num_labels __a : str = mask_feature_size def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( _lowercase ) __a : Dict = torch.ones([self.batch_size, self.min_size, self.max_size] , device=_lowercase ) __a : str = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=_lowercase ) > 0.5 ).float() __a : Dict = (torch.rand((self.batch_size, self.num_labels) , device=_lowercase ) > 0.5).long() __a : Optional[Any] = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def lowerCAmelCase__(self ): '''simple docstring''' return MaskFormerConfig.from_backbone_and_decoder_configs( backbone_config=SwinConfig( depths=[1, 1, 1, 1] , ) , decoder_config=DetrConfig( decoder_ffn_dim=128 , num_queries=self.num_queries , decoder_attention_heads=2 , d_model=self.mask_feature_size , ) , mask_feature_size=self.mask_feature_size , fpn_feature_size=self.mask_feature_size , num_channels=self.num_channels , num_labels=self.num_labels , ) def lowerCAmelCase__(self ): '''simple docstring''' __a : List[str] = self.prepare_config_and_inputs() __a : Dict = {"""pixel_values""": pixel_values, """pixel_mask""": pixel_mask} return config, inputs_dict def lowerCAmelCase__(self , _lowercase , _lowercase ): '''simple docstring''' __a : Optional[int] = output.encoder_hidden_states __a : Optional[Any] = output.pixel_decoder_hidden_states __a : Optional[Any] = output.transformer_decoder_hidden_states self.parent.assertTrue(len(_lowercase ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_lowercase ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_lowercase ) , config.decoder_config.decoder_layers ) def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase , _lowercase=False ): '''simple docstring''' with torch.no_grad(): __a : Any = MaskFormerModel(config=_lowercase ) model.to(_lowercase ) model.eval() __a : Dict = model(pixel_values=_lowercase , pixel_mask=_lowercase ) __a : str = model(_lowercase , output_hidden_states=_lowercase ) # the correct shape of output.transformer_decoder_hidden_states ensure the correcteness of the # encoder and pixel decoder self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.mask_feature_size) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(_lowercase , _lowercase ) def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase ): '''simple docstring''' __a : Optional[Any] = MaskFormerForInstanceSegmentation(config=_lowercase ) model.to(_lowercase ) model.eval() def comm_check_on_output(_lowercase ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): __a : Optional[int] = model(pixel_values=_lowercase , pixel_mask=_lowercase ) __a : List[Any] = model(_lowercase ) comm_check_on_output(_lowercase ) __a : Dict = model( pixel_values=_lowercase , pixel_mask=_lowercase , mask_labels=_lowercase , class_labels=_lowercase ) comm_check_on_output(_lowercase ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class SCREAMING_SNAKE_CASE__ ( __snake_case , __snake_case , unittest.TestCase ): _lowerCAmelCase = (MaskFormerModel, MaskFormerForInstanceSegmentation) if is_torch_available() else () _lowerCAmelCase = ( {"feature-extraction": MaskFormerModel, "image-segmentation": MaskFormerForInstanceSegmentation} if is_torch_available() else {} ) _lowerCAmelCase = False _lowerCAmelCase = False _lowerCAmelCase = False _lowerCAmelCase = False def lowerCAmelCase__(self ): '''simple docstring''' __a : int = MaskFormerModelTester(self ) __a : int = ConfigTester(self , config_class=_lowercase , has_text_modality=_lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCAmelCase__(self ): '''simple docstring''' __a : Dict = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(_lowercase , **_lowercase , output_hidden_states=_lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskformer_instance_segmentation_head_model(*_lowercase ) @unittest.skip(reason="""MaskFormer does not use inputs_embeds""" ) def lowerCAmelCase__(self ): '''simple docstring''' pass @unittest.skip(reason="""MaskFormer does not have a get_input_embeddings method""" ) def lowerCAmelCase__(self ): '''simple docstring''' pass @unittest.skip(reason="""MaskFormer is not a generative model""" ) def lowerCAmelCase__(self ): '''simple docstring''' pass @unittest.skip(reason="""MaskFormer does not use token embeddings""" ) def lowerCAmelCase__(self ): '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip( reason="""MaskFormer has some layers using `add_module` which doesn't work well with `nn.DataParallel`""" ) def lowerCAmelCase__(self ): '''simple docstring''' pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowerCAmelCase__(self ): '''simple docstring''' pass def lowerCAmelCase__(self ): '''simple docstring''' __a : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a : Optional[int] = model_class(_lowercase ) __a : Optional[int] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __a : Union[str, Any] = [*signature.parameters.keys()] __a : Tuple = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowercase ) @slow def lowerCAmelCase__(self ): '''simple docstring''' for model_name in ["facebook/maskformer-swin-small-coco"]: __a : str = MaskFormerModel.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' __a : int = (self.model_tester.min_size,) * 2 __a : Optional[int] = { """pixel_values""": torch.randn((2, 3, *size) , device=_lowercase ), """mask_labels""": torch.randn((2, 10, *size) , device=_lowercase ), """class_labels""": torch.zeros(2 , 10 , device=_lowercase ).long(), } __a : Union[str, Any] = MaskFormerForInstanceSegmentation(MaskFormerConfig() ).to(_lowercase ) __a : Optional[Any] = model(**_lowercase ) self.assertTrue(outputs.loss is not None ) def lowerCAmelCase__(self ): '''simple docstring''' __a : str = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(_lowercase , **_lowercase , output_hidden_states=_lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' __a : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a : Optional[Any] = model_class(_lowercase ).to(_lowercase ) __a : int = model(**_lowercase , output_attentions=_lowercase ) self.assertTrue(outputs.attentions is not None ) def lowerCAmelCase__(self ): '''simple docstring''' if not self.model_tester.is_training: return # only MaskFormerForInstanceSegmentation has the loss __a : Union[str, Any] = self.all_model_classes[1] __a : List[Any] = self.model_tester.prepare_config_and_inputs() __a : Tuple = model_class(_lowercase ) model.to(_lowercase ) model.train() __a : Optional[Any] = model(_lowercase , mask_labels=_lowercase , class_labels=_lowercase ).loss loss.backward() def lowerCAmelCase__(self ): '''simple docstring''' __a : List[Any] = self.all_model_classes[1] __a : Tuple = self.model_tester.prepare_config_and_inputs() __a : Optional[int] = True __a : Dict = True __a : Union[str, Any] = model_class(_lowercase ) model.to(_lowercase ) model.train() __a : Optional[int] = model(_lowercase , mask_labels=_lowercase , class_labels=_lowercase ) __a : str = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() __a : List[Any] = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() # we requires_grad=True in inputs_embeds (line 2152), the original implementation don't __a : Tuple = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() __a : Tuple = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=_lowercase ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) lowercase__ = 1e-4 def __magic_name__ ( ): __a : Optional[Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_vision @slow class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @cached_property def lowerCAmelCase__(self ): '''simple docstring''' return ( MaskFormerImageProcessor.from_pretrained("""facebook/maskformer-swin-small-coco""" ) if is_vision_available() else None ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[int] = MaskFormerModel.from_pretrained("""facebook/maskformer-swin-small-coco""" ).to(_lowercase ) __a : int = self.default_image_processor __a : List[Any] = prepare_img() __a : Any = image_processor(_lowercase , return_tensors="""pt""" ).to(_lowercase ) __a : Dict = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_lowercase , (1, 3, 800, 1088) ) with torch.no_grad(): __a : List[Any] = model(**_lowercase ) __a : Dict = torch.tensor( [[-0.0482, 0.9228, 0.4951], [-0.2547, 0.8017, 0.8527], [-0.0069, 0.3385, -0.0089]] ).to(_lowercase ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , _lowercase , atol=_lowercase ) ) __a : int = torch.tensor( [[-0.8422, -0.8434, -0.9718], [-1.0144, -0.5565, -0.4195], [-1.0038, -0.4484, -0.1961]] ).to(_lowercase ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , _lowercase , atol=_lowercase ) ) __a : List[Any] = torch.tensor( [[0.2852, -0.0159, 0.9735], [0.6254, 0.1858, 0.8529], [-0.0680, -0.4116, 1.8413]] ).to(_lowercase ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , _lowercase , atol=_lowercase ) ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[int] = ( MaskFormerForInstanceSegmentation.from_pretrained("""facebook/maskformer-swin-small-coco""" ) .to(_lowercase ) .eval() ) __a : int = self.default_image_processor __a : Optional[Any] = prepare_img() __a : Optional[Any] = image_processor(_lowercase , return_tensors="""pt""" ).to(_lowercase ) __a : Union[str, Any] = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_lowercase , (1, 3, 800, 1088) ) with torch.no_grad(): __a : Optional[int] = model(**_lowercase ) # masks_queries_logits __a : Optional[int] = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) __a : Dict = [ [-1.373_7124, -1.772_4937, -1.936_4233], [-1.597_7281, -1.986_7939, -2.152_3695], [-1.579_5398, -1.926_9832, -2.09_3942], ] __a : Dict = torch.tensor(_lowercase ).to(_lowercase ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , _lowercase , atol=_lowercase ) ) # class_queries_logits __a : Union[str, Any] = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) __a : List[str] = torch.tensor( [ [1.6512e00, -5.2572e00, -3.3519e00], [3.6169e-02, -5.9025e00, -2.9313e00], [1.0766e-04, -7.7630e00, -5.1263e00], ] ).to(_lowercase ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , _lowercase , atol=_lowercase ) ) def lowerCAmelCase__(self ): '''simple docstring''' __a : List[str] = ( MaskFormerForInstanceSegmentation.from_pretrained("""facebook/maskformer-resnet101-coco-stuff""" ) .to(_lowercase ) .eval() ) __a : List[str] = self.default_image_processor __a : Tuple = prepare_img() __a : int = image_processor(_lowercase , return_tensors="""pt""" ).to(_lowercase ) __a : str = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_lowercase , (1, 3, 800, 1088) ) with torch.no_grad(): __a : Tuple = model(**_lowercase ) # masks_queries_logits __a : Optional[int] = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) __a : Optional[int] = [[-0.9046, -2.6366, -4.6062], [-3.4179, -5.7890, -8.8057], [-4.9179, -7.6560, -10.7711]] __a : Dict = torch.tensor(_lowercase ).to(_lowercase ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , _lowercase , atol=_lowercase ) ) # class_queries_logits __a : int = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) __a : Optional[Any] = torch.tensor( [[4.7188, -3.2585, -2.8857], [6.6871, -2.9181, -1.2487], [7.2449, -2.2764, -2.1874]] ).to(_lowercase ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , _lowercase , atol=_lowercase ) ) def lowerCAmelCase__(self ): '''simple docstring''' __a : List[Any] = ( MaskFormerForInstanceSegmentation.from_pretrained("""facebook/maskformer-swin-small-coco""" ) .to(_lowercase ) .eval() ) __a : Optional[Any] = self.default_image_processor __a : Tuple = image_processor( [np.zeros((3, 800, 1333) ), np.zeros((3, 800, 1333) )] , segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] , return_tensors="""pt""" , ) __a : Optional[int] = inputs["""pixel_values"""].to(_lowercase ) __a : List[Any] = [el.to(_lowercase ) for el in inputs["""mask_labels"""]] __a : Optional[int] = [el.to(_lowercase ) for el in inputs["""class_labels"""]] with torch.no_grad(): __a : Union[str, Any] = model(**_lowercase ) self.assertTrue(outputs.loss is not None )
718
"""simple docstring""" import os def __magic_name__ ( _lowerCamelCase : Dict ): __a : List[str] = len(grid[0] ) __a : int = len(_lowerCamelCase ) __a : Tuple = 0 __a : List[Any] = 0 __a : List[str] = 0 # Check vertically, horizontally, diagonally at the same time (only works # for nxn grid) for i in range(_lowerCamelCase ): for j in range(n_rows - 3 ): __a : List[Any] = grid[j][i] * grid[j + 1][i] * grid[j + 2][i] * grid[j + 3][i] __a : Tuple = grid[i][j] * grid[i][j + 1] * grid[i][j + 2] * grid[i][j + 3] # Left-to-right diagonal (\) product if i < n_columns - 3: __a : List[Any] = ( grid[i][j] * grid[i + 1][j + 1] * grid[i + 2][j + 2] * grid[i + 3][j + 3] ) # Right-to-left diagonal(/) product if i > 2: __a : List[Any] = ( grid[i][j] * grid[i - 1][j + 1] * grid[i - 2][j + 2] * grid[i - 3][j + 3] ) __a : str = max( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) if max_product > largest: __a : Optional[Any] = max_product return largest def __magic_name__ ( ): __a : Tuple = [] with open(os.path.dirname(_lowerCamelCase ) + """/grid.txt""" ) as file: for line in file: grid.append(line.strip("""\n""" ).split(""" """ ) ) __a : Tuple = [[int(_lowerCamelCase ) for i in grid[j]] for j in range(len(_lowerCamelCase ) )] return largest_product(_lowerCamelCase ) if __name__ == "__main__": print(solution())
63
0
"""simple docstring""" from collections.abc import Iterable from typing import Any class SCREAMING_SNAKE_CASE__ : def __init__(self , _lowercase = None ): '''simple docstring''' __a : int = value __a : Node | None = None # Added in order to delete a node easier __a : Node | None = None __a : Node | None = None def __repr__(self ): '''simple docstring''' from pprint import pformat if self.left is None and self.right is None: return str(self.value ) return pformat({F'''{self.value}''': (self.left, self.right)} , indent=1 ) class SCREAMING_SNAKE_CASE__ : def __init__(self , _lowercase = None ): '''simple docstring''' __a : str = root def __str__(self ): '''simple docstring''' return str(self.root ) def lowerCAmelCase__(self , _lowercase , _lowercase ): '''simple docstring''' if new_children is not None: # reset its kids __a : Tuple = node.parent if node.parent is not None: # reset its parent if self.is_right(_lowercase ): # If it is the right children __a : Dict = new_children else: __a : List[Any] = new_children else: __a : str = new_children def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' if node.parent and node.parent.right: return node == node.parent.right return False def lowerCAmelCase__(self ): '''simple docstring''' return self.root is None def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' __a : Optional[Any] = Node(_lowercase ) # create a new Node if self.empty(): # if Tree is empty __a : Any = new_node # set its root else: # Tree is not empty __a : Dict = self.root # from root if parent_node is None: return while True: # While we don't get to a leaf if value < parent_node.value: # We go left if parent_node.left is None: __a : Optional[int] = new_node # We insert the new node in a leaf break else: __a : Optional[int] = parent_node.left else: if parent_node.right is None: __a : Optional[Any] = new_node break else: __a : str = parent_node.right __a : Union[str, Any] = parent_node def lowerCAmelCase__(self , *_lowercase ): '''simple docstring''' for value in values: self.__insert(_lowercase ) def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' if self.empty(): raise IndexError("""Warning: Tree is empty! please use another.""" ) else: __a : Any = self.root # use lazy evaluation here to avoid NoneType Attribute error while node is not None and node.value is not value: __a : int = node.left if value < node.value else node.right return node def lowerCAmelCase__(self , _lowercase = None ): '''simple docstring''' if node is None: if self.root is None: return None __a : Union[str, Any] = self.root if not self.empty(): while node.right is not None: __a : Dict = node.right return node def lowerCAmelCase__(self , _lowercase = None ): '''simple docstring''' if node is None: __a : Any = self.root if self.root is None: return None if not self.empty(): __a : str = self.root while node.left is not None: __a : List[Any] = node.left return node def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' __a : int = self.search(_lowercase ) # Look for the node with that label if node is not None: if node.left is None and node.right is None: # If it has no children self.__reassign_nodes(_lowercase , _lowercase ) elif node.left is None: # Has only right children self.__reassign_nodes(_lowercase , node.right ) elif node.right is None: # Has only left children self.__reassign_nodes(_lowercase , node.left ) else: __a : Optional[Any] = self.get_max( node.left ) # Gets the max value of the left branch self.remove(tmp_node.value ) # type: ignore __a : Tuple = ( tmp_node.value # type: ignore ) # Assigns the value to the node to delete and keep tree structure def lowerCAmelCase__(self , _lowercase ): '''simple docstring''' if node is not None: yield node # Preorder Traversal yield from self.preorder_traverse(node.left ) yield from self.preorder_traverse(node.right ) def lowerCAmelCase__(self , _lowercase=None ): '''simple docstring''' if traversal_function is None: return self.preorder_traverse(self.root ) else: return traversal_function(self.root ) def lowerCAmelCase__(self , _lowercase , _lowercase ): '''simple docstring''' if node: self.inorder(_lowercase , node.left ) arr.append(node.value ) self.inorder(_lowercase , node.right ) def lowerCAmelCase__(self , _lowercase , _lowercase ): '''simple docstring''' __a : list[int] = [] self.inorder(_lowercase , _lowercase ) # append all values to list using inorder traversal return arr[k - 1] def __magic_name__ ( _lowerCamelCase : Node | None ): __a : str = [] if curr_node is not None: __a : Dict = postorder(curr_node.left ) + postorder(curr_node.right ) + [curr_node] return node_list def __magic_name__ ( ): __a : Dict = (8, 3, 6, 1, 1_0, 1_4, 1_3, 4, 7) __a : List[str] = BinarySearchTree() for i in testlist: t.insert(_lowerCamelCase ) # Prints all the elements of the list in order traversal print(_lowerCamelCase ) if t.search(6 ) is not None: print("""The value 6 exists""" ) else: print("""The value 6 doesn't exist""" ) if t.search(-1 ) is not None: print("""The value -1 exists""" ) else: print("""The value -1 doesn't exist""" ) if not t.empty(): print("""Max Value: """ , t.get_max().value ) # type: ignore print("""Min Value: """ , t.get_min().value ) # type: ignore for i in testlist: t.remove(_lowerCamelCase ) print(_lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
719
"""simple docstring""" from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import BaseOutput, is_torch_available, is_transformers_available @dataclass class SCREAMING_SNAKE_CASE__ ( __snake_case ): _lowerCAmelCase = 42 _lowerCAmelCase = 42 if is_transformers_available() and is_torch_available(): from .pipeline_semantic_stable_diffusion import SemanticStableDiffusionPipeline
63
0
"""simple docstring""" from math import cos, sin, sqrt, tau from audio_filters.iir_filter import IIRFilter def __magic_name__ ( _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : float = 1 / sqrt(2 ) ): __a : Optional[int] = tau * frequency / samplerate __a : Dict = sin(_lowerCamelCase ) __a : Optional[int] = cos(_lowerCamelCase ) __a : int = _sin / (2 * q_factor) __a : List[Any] = (1 - _cos) / 2 __a : Dict = 1 - _cos __a : Optional[int] = 1 + alpha __a : List[str] = -2 * _cos __a : Optional[Any] = 1 - alpha __a : Optional[Any] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def __magic_name__ ( _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : float = 1 / sqrt(2 ) ): __a : str = tau * frequency / samplerate __a : Optional[Any] = sin(_lowerCamelCase ) __a : List[str] = cos(_lowerCamelCase ) __a : Any = _sin / (2 * q_factor) __a : Optional[int] = (1 + _cos) / 2 __a : int = -1 - _cos __a : Optional[Any] = 1 + alpha __a : int = -2 * _cos __a : str = 1 - alpha __a : Tuple = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def __magic_name__ ( _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : float = 1 / sqrt(2 ) ): __a : Any = tau * frequency / samplerate __a : List[Any] = sin(_lowerCamelCase ) __a : Union[str, Any] = cos(_lowerCamelCase ) __a : List[str] = _sin / (2 * q_factor) __a : List[str] = _sin / 2 __a : str = 0 __a : int = -ba __a : List[str] = 1 + alpha __a : Any = -2 * _cos __a : Optional[Any] = 1 - alpha __a : List[str] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def __magic_name__ ( _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : float = 1 / sqrt(2 ) ): __a : Optional[int] = tau * frequency / samplerate __a : List[Any] = sin(_lowerCamelCase ) __a : Dict = cos(_lowerCamelCase ) __a : Optional[int] = _sin / (2 * q_factor) __a : List[str] = 1 - alpha __a : Any = -2 * _cos __a : Tuple = 1 + alpha __a : str = IIRFilter(2 ) filt.set_coefficients([ba, ba, ba] , [ba, ba, ba] ) return filt def __magic_name__ ( _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : float , _lowerCamelCase : float = 1 / sqrt(2 ) , ): __a : Union[str, Any] = tau * frequency / samplerate __a : Optional[Any] = sin(_lowerCamelCase ) __a : str = cos(_lowerCamelCase ) __a : Any = _sin / (2 * q_factor) __a : Any = 1_0 ** (gain_db / 4_0) __a : str = 1 + alpha * big_a __a : str = -2 * _cos __a : Optional[int] = 1 - alpha * big_a __a : Optional[int] = 1 + alpha / big_a __a : str = -2 * _cos __a : Tuple = 1 - alpha / big_a __a : Tuple = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def __magic_name__ ( _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : float , _lowerCamelCase : float = 1 / sqrt(2 ) , ): __a : Union[str, Any] = tau * frequency / samplerate __a : str = sin(_lowerCamelCase ) __a : List[Any] = cos(_lowerCamelCase ) __a : Dict = _sin / (2 * q_factor) __a : Any = 1_0 ** (gain_db / 4_0) __a : int = (big_a + 1) - (big_a - 1) * _cos __a : str = (big_a + 1) + (big_a - 1) * _cos __a : Union[str, Any] = (big_a - 1) - (big_a + 1) * _cos __a : Dict = (big_a - 1) + (big_a + 1) * _cos __a : int = 2 * sqrt(_lowerCamelCase ) * alpha __a : str = big_a * (pmc + aaa) __a : Optional[int] = 2 * big_a * mpc __a : Optional[int] = big_a * (pmc - aaa) __a : Optional[int] = ppmc + aaa __a : List[Any] = -2 * pmpc __a : List[str] = ppmc - aaa __a : Dict = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def __magic_name__ ( _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : float , _lowerCamelCase : float = 1 / sqrt(2 ) , ): __a : str = tau * frequency / samplerate __a : Dict = sin(_lowerCamelCase ) __a : List[str] = cos(_lowerCamelCase ) __a : Tuple = _sin / (2 * q_factor) __a : Optional[int] = 1_0 ** (gain_db / 4_0) __a : str = (big_a + 1) - (big_a - 1) * _cos __a : Optional[int] = (big_a + 1) + (big_a - 1) * _cos __a : Optional[Any] = (big_a - 1) - (big_a + 1) * _cos __a : Dict = (big_a - 1) + (big_a + 1) * _cos __a : Optional[Any] = 2 * sqrt(_lowerCamelCase ) * alpha __a : Union[str, Any] = big_a * (ppmc + aaa) __a : Tuple = -2 * big_a * pmpc __a : str = big_a * (ppmc - aaa) __a : List[str] = pmc + aaa __a : Tuple = 2 * mpc __a : List[str] = pmc - aaa __a : Tuple = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt
720
"""simple docstring""" import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TextClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. lowercase__ = {"LayoutLMv2Config", "LayoutLMv3Config"} @is_pipeline_test class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): _lowerCAmelCase = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING _lowerCAmelCase = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: _lowerCAmelCase = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: _lowerCAmelCase = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } @require_torch def lowerCAmelCase__(self ): '''simple docstring''' __a : int = pipeline( task="""text-classification""" , model="""hf-internal-testing/tiny-random-distilbert""" , framework="""pt""" ) __a : Tuple = text_classifier("""This is great !""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """LABEL_0""", """score""": 0.504}] ) __a : Optional[Any] = text_classifier("""This is great !""" , top_k=2 ) self.assertEqual( nested_simplify(_lowercase ) , [{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}] ) __a : int = text_classifier(["""This is great !""", """This is bad"""] , top_k=2 ) self.assertEqual( nested_simplify(_lowercase ) , [ [{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}], [{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}], ] , ) __a : List[str] = text_classifier("""This is great !""" , top_k=1 ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """LABEL_0""", """score""": 0.504}] ) # Legacy behavior __a : Optional[int] = text_classifier("""This is great !""" , return_all_scores=_lowercase ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """LABEL_0""", """score""": 0.504}] ) __a : Tuple = text_classifier("""This is great !""" , return_all_scores=_lowercase ) self.assertEqual( nested_simplify(_lowercase ) , [[{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}]] ) __a : Any = text_classifier(["""This is great !""", """Something else"""] , return_all_scores=_lowercase ) self.assertEqual( nested_simplify(_lowercase ) , [ [{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}], [{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}], ] , ) __a : Union[str, Any] = text_classifier(["""This is great !""", """Something else"""] , return_all_scores=_lowercase ) self.assertEqual( nested_simplify(_lowercase ) , [ {"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_0""", """score""": 0.504}, ] , ) @require_torch def lowerCAmelCase__(self ): '''simple docstring''' import torch __a : Any = pipeline( task="""text-classification""" , model="""hf-internal-testing/tiny-random-distilbert""" , framework="""pt""" , device=torch.device("""cpu""" ) , ) __a : Optional[int] = text_classifier("""This is great !""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """LABEL_0""", """score""": 0.504}] ) @require_tf def lowerCAmelCase__(self ): '''simple docstring''' __a : List[Any] = pipeline( task="""text-classification""" , model="""hf-internal-testing/tiny-random-distilbert""" , framework="""tf""" ) __a : List[str] = text_classifier("""This is great !""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """LABEL_0""", """score""": 0.504}] ) @slow @require_torch def lowerCAmelCase__(self ): '''simple docstring''' __a : Tuple = pipeline("""text-classification""" ) __a : Tuple = text_classifier("""This is great !""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """POSITIVE""", """score""": 1.0}] ) __a : Optional[int] = text_classifier("""This is bad !""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """NEGATIVE""", """score""": 1.0}] ) __a : Union[str, Any] = text_classifier("""Birds are a type of animal""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """POSITIVE""", """score""": 0.988}] ) @slow @require_tf def lowerCAmelCase__(self ): '''simple docstring''' __a : List[str] = pipeline("""text-classification""" , framework="""tf""" ) __a : str = text_classifier("""This is great !""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """POSITIVE""", """score""": 1.0}] ) __a : Tuple = text_classifier("""This is bad !""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """NEGATIVE""", """score""": 1.0}] ) __a : str = text_classifier("""Birds are a type of animal""" ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": """POSITIVE""", """score""": 0.988}] ) def lowerCAmelCase__(self , _lowercase , _lowercase , _lowercase ): '''simple docstring''' __a : Dict = TextClassificationPipeline(model=_lowercase , tokenizer=_lowercase ) return text_classifier, ["HuggingFace is in", "This is another test"] def lowerCAmelCase__(self , _lowercase , _lowercase ): '''simple docstring''' __a : List[str] = text_classifier.model # Small inputs because BartTokenizer tiny has maximum position embeddings = 22 __a : Union[str, Any] = """HuggingFace is in""" __a : List[str] = text_classifier(_lowercase ) self.assertEqual(nested_simplify(_lowercase ) , [{"""label""": ANY(_lowercase ), """score""": ANY(_lowercase )}] ) self.assertTrue(outputs[0]["""label"""] in model.config.idalabel.values() ) __a : Optional[int] = ["""HuggingFace is in """, """Paris is in France"""] __a : Dict = text_classifier(_lowercase ) self.assertEqual( nested_simplify(_lowercase ) , [{"""label""": ANY(_lowercase ), """score""": ANY(_lowercase )}, {"""label""": ANY(_lowercase ), """score""": ANY(_lowercase )}] , ) self.assertTrue(outputs[0]["""label"""] in model.config.idalabel.values() ) self.assertTrue(outputs[1]["""label"""] in model.config.idalabel.values() ) # Forcing to get all results with `top_k=None` # This is NOT the legacy format __a : Dict = text_classifier(_lowercase , top_k=_lowercase ) __a : Dict = len(model.config.idalabel.values() ) self.assertEqual( nested_simplify(_lowercase ) , [[{"""label""": ANY(_lowercase ), """score""": ANY(_lowercase )}] * N, [{"""label""": ANY(_lowercase ), """score""": ANY(_lowercase )}] * N] , ) __a : Dict = {"""text""": """HuggingFace is in """, """text_pair""": """Paris is in France"""} __a : Any = text_classifier(_lowercase ) self.assertEqual( nested_simplify(_lowercase ) , {"""label""": ANY(_lowercase ), """score""": ANY(_lowercase )} , ) self.assertTrue(outputs["""label"""] in model.config.idalabel.values() ) # This might be used a text pair, but tokenizer + pipe interaction # makes it hard to understand that it's not using the pair properly # https://github.com/huggingface/transformers/issues/17305 # We disabled this usage instead as it was outputting wrong outputs. __a : Dict = [["""HuggingFace is in """, """Paris is in France"""]] with self.assertRaises(_lowercase ): text_classifier(_lowercase ) # This used to be valid for doing text pairs # We're keeping it working because of backward compatibility __a : Optional[int] = text_classifier([[["""HuggingFace is in """, """Paris is in France"""]]] ) self.assertEqual( nested_simplify(_lowercase ) , [{"""label""": ANY(_lowercase ), """score""": ANY(_lowercase )}] , ) self.assertTrue(outputs[0]["""label"""] in model.config.idalabel.values() )
63
0
"""simple docstring""" import tempfile import unittest from make_student import create_student_by_copying_alternating_layers from transformers import AutoConfig from transformers.file_utils import cached_property from transformers.testing_utils import require_torch lowercase__ = "sshleifer/bart-tiny-random" lowercase__ = "patrickvonplaten/t5-tiny-random" @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @cached_property def lowerCAmelCase__(self ): '''simple docstring''' return AutoConfig.from_pretrained(_lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Any = create_student_by_copying_alternating_layers(_lowercase , tempfile.mkdtemp() , e=1 , d=1 ) self.assertEqual(student.config.num_hidden_layers , 1 ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Tuple = create_student_by_copying_alternating_layers(_lowercase , tempfile.mkdtemp() , e=1 , d=_lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Optional[int] = create_student_by_copying_alternating_layers(_lowercase , tempfile.mkdtemp() , e=1 , d=_lowercase ) self.assertEqual(student.config.encoder_layers , 1 ) self.assertEqual(student.config.decoder_layers , self.teacher_config.encoder_layers ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Tuple = create_student_by_copying_alternating_layers(_lowercase , tempfile.mkdtemp() , e=1 , d=1 ) self.assertEqual(student.config.encoder_layers , 1 ) self.assertEqual(student.config.decoder_layers , 1 ) def lowerCAmelCase__(self ): '''simple docstring''' with self.assertRaises(_lowercase ): create_student_by_copying_alternating_layers(_lowercase , tempfile.mkdtemp() , e=_lowercase , d=_lowercase )
721
"""simple docstring""" import unittest from knapsack import knapsack as k class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def lowerCAmelCase__(self ): '''simple docstring''' __a : str = 0 __a : Optional[Any] = [0] __a : int = [0] __a : str = len(_lowercase ) self.assertEqual(k.knapsack(_lowercase , _lowercase , _lowercase , _lowercase ) , 0 ) __a : int = [60] __a : Union[str, Any] = [10] __a : Tuple = len(_lowercase ) self.assertEqual(k.knapsack(_lowercase , _lowercase , _lowercase , _lowercase ) , 0 ) def lowerCAmelCase__(self ): '''simple docstring''' __a : int = 3 __a : str = [1, 2, 3] __a : Optional[Any] = [3, 2, 1] __a : int = len(_lowercase ) self.assertEqual(k.knapsack(_lowercase , _lowercase , _lowercase , _lowercase ) , 5 ) def lowerCAmelCase__(self ): '''simple docstring''' __a : Dict = 50 __a : Tuple = [60, 100, 120] __a : List[str] = [10, 20, 30] __a : Union[str, Any] = len(_lowercase ) self.assertEqual(k.knapsack(_lowercase , _lowercase , _lowercase , _lowercase ) , 220 ) if __name__ == "__main__": unittest.main()
63
0
from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_torch_available, is_torch_tpu_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_torch_available(): import torch if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm A_ : Optional[int] = logging.get_logger(__name__) @dataclass class _a (__magic_name__ ): '''simple docstring''' UpperCAmelCase__: Optional[Any] = [ '''no_inference''', '''no_cuda''', '''no_tpu''', '''no_speed''', '''no_memory''', '''no_env_print''', '''no_multi_process''', ] def __init__( self , **A__ ): for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: A__ : List[Any] = deprecated_arg[3:] setattr(self , A__ , not kwargs.pop(A__ ) ) logger.warning( F"""{deprecated_arg} is depreciated. Please use --no_{positive_arg} or""" F""" {positive_arg}={kwargs[positive_arg]}""" ) A__ : Union[str, Any] = kwargs.pop("""torchscript""" , self.torchscript ) A__ : Union[str, Any] = kwargs.pop("""torch_xla_tpu_print_metrics""" , self.torch_xla_tpu_print_metrics ) A__ : Optional[int] = kwargs.pop("""fp16_opt_level""" , self.fpaa_opt_level ) super().__init__(**A__ ) UpperCAmelCase__: bool = field(default=__magic_name__ , metadata={'''help''': '''Trace the models using torchscript'''} ) UpperCAmelCase__: bool = field(default=__magic_name__ , metadata={'''help''': '''Print Xla/PyTorch tpu metrics'''} ) UpperCAmelCase__: str = field( default='''O1''' , metadata={ '''help''': ( '''For fp16: Apex AMP optimization level selected in [\'O0\', \'O1\', \'O2\', and \'O3\']. ''' '''See details at https://nvidia.github.io/apex/amp.html''' ) } , ) @cached_property def __A ( self ): requires_backends(self , ["""torch"""] ) logger.info("""PyTorch: setting up devices""" ) if not self.cuda: A__ : int = torch.device("""cpu""" ) A__ : Dict = 0 elif is_torch_tpu_available(): A__ : Union[str, Any] = xm.xla_device() A__ : str = 0 else: A__ : Optional[int] = torch.device("""cuda""" if torch.cuda.is_available() else """cpu""" ) A__ : str = torch.cuda.device_count() return device, n_gpu @property def __A ( self ): return is_torch_tpu_available() and self.tpu @property def __A ( self ): requires_backends(self , ["""torch"""] ) # TODO(PVP): currently only single GPU is supported return torch.cuda.current_device() @property def __A ( self ): requires_backends(self , ["""torch"""] ) return self._setup_devices[0] @property def __A ( self ): requires_backends(self , ["""torch"""] ) return self._setup_devices[1] @property def __A ( self ): return self.n_gpu > 0
64
import os try: from .build_directory_md import good_file_paths except ImportError: from build_directory_md import good_file_paths # type: ignore A_ : Dict = list(good_file_paths()) assert filepaths, "good_file_paths() failed!" A_ : Optional[Any] = [file for file in filepaths if file != file.lower()] if upper_files: print(f'''{len(upper_files)} files contain uppercase characters:''') print('\n'.join(upper_files) + '\n') A_ : Tuple = [file for file in filepaths if ' ' in file] if space_files: print(f'''{len(space_files)} files contain space characters:''') print('\n'.join(space_files) + '\n') A_ : Any = [file for file in filepaths if '-' in file] if hyphen_files: print(f'''{len(hyphen_files)} files contain hyphen characters:''') print('\n'.join(hyphen_files) + '\n') A_ : List[str] = [file for file in filepaths if os.sep not in file] if nodir_files: print(f'''{len(nodir_files)} files are not in a directory:''') print('\n'.join(nodir_files) + '\n') A_ : Any = len(upper_files + space_files + hyphen_files + nodir_files) if bad_files: import sys sys.exit(bad_files)
64
1
import inspect import logging import os import random import shutil import tempfile import unittest import pytest import torch from torch import nn from torch.utils.data import DataLoader, TensorDataset from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_cuda from accelerate.utils import ProjectConfiguration, set_seed A_ : Any = logging.getLogger(__name__) def UpperCamelCase (lowercase_: Optional[Any]=2 , lowercase_: Union[str, Any]=3 , lowercase_: int=16 , lowercase_: int = 10 , lowercase_: int = 2 ) -> int: def get_dataset(lowercase_: Optional[int] ): A__ : Optional[Any] = torch.randn(batch_size * n_batches , 1 ) return TensorDataset(lowercase_ , a * x + b + 0.1 * torch.randn(batch_size * n_batches , 1 ) ) A__ : Dict = get_dataset(lowercase_ ) A__ : Any = get_dataset(lowercase_ ) A__ : Dict = DataLoader(lowercase_ , shuffle=lowercase_ , batch_size=lowercase_ , num_workers=4 ) A__ : Optional[Any] = DataLoader(lowercase_ , shuffle=lowercase_ , batch_size=lowercase_ , num_workers=4 ) return (train_dataloader, valid_dataloader) def UpperCamelCase (lowercase_: Optional[Any] , lowercase_: List[str] , lowercase_: int , lowercase_: int , lowercase_: List[str] , lowercase_: Dict=None ) -> List[Any]: A__ : List[Any] = [] for epoch in range(lowercase_ ): # Train quickly model.train() for batch in dataloader: A__ , A__ : Any = batch A__ : Any = model(lowercase_ ) A__ : Any = torch.nn.functional.mse_loss(lowercase_ , lowercase_ ) accelerator.backward(lowercase_ ) optimizer.step() optimizer.zero_grad() rands.append(random.random() ) # Introduce some randomness if scheduler is not None: scheduler.step() return rands class _a (nn.Module ): '''simple docstring''' def __init__( self ): super().__init__() A__ : str = nn.Parameter(torch.randn(1 ) ) A__ : Any = nn.Parameter(torch.randn(1 ) ) def __A ( self , A__ ): return x * self.a + self.b class _a (unittest.TestCase ): '''simple docstring''' def __A ( self ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) A__ : Optional[Any] = DummyModel() A__ : Optional[Any] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) A__ , A__ : str = dummy_dataloaders() A__ : Dict = ProjectConfiguration(total_limit=1 , project_dir=A__ , automatic_checkpoint_naming=A__ ) # Train baseline A__ : List[str] = Accelerator(project_config=A__ ) A__ , A__ , A__ , A__ : Any = accelerator.prepare( A__ , A__ , A__ , A__ ) # Save initial accelerator.save_state() # Save second state accelerator.save_state() self.assertEqual(len(os.listdir(accelerator.project_dir ) ) , 1 ) def __A ( self ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) A__ : str = DummyModel() A__ : Optional[int] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) A__ , A__ : int = dummy_dataloaders() # Train baseline A__ : str = Accelerator() A__ , A__ , A__ , A__ : List[str] = accelerator.prepare( A__ , A__ , A__ , A__ ) # Save initial A__ : List[Any] = os.path.join(A__ , """initial""" ) accelerator.save_state(A__ ) ((A__) , (A__)) : str = model.a.item(), model.b.item() A__ : Dict = optimizer.state_dict() A__ : List[str] = train(3 , A__ , A__ , A__ , A__ ) ((A__) , (A__)) : str = model.a.item(), model.b.item() A__ : Any = optimizer.state_dict() # Train partially set_seed(42 ) A__ : Optional[int] = DummyModel() A__ : Dict = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) A__ , A__ : Dict = dummy_dataloaders() A__ : List[str] = Accelerator() A__ , A__ , A__ , A__ : Optional[Any] = accelerator.prepare( A__ , A__ , A__ , A__ ) accelerator.load_state(A__ ) ((A__) , (A__)) : Tuple = model.a.item(), model.b.item() A__ : Union[str, Any] = optimizer.state_dict() self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) A__ : List[str] = train(2 , A__ , A__ , A__ , A__ ) # Save everything A__ : Optional[int] = os.path.join(A__ , """checkpoint""" ) accelerator.save_state(A__ ) # Load everything back in and make sure all states work accelerator.load_state(A__ ) test_rands += train(1 , A__ , A__ , A__ , A__ ) ((A__) , (A__)) : Union[str, Any] = model.a.item(), model.b.item() A__ : Optional[int] = optimizer.state_dict() self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) def __A ( self ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) A__ : int = DummyModel() A__ : Optional[int] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) A__ , A__ : List[str] = dummy_dataloaders() A__ : str = ProjectConfiguration(automatic_checkpoint_naming=A__ ) # Train baseline A__ : Any = Accelerator(project_dir=A__ , project_config=A__ ) A__ , A__ , A__ , A__ : str = accelerator.prepare( A__ , A__ , A__ , A__ ) # Save initial accelerator.save_state() ((A__) , (A__)) : Tuple = model.a.item(), model.b.item() A__ : int = optimizer.state_dict() A__ : int = train(3 , A__ , A__ , A__ , A__ ) ((A__) , (A__)) : Optional[Any] = model.a.item(), model.b.item() A__ : Any = optimizer.state_dict() # Train partially set_seed(42 ) A__ : Dict = DummyModel() A__ : List[Any] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) A__ , A__ : Union[str, Any] = dummy_dataloaders() A__ : List[Any] = ProjectConfiguration(iteration=1 , automatic_checkpoint_naming=A__ ) A__ : Dict = Accelerator(project_dir=A__ , project_config=A__ ) A__ , A__ , A__ , A__ : Union[str, Any] = accelerator.prepare( A__ , A__ , A__ , A__ ) accelerator.load_state(os.path.join(A__ , """checkpoints""" , """checkpoint_0""" ) ) ((A__) , (A__)) : Optional[int] = model.a.item(), model.b.item() A__ : Tuple = optimizer.state_dict() self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) A__ : str = train(2 , A__ , A__ , A__ , A__ ) # Save everything accelerator.save_state() # Load everything back in and make sure all states work accelerator.load_state(os.path.join(A__ , """checkpoints""" , """checkpoint_1""" ) ) test_rands += train(1 , A__ , A__ , A__ , A__ ) ((A__) , (A__)) : Optional[int] = model.a.item(), model.b.item() A__ : List[Any] = optimizer.state_dict() self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) def __A ( self ): A__ : Union[str, Any] = torch.tensor([1, 2, 3] ) A__ : int = torch.tensor([2, 3, 4] ) A__ : List[Any] = DummyModel() A__ : List[Any] = torch.optim.Adam(net.parameters() ) A__ : Tuple = Accelerator() with self.assertRaises(A__ ) as ve: accelerator.register_for_checkpointing(A__ , A__ , A__ , A__ ) A__ : Any = str(ve.exception ) self.assertTrue("""Item at index 0""" in message ) self.assertTrue("""Item at index 1""" in message ) self.assertFalse("""Item at index 2""" in message ) self.assertFalse("""Item at index 3""" in message ) def __A ( self ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) A__ : Any = DummyModel() A__ : Union[str, Any] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) A__ : Dict = torch.optim.lr_scheduler.StepLR(A__ , step_size=1 , gamma=0.9_9 ) A__ , A__ : List[Any] = dummy_dataloaders() A__ : Tuple = ProjectConfiguration(automatic_checkpoint_naming=A__ ) # Train baseline A__ : Optional[Any] = Accelerator(project_dir=A__ , project_config=A__ ) A__ , A__ , A__ , A__ , A__ : Union[str, Any] = accelerator.prepare( A__ , A__ , A__ , A__ , A__ ) # Save initial accelerator.save_state() A__ : Tuple = scheduler.state_dict() train(3 , A__ , A__ , A__ , A__ , A__ ) self.assertNotEqual(A__ , scheduler.state_dict() ) # Load everything back in and make sure all states work accelerator.load_state(os.path.join(A__ , """checkpoints""" , """checkpoint_0""" ) ) self.assertEqual(A__ , scheduler.state_dict() ) def __A ( self ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) A__ : Optional[Any] = DummyModel() A__ : int = ProjectConfiguration(automatic_checkpoint_naming=A__ , total_limit=2 ) # Train baseline A__ : List[str] = Accelerator(project_dir=A__ , project_config=A__ ) A__ : Union[str, Any] = accelerator.prepare(A__ ) # Save 3 states: for _ in range(11 ): accelerator.save_state() self.assertTrue(not os.path.exists(os.path.join(A__ , """checkpoints""" , """checkpoint_0""" ) ) ) self.assertTrue(os.path.exists(os.path.join(A__ , """checkpoints""" , """checkpoint_9""" ) ) ) self.assertTrue(os.path.exists(os.path.join(A__ , """checkpoints""" , """checkpoint_10""" ) ) ) @require_cuda def __A ( self ): A__ : Dict = ["""torchrun""", F"""--nproc_per_node={torch.cuda.device_count()}""", inspect.getfile(self.__class__ )] execute_subprocess_async(A__ , env=os.environ.copy() ) if __name__ == "__main__": A_ : List[str] = '/tmp/accelerate/state_checkpointing' A_ : Optional[Any] = DummyModel() A_ : Union[str, Any] = torch.optim.Adam(params=model.parameters(), lr=1E-3) A_ : str = torch.optim.lr_scheduler.StepLR(optimizer, step_size=1, gamma=0.99) A_ , A_ : List[Any] = dummy_dataloaders() A_ : int = ProjectConfiguration(automatic_checkpoint_naming=True) # Train baseline A_ : List[str] = Accelerator(project_dir=savedir, project_config=project_config, mixed_precision='no') if accelerator.process_index == 0: if os.path.exists(savedir): shutil.rmtree(savedir) os.makedirs(savedir) A_ , A_ , A_ , A_ , A_ : List[Any] = accelerator.prepare( model, optimizer, train_dataloader, valid_dataloader, scheduler ) A_ , A_ : Dict = accelerator.prepare(model, optimizer) train(3, model, train_dataloader, optimizer, accelerator, scheduler) # Check that the intial optimizer is loaded on the GPU for group in optimizer.param_groups: A_ : str = group['params'][0].device break assert param_device.type == accelerator.device.type A_ : Optional[Any] = model.cpu() accelerator.wait_for_everyone() accelerator.save_state() accelerator.wait_for_everyone() # Check CPU state accelerator.load_state(os.path.join(savedir, 'checkpoints', 'checkpoint_0'), map_location='cpu') for group in optimizer.param_groups: A_ : str = group['params'][0].device break assert ( param_device.type == torch.device('cpu').type ), f"Loaded optimizer states did not match, expected to be loaded on the CPU but got {param_device}" # Check device state model.to(accelerator.device) accelerator.load_state(os.path.join(savedir, 'checkpoints', 'checkpoint_0'), map_location='on_device') for group in optimizer.param_groups: A_ : Tuple = group['params'][0].device break assert ( param_device.type == accelerator.device.type ), f"Loaded optimizer states did not match, expected to be loaded on {accelerator.device} but got {param_device}" # Check error with pytest.raises(TypeError, match='Unsupported optimizer map location passed'): accelerator.load_state(os.path.join(savedir, 'checkpoints', 'checkpoint_0'), map_location='invalid') accelerator.wait_for_everyone() if accelerator.process_index == 0: shutil.rmtree(savedir) accelerator.wait_for_everyone()
64
import inspect import warnings from typing import Any, Dict, Optional, Union from packaging import version def UpperCamelCase (*lowercase_: Optional[int] , lowercase_: Optional[Union[Dict, Any]] = None , lowercase_: Dict=True , lowercase_: Tuple=2 ) -> Dict: from .. import __version__ A__ : Dict = take_from A__ : str = () if not isinstance(args[0] , lowercase_ ): A__ : int = (args,) for attribute, version_name, message in args: if version.parse(version.parse(lowercase_ ).base_version ) >= version.parse(lowercase_ ): raise ValueError( f"""The deprecation tuple {(attribute, version_name, message)} should be removed since diffusers'""" f""" version {__version__} is >= {version_name}""" ) A__ : Any = None if isinstance(lowercase_ , lowercase_ ) and attribute in deprecated_kwargs: values += (deprecated_kwargs.pop(lowercase_ ),) A__ : List[str] = f"""The `{attribute}` argument is deprecated and will be removed in version {version_name}.""" elif hasattr(lowercase_ , lowercase_ ): values += (getattr(lowercase_ , lowercase_ ),) A__ : Optional[Any] = f"""The `{attribute}` attribute is deprecated and will be removed in version {version_name}.""" elif deprecated_kwargs is None: A__ : int = f"""`{attribute}` is deprecated and will be removed in version {version_name}.""" if warning is not None: A__ : int = warning + """ """ if standard_warn else """""" warnings.warn(warning + message , lowercase_ , stacklevel=lowercase_ ) if isinstance(lowercase_ , lowercase_ ) and len(lowercase_ ) > 0: A__ : Union[str, Any] = inspect.getouterframes(inspect.currentframe() )[1] A__ : Optional[Any] = call_frame.filename A__ : Optional[int] = call_frame.lineno A__ : Any = call_frame.function A__ , A__ : List[str] = next(iter(deprecated_kwargs.items() ) ) raise TypeError(f"""{function} in {filename} line {line_number-1} got an unexpected keyword argument `{key}`""" ) if len(lowercase_ ) == 0: return elif len(lowercase_ ) == 1: return values[0] return values
64
1
import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, PerceiverTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): A_ : Tuple = 'pt' elif is_tf_available(): A_ : Optional[int] = 'tf' else: A_ : str = 'jax' class _a (__magic_name__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__: Dict = PerceiverTokenizer UpperCAmelCase__: str = False def __A ( self ): super().setUp() A__ : int = PerceiverTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def __A ( self ): return PerceiverTokenizer.from_pretrained("""deepmind/language-perceiver""" ) def __A ( self , **A__ ): return self.tokenizer_class.from_pretrained(self.tmpdirname , **A__ ) def __A ( self , A__ , A__=False , A__=20 , A__=5 ): # XXX The default common tokenizer tests assume that every ID is decodable on its own. # This assumption is invalid for Perceiver because single bytes might not be # valid utf-8 (byte 128 for instance). # Here we're overriding the smallest possible method to provide # a clean sequence without making the same assumption. A__ : int = [] for i in range(len(A__ ) ): try: A__ : List[Any] = tokenizer.decode([i] , clean_up_tokenization_spaces=A__ ) except UnicodeDecodeError: pass toks.append((i, tok) ) A__ : Dict = list(filter(lambda A__ : re.match(r"""^[ a-zA-Z]+$""" , t[1] ) , A__ ) ) A__ : str = list(filter(lambda A__ : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=A__ ) , A__ ) ) if max_length is not None and len(A__ ) > max_length: A__ : List[Any] = toks[:max_length] if min_length is not None and len(A__ ) < min_length and len(A__ ) > 0: while len(A__ ) < min_length: A__ : List[Any] = toks + toks # toks_str = [t[1] for t in toks] A__ : Tuple = [t[0] for t in toks] # Ensure consistency A__ : Union[str, Any] = tokenizer.decode(A__ , clean_up_tokenization_spaces=A__ ) if " " not in output_txt and len(A__ ) > 1: A__ : List[Any] = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=A__ ) + """ """ + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=A__ ) ) if with_prefix_space: A__ : List[Any] = """ """ + output_txt A__ : Optional[Any] = tokenizer.encode(A__ , add_special_tokens=A__ ) return output_txt, output_ids def __A ( self ): A__ : Union[str, Any] = self.perceiver_tokenizer A__ : str = """Unicode €.""" A__ : Union[str, Any] = tokenizer(A__ ) A__ : Any = [4, 91, 116, 111, 105, 117, 106, 107, 38, 232, 136, 178, 52, 5] self.assertEqual(encoded["""input_ids"""] , A__ ) # decoding A__ : List[Any] = tokenizer.decode(A__ ) self.assertEqual(A__ , """[CLS]Unicode €.[SEP]""" ) A__ : Any = tokenizer("""e è é ê ë""" ) A__ : str = [4, 107, 38, 201, 174, 38, 201, 175, 38, 201, 176, 38, 201, 177, 5] self.assertEqual(encoded["""input_ids"""] , A__ ) # decoding A__ : Tuple = tokenizer.decode(A__ ) self.assertEqual(A__ , """[CLS]e è é ê ë[SEP]""" ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode("""e è é ê ë""" ) ) , """[CLS]e è é ê ë[SEP]""" ) def __A ( self ): A__ : int = self.perceiver_tokenizer A__ : Optional[int] = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] # fmt: off A__ : str = [4, 71, 38, 114, 117, 116, 109, 38, 118, 103, 120, 103, 109, 120, 103, 118, 110, 38, 108, 117, 120, 38, 121, 123, 115, 115, 103, 120, 111, 128, 103, 122, 111, 117, 116, 52, 5, 0] # fmt: on A__ : Union[str, Any] = tokenizer(A__ , padding=A__ , return_tensors=A__ ) self.assertIsInstance(A__ , A__ ) if FRAMEWORK != "jax": A__ : Dict = list(batch.input_ids.numpy()[0] ) else: A__ : List[Any] = list(batch.input_ids.tolist()[0] ) self.assertListEqual(A__ , A__ ) self.assertEqual((2, 38) , batch.input_ids.shape ) self.assertEqual((2, 38) , batch.attention_mask.shape ) def __A ( self ): A__ : Tuple = self.perceiver_tokenizer A__ : Tuple = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] A__ : List[str] = tokenizer(A__ , padding=A__ , return_tensors=A__ ) # check if input_ids are returned and no decoder_input_ids self.assertIn("""input_ids""" , A__ ) self.assertIn("""attention_mask""" , A__ ) self.assertNotIn("""decoder_input_ids""" , A__ ) self.assertNotIn("""decoder_attention_mask""" , A__ ) def __A ( self ): A__ : Optional[int] = self.perceiver_tokenizer A__ : Optional[Any] = [ """Summary of the text.""", """Another summary.""", ] A__ : Optional[Any] = tokenizer( text_target=A__ , max_length=32 , padding="""max_length""" , truncation=A__ , return_tensors=A__ ) self.assertEqual(32 , targets["""input_ids"""].shape[1] ) def __A ( self ): # safety check on max_len default value so we are sure the test works A__ : List[str] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): self.assertNotEqual(tokenizer.model_max_length , 42 ) # Now let's start the test A__ : Optional[int] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc A__ : Tuple = tempfile.mkdtemp() A__ : Tuple = """ He is very happy, UNwant\u00E9d,running""" A__ : Optional[Any] = tokenizer.encode(A__ , add_special_tokens=A__ ) tokenizer.save_pretrained(A__ ) A__ : List[str] = tokenizer.__class__.from_pretrained(A__ ) A__ : int = after_tokenizer.encode(A__ , add_special_tokens=A__ ) self.assertListEqual(A__ , A__ ) shutil.rmtree(A__ ) A__ : List[str] = self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc A__ : Optional[Any] = tempfile.mkdtemp() A__ : Any = """ He is very happy, UNwant\u00E9d,running""" tokenizer.add_tokens(["""bim""", """bambam"""] ) A__ : Optional[Any] = tokenizer.additional_special_tokens additional_special_tokens.append("""new_additional_special_token""" ) tokenizer.add_special_tokens({"""additional_special_tokens""": additional_special_tokens} ) A__ : Optional[Any] = tokenizer.encode(A__ , add_special_tokens=A__ ) tokenizer.save_pretrained(A__ ) A__ : List[Any] = tokenizer.__class__.from_pretrained(A__ ) A__ : Optional[int] = after_tokenizer.encode(A__ , add_special_tokens=A__ ) self.assertListEqual(A__ , A__ ) self.assertIn("""new_additional_special_token""" , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 42 ) A__ : Dict = tokenizer.__class__.from_pretrained(A__ , model_max_length=43 ) self.assertEqual(tokenizer.model_max_length , 43 ) shutil.rmtree(A__ ) def __A ( self ): A__ : Dict = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(A__ ) with open(os.path.join(A__ , """special_tokens_map.json""" ) , encoding="""utf-8""" ) as json_file: A__ : int = json.load(A__ ) with open(os.path.join(A__ , """tokenizer_config.json""" ) , encoding="""utf-8""" ) as json_file: A__ : Tuple = json.load(A__ ) A__ : str = [F"""<extra_id_{i}>""" for i in range(125 )] A__ : Optional[Any] = added_tokens_extra_ids + [ """an_additional_special_token""" ] A__ : Dict = added_tokens_extra_ids + [ """an_additional_special_token""" ] with open(os.path.join(A__ , """special_tokens_map.json""" ) , """w""" , encoding="""utf-8""" ) as outfile: json.dump(A__ , A__ ) with open(os.path.join(A__ , """tokenizer_config.json""" ) , """w""" , encoding="""utf-8""" ) as outfile: json.dump(A__ , A__ ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files A__ : List[str] = tokenizer_class.from_pretrained( A__ , ) self.assertIn( """an_additional_special_token""" , tokenizer_without_change_in_init.additional_special_tokens ) self.assertEqual( ["""an_additional_special_token"""] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(["""an_additional_special_token"""] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained A__ : Tuple = added_tokens_extra_ids + [AddedToken("""a_new_additional_special_token""" , lstrip=A__ )] A__ : Tuple = tokenizer_class.from_pretrained( A__ , additional_special_tokens=A__ , ) self.assertIn("""a_new_additional_special_token""" , tokenizer.additional_special_tokens ) self.assertEqual( ["""a_new_additional_special_token"""] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(["""a_new_additional_special_token"""] ) ) , ) def __A ( self ): A__ : Any = self.perceiver_tokenizer self.assertEqual(tokenizer.decode([178] ) , """�""" ) def __A ( self ): pass def __A ( self ): pass def __A ( self ): pass def __A ( self ): pass def __A ( self ): # The default common tokenizer tests uses invalid tokens for Perceiver that can only accept one-character # strings and special added tokens as tokens A__ : int = self.get_tokenizers(fast=A__ , do_lower_case=A__ ) for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): A__ : Tuple = ["""[CLS]""", """t""", """h""", """i""", """s""", """ """, """i""", """s""", """ """, """a""", """ """, """t""", """e""", """s""", """t""", """[SEP]"""] A__ : Dict = tokenizer.convert_tokens_to_string(A__ ) self.assertIsInstance(A__ , A__ )
64
import argparse import json import os import tensorstore as ts import torch from flax import serialization from flax.traverse_util import flatten_dict, unflatten_dict from tensorflow.io import gfile from transformers.modeling_utils import dtype_byte_size from transformers.models.switch_transformers.convert_switch_transformers_original_flax_checkpoint_to_pytorch import ( rename_keys, ) from transformers.utils import WEIGHTS_INDEX_NAME, WEIGHTS_NAME from transformers.utils.hub import convert_file_size_to_int def UpperCamelCase (lowercase_: List[str] , lowercase_: str ) -> Optional[Any]: if flax_key_tuple[-1] == "kernel" and flax_tensor.ndim == 3: # expert layer A__ : Union[str, Any] = flax_key_tuple[:-1] + ("""weight""",) A__ : Optional[int] = torch.permute(lowercase_ , (0, 2, 1) ) elif flax_key_tuple[-1] == "kernel" and ".".join(lowercase_ ): # linear layer A__ : Optional[Any] = flax_key_tuple[:-1] + ("""weight""",) A__ : int = flax_tensor.T elif flax_key_tuple[-1] in ["scale", "embedding"]: A__ : Optional[int] = flax_key_tuple[:-1] + ("""weight""",) return flax_key_tuple, flax_tensor def UpperCamelCase (lowercase_: Tuple , lowercase_: Optional[int] , lowercase_: str ) -> Union[str, Any]: if "metadata" in layer: A__ : Tuple = layer.split("""metadata""" ) A__ : Optional[Any] = """""".join(split_layer[0] )[:-1] A__ : Optional[Any] = [tuple(("""metadata""" + split_layer[1]).split("""/""" ) )] elif "kvstore" in layer: A__ : str = layer.split("""kvstore""" ) A__ : int = """""".join(split_layer[0] )[:-1] A__ : Optional[int] = [tuple(("""kvstore""" + split_layer[1]).split("""/""" ) )] else: A__ : Any = layer.split("""/""" ) A__ : int = """/""".join(split_layer[:-1] ) A__ : str = (split_layer[-1],) if "kvstore/path" in layer: A__ : Dict = f"""{switch_checkpoint_path}/{checkpoint_info[layer]}""" elif "kvstore/driver" in layer: A__ : Optional[int] = """file""" else: A__ : str = checkpoint_info[layer] return curr_real_layer_name, split_layer, content def UpperCamelCase (lowercase_: str , lowercase_: List[Any] ) -> int: A__ : int = rename_keys(lowercase_ ) A__ : Any = {} for k, v in current_block.items(): A__ : Dict = v A__ : str = new_current_block torch.save(lowercase_ , lowercase_ ) def UpperCamelCase (lowercase_: Dict , lowercase_: Optional[Any] , lowercase_: Optional[Any] , lowercase_: Optional[int] , lowercase_: str = WEIGHTS_NAME ) -> Tuple: A__ : Optional[int] = convert_file_size_to_int(lowercase_ ) A__ : List[Any] = [] A__ : int = {} A__ : List[str] = 0 A__ : Any = 0 os.makedirs(lowercase_ , exist_ok=lowercase_ ) with gfile.GFile(switch_checkpoint_path + """/checkpoint""" , """rb""" ) as fp: A__ : Optional[Any] = serialization.msgpack_restore(fp.read() )["""optimizer"""]["""target"""] A__ : Dict = flatten_dict(lowercase_ , sep="""/""" ) A__ : Any = {} for layer in checkpoint_info.keys(): A__ , A__ , A__ : Union[str, Any] = get_key_and_tensorstore_dict( lowercase_ , lowercase_ , lowercase_ ) if curr_real_layer_name in all_layers: A__ : Optional[int] = content else: A__ : List[Any] = {split_layer[-1]: content} for key in all_layers.keys(): # open tensorstore file A__ : Optional[Any] = ts.open(unflatten_dict(all_layers[key] ) ).result().read().result() A__ : List[Any] = torch.tensor(lowercase_ ) A__ : List[Any] = raw_weights.numel() * dtype_byte_size(raw_weights.dtype ) # use the renaming pattern from the small conversion scripts A__ , A__ : Any = rename_base_flax_keys(tuple(key.split("""/""" ) ) , lowercase_ ) A__ : Any = """/""".join(lowercase_ ) # If this weight is going to tip up over the maximal size, we split. if current_block_size + weight_size > max_shard_size: A__ : List[Any] = os.path.join( lowercase_ , weights_name.replace(""".bin""" , f"""-{len(lowercase_ )+1:05d}-of-???.bin""" ) ) rename_and_save_block(lowercase_ , lowercase_ ) sharded_state_dicts.append(current_block.keys() ) del current_block A__ : Any = {} A__ : str = 0 A__ : List[str] = raw_weights.to(getattr(lowercase_ , lowercase_ ) ) current_block_size += weight_size total_size += weight_size # Add the last block A__ : Union[str, Any] = os.path.join(lowercase_ , weights_name.replace(""".bin""" , f"""-{len(lowercase_ )+1:05d}-of-???.bin""" ) ) rename_and_save_block(lowercase_ , lowercase_ ) sharded_state_dicts.append(current_block.keys() ) # If we only have one shard, we return it if len(lowercase_ ) == 1: return {weights_name: sharded_state_dicts[0]}, None # Otherwise, let's build the index A__ : str = {} A__ : Any = {} for idx, shard in enumerate(lowercase_ ): A__ : Any = weights_name.replace( """.bin""" , f"""-{idx+1:05d}-of-{len(lowercase_ ):05d}.bin""" ) # len(sharded_state_dicts):05d} A__ : Dict = os.path.join(lowercase_ , weights_name.replace(""".bin""" , f"""-{idx+1:05d}-of-???.bin""" ) ) os.rename(lowercase_ , os.path.join(lowercase_ , lowercase_ ) ) A__ : str = shard for key in shard: A__ : Any = shard_file # Add the metadata A__ : Tuple = {"""total_size""": total_size} A__ : Union[str, Any] = {"""metadata""": metadata, """weight_map""": weight_map} with open(os.path.join(lowercase_ , lowercase_ ) , """w""" , encoding="""utf-8""" ) as f: A__ : Dict = json.dumps(lowercase_ , indent=2 , sort_keys=lowercase_ ) + """\n""" f.write(lowercase_ ) return metadata, index if __name__ == "__main__": A_ : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--switch_t5x_checkpoint_path', default='/mnt/disks/disk_switch/original_checkpoints/switch-xxl-128/checkpoint_634600', type=str, required=False, help='Path to a directory containing a folder per layer. Follows the original Google format.', ) parser.add_argument('--max_shard_size', default='10GB', required=False, help='Max shard size') parser.add_argument('--dtype', default='bfloat16', type=str, required=False, help='dtype of the saved model') parser.add_argument( '--pytorch_dump_folder_path', default='/mnt/disks/disk_switch/original_checkpoints/switch-xxl-128-converted', type=str, required=False, help='Path to the output pytorch model.', ) A_ : Dict = parser.parse_args() shard_on_the_fly( args.switch_tax_checkpoint_path, args.pytorch_dump_folder_path, args.max_shard_size, args.dtype, ) def UpperCamelCase () -> int: from transformers import SwitchTransformersConfig, SwitchTransformersForConditionalGeneration, TaTokenizer A__ : str = SwitchTransformersConfig.from_pretrained("""google/switch-base-8""" ) config.save_pretrained("""/home/arthur_huggingface_co/transformers/switch_converted""" ) A__ : str = SwitchTransformersForConditionalGeneration.from_pretrained( """/home/arthur_huggingface_co/transformers/switch_converted""" , device_map="""auto""" ) A__ : Tuple = TaTokenizer.from_pretrained("""t5-small""" ) A__ : Dict = """A <extra_id_0> walks into a bar a orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.""" A__ : Union[str, Any] = tokenizer(lowercase_ , return_tensors="""pt""" ).input_ids A__ : Tuple = model.generate(lowercase_ , decoder_start_token_id=0 ) print(tokenizer.decode(out[0] ) )
64
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A_ : str = { 'configuration_blip_2': [ 'BLIP_2_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Blip2Config', 'Blip2QFormerConfig', 'Blip2VisionConfig', ], 'processing_blip_2': ['Blip2Processor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Dict = [ 'BLIP_2_PRETRAINED_MODEL_ARCHIVE_LIST', 'Blip2Model', 'Blip2QFormerModel', 'Blip2PreTrainedModel', 'Blip2ForConditionalGeneration', 'Blip2VisionModel', ] if TYPE_CHECKING: from .configuration_blip_a import ( BLIP_2_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipaConfig, BlipaQFormerConfig, BlipaVisionConfig, ) from .processing_blip_a import BlipaProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip_a import ( BLIP_2_PRETRAINED_MODEL_ARCHIVE_LIST, BlipaForConditionalGeneration, BlipaModel, BlipaPreTrainedModel, BlipaQFormerModel, BlipaVisionModel, ) else: import sys A_ : List[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
64
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available A_ : Optional[Any] = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : str = ['BartphoTokenizer'] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bartpho import BartphoTokenizer else: import sys A_ : Optional[int] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
64
1
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import PoolFormerImageProcessor class _a (unittest.TestCase ): '''simple docstring''' def __init__( self , A__ , A__=7 , A__=3 , A__=30 , A__=400 , A__=True , A__=None , A__=0.9 , A__=None , A__=True , A__=[0.5, 0.5, 0.5] , A__=[0.5, 0.5, 0.5] , ): A__ : List[Any] = size if size is not None else {"""shortest_edge""": 30} A__ : Any = crop_size if crop_size is not None else {"""height""": 30, """width""": 30} A__ : Union[str, Any] = parent A__ : Optional[int] = batch_size A__ : Any = num_channels A__ : List[Any] = min_resolution A__ : Optional[Any] = max_resolution A__ : List[str] = do_resize_and_center_crop A__ : int = size A__ : Tuple = crop_pct A__ : List[Any] = crop_size A__ : int = do_normalize A__ : str = image_mean A__ : Any = image_std def __A ( self ): return { "size": self.size, "do_resize_and_center_crop": self.do_resize_and_center_crop, "crop_pct": self.crop_pct, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class _a (__magic_name__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__: Tuple = PoolFormerImageProcessor if is_vision_available() else None def __A ( self ): A__ : Tuple = PoolFormerImageProcessingTester(self ) @property def __A ( self ): return self.image_processor_tester.prepare_image_processor_dict() def __A ( self ): A__ : Dict = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A__ , """do_resize_and_center_crop""" ) ) self.assertTrue(hasattr(A__ , """size""" ) ) self.assertTrue(hasattr(A__ , """crop_pct""" ) ) self.assertTrue(hasattr(A__ , """do_normalize""" ) ) self.assertTrue(hasattr(A__ , """image_mean""" ) ) self.assertTrue(hasattr(A__ , """image_std""" ) ) def __A ( self ): A__ : Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""shortest_edge""": 30} ) self.assertEqual(image_processor.crop_size , {"""height""": 30, """width""": 30} ) A__ : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {"""shortest_edge""": 42} ) self.assertEqual(image_processor.crop_size , {"""height""": 84, """width""": 84} ) def __A ( self ): pass def __A ( self ): # Initialize image_processing A__ : int = self.image_processing_class(**self.image_processor_dict ) # create random PIL images A__ : Tuple = prepare_image_inputs(self.image_processor_tester , equal_resolution=A__ ) for image in image_inputs: self.assertIsInstance(A__ , Image.Image ) # Test not batched input A__ : Tuple = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A__ : List[str] = image_processing(A__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def __A ( self ): # Initialize image_processing A__ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors A__ : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A__ , numpify=A__ ) for image in image_inputs: self.assertIsInstance(A__ , np.ndarray ) # Test not batched input A__ : Tuple = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A__ : List[Any] = image_processing(A__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def __A ( self ): # Initialize image_processing A__ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A__ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=A__ , torchify=A__ ) for image in image_inputs: self.assertIsInstance(A__ , torch.Tensor ) # Test not batched input A__ : Dict = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A__ : Union[str, Any] = image_processing(A__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , )
64
import argparse import hashlib import os import urllib import warnings import torch from torch import nn from tqdm import tqdm from transformers import WhisperConfig, WhisperForConditionalGeneration A_ : Dict = { 'tiny.en': 'https://openaipublic.azureedge.net/main/whisper/models/d3dd57d32accea0b295c96e26691aa14d8822fac7d9d27d5dc00b4ca2826dd03/tiny.en.pt', 'tiny': 'https://openaipublic.azureedge.net/main/whisper/models/65147644a518d12f04e32d6f3b26facc3f8dd46e5390956a9424a650c0ce22b9/tiny.pt', 'base.en': 'https://openaipublic.azureedge.net/main/whisper/models/25a8566e1d0c1e2231d1c762132cd20e0f96a85d16145c3a00adf5d1ac670ead/base.en.pt', 'base': 'https://openaipublic.azureedge.net/main/whisper/models/ed3a0b6b1c0edf879ad9b11b1af5a0e6ab5db9205f891f668f8b0e6c6326e34e/base.pt', 'small.en': 'https://openaipublic.azureedge.net/main/whisper/models/f953ad0fd29cacd07d5a9eda5624af0f6bcf2258be67c92b79389873d91e0872/small.en.pt', 'small': 'https://openaipublic.azureedge.net/main/whisper/models/9ecf779972d90ba49c06d968637d720dd632c55bbf19d441fb42bf17a411e794/small.pt', 'medium.en': 'https://openaipublic.azureedge.net/main/whisper/models/d7440d1dc186f76616474e0ff0b3b6b879abc9d1a4926b7adfa41db2d497ab4f/medium.en.pt', 'medium': 'https://openaipublic.azureedge.net/main/whisper/models/345ae4da62f9b3d59415adc60127b97c714f32e89e936602e85993674d08dcb1/medium.pt', 'large': 'https://openaipublic.azureedge.net/main/whisper/models/e4b87e7e0bf463eb8e6956e646f1e277e901512310def2c24bf0e11bd3c28e9a/large.pt', 'large-v2': 'https://openaipublic.azureedge.net/main/whisper/models/81f7c96c852ee8fc832187b0132e569d6c3065a3252ed18e56effd0b6a73e524/large-v2.pt', } def UpperCamelCase (lowercase_: Optional[Any] ) -> Optional[int]: A__ : List[Any] = ["""layers""", """blocks"""] for k in ignore_keys: state_dict.pop(lowercase_ , lowercase_ ) A_ : Any = { 'blocks': 'layers', 'mlp.0': 'fc1', 'mlp.2': 'fc2', 'mlp_ln': 'final_layer_norm', '.attn.query': '.self_attn.q_proj', '.attn.key': '.self_attn.k_proj', '.attn.value': '.self_attn.v_proj', '.attn_ln': '.self_attn_layer_norm', '.attn.out': '.self_attn.out_proj', '.cross_attn.query': '.encoder_attn.q_proj', '.cross_attn.key': '.encoder_attn.k_proj', '.cross_attn.value': '.encoder_attn.v_proj', '.cross_attn_ln': '.encoder_attn_layer_norm', '.cross_attn.out': '.encoder_attn.out_proj', 'decoder.ln.': 'decoder.layer_norm.', 'encoder.ln.': 'encoder.layer_norm.', 'token_embedding': 'embed_tokens', 'encoder.positional_embedding': 'encoder.embed_positions.weight', 'decoder.positional_embedding': 'decoder.embed_positions.weight', 'ln_post': 'layer_norm', } def UpperCamelCase (lowercase_: str ) -> Any: A__ : Dict = list(s_dict.keys() ) for key in keys: A__ : List[str] = key for k, v in WHISPER_MAPPING.items(): if k in key: A__ : List[Any] = new_key.replace(lowercase_ , lowercase_ ) print(f"""{key} -> {new_key}""" ) A__ : Tuple = s_dict.pop(lowercase_ ) return s_dict def UpperCamelCase (lowercase_: Tuple ) -> Optional[int]: A__ , A__ : Any = emb.weight.shape A__ : str = nn.Linear(lowercase_ , lowercase_ , bias=lowercase_ ) A__ : Union[str, Any] = emb.weight.data return lin_layer def UpperCamelCase (lowercase_: str , lowercase_: str ) -> bytes: os.makedirs(lowercase_ , exist_ok=lowercase_ ) A__ : Tuple = os.path.basename(lowercase_ ) A__ : int = url.split("""/""" )[-2] A__ : Dict = os.path.join(lowercase_ , lowercase_ ) if os.path.exists(lowercase_ ) and not os.path.isfile(lowercase_ ): raise RuntimeError(f"""{download_target} exists and is not a regular file""" ) if os.path.isfile(lowercase_ ): A__ : Optional[Any] = open(lowercase_ , """rb""" ).read() if hashlib.shaaaa(lowercase_ ).hexdigest() == expected_shaaaa: return model_bytes else: warnings.warn(f"""{download_target} exists, but the SHA256 checksum does not match; re-downloading the file""" ) with urllib.request.urlopen(lowercase_ ) as source, open(lowercase_ , """wb""" ) as output: with tqdm( total=int(source.info().get("""Content-Length""" ) ) , ncols=80 , unit="""iB""" , unit_scale=lowercase_ , unit_divisor=1024 ) as loop: while True: A__ : Any = source.read(8192 ) if not buffer: break output.write(lowercase_ ) loop.update(len(lowercase_ ) ) A__ : Dict = open(lowercase_ , """rb""" ).read() if hashlib.shaaaa(lowercase_ ).hexdigest() != expected_shaaaa: raise RuntimeError( """Model has been downloaded but the SHA256 checksum does not not match. Please retry loading the model.""" ) return model_bytes def UpperCamelCase (lowercase_: List[Any] , lowercase_: Tuple ) -> Optional[Any]: if ".pt" not in checkpoint_path: A__ : Tuple = _download(_MODELS[checkpoint_path] ) else: A__ : Optional[int] = torch.load(lowercase_ , map_location="""cpu""" ) A__ : str = original_checkpoint["""dims"""] A__ : List[Any] = original_checkpoint["""model_state_dict"""] A__ : Optional[Any] = state_dict["""decoder.token_embedding.weight"""] remove_ignore_keys_(lowercase_ ) rename_keys(lowercase_ ) A__ : List[str] = True A__ : Optional[Any] = state_dict["""decoder.layers.0.fc1.weight"""].shape[0] A__ : List[Any] = WhisperConfig( vocab_size=dimensions["""n_vocab"""] , encoder_ffn_dim=lowercase_ , decoder_ffn_dim=lowercase_ , num_mel_bins=dimensions["""n_mels"""] , d_model=dimensions["""n_audio_state"""] , max_target_positions=dimensions["""n_text_ctx"""] , encoder_layers=dimensions["""n_audio_layer"""] , encoder_attention_heads=dimensions["""n_audio_head"""] , decoder_layers=dimensions["""n_text_layer"""] , decoder_attention_heads=dimensions["""n_text_state"""] , max_source_positions=dimensions["""n_audio_ctx"""] , ) A__ : Optional[Any] = WhisperForConditionalGeneration(lowercase_ ) A__ , A__ : List[Any] = model.model.load_state_dict(lowercase_ , strict=lowercase_ ) if len(lowercase_ ) > 0 and not set(lowercase_ ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( """Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing,""" f""" but all the following weights are missing {missing}""" ) if tie_embeds: A__ : Any = make_linear_from_emb(model.model.decoder.embed_tokens ) else: A__ : str = proj_out_weights model.save_pretrained(lowercase_ ) if __name__ == "__main__": A_ : Any = argparse.ArgumentParser() # # Required parameters parser.add_argument('--checkpoint_path', type=str, help='Patht to the downloaded checkpoints') parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') A_ : Tuple = parser.parse_args() convert_openai_whisper_to_tfms(args.checkpoint_path, args.pytorch_dump_folder_path)
64
1
import warnings from .generation import TFGenerationMixin class _a (__magic_name__ ): '''simple docstring''' warnings.warn( '''Importing `TFGenerationMixin` from `src/transformers/generation_tf_utils.py` is deprecated and will ''' '''be removed in Transformers v5. Import as `from transformers import TFGenerationMixin` instead.''' , __magic_name__ , )
64
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class _a (__magic_name__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__: Any = TextToVideoSDPipeline UpperCAmelCase__: Any = TEXT_TO_IMAGE_PARAMS UpperCAmelCase__: Optional[Any] = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. UpperCAmelCase__: Optional[int] = frozenset( [ '''num_inference_steps''', '''generator''', '''latents''', '''return_dict''', '''callback''', '''callback_steps''', ] ) def __A ( self ): torch.manual_seed(0 ) A__ : Optional[int] = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """DownBlock3D""") , up_block_types=("""UpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""") , cross_attention_dim=32 , attention_head_dim=4 , ) A__ : Optional[int] = DDIMScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule="""scaled_linear""" , clip_sample=A__ , set_alpha_to_one=A__ , ) torch.manual_seed(0 ) A__ : Union[str, Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) A__ : Union[str, Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act="""gelu""" , projection_dim=512 , ) A__ : Union[str, Any] = CLIPTextModel(A__ ) A__ : Tuple = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) A__ : Dict = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, } return components def __A ( self , A__ , A__=0 ): if str(A__ ).startswith("""mps""" ): A__ : Tuple = torch.manual_seed(A__ ) else: A__ : List[str] = torch.Generator(device=A__ ).manual_seed(A__ ) A__ : List[str] = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """pt""", } return inputs def __A ( self ): A__ : List[str] = """cpu""" # ensure determinism for the device-dependent torch.Generator A__ : Union[str, Any] = self.get_dummy_components() A__ : Union[str, Any] = TextToVideoSDPipeline(**A__ ) A__ : int = sd_pipe.to(A__ ) sd_pipe.set_progress_bar_config(disable=A__ ) A__ : int = self.get_dummy_inputs(A__ ) A__ : int = """np""" A__ : Any = sd_pipe(**A__ ).frames A__ : Dict = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) A__ : Optional[Any] = np.array([1_5_8.0, 1_6_0.0, 1_5_3.0, 1_2_5.0, 1_0_0.0, 1_2_1.0, 1_1_1.0, 9_3.0, 1_1_3.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def __A ( self ): self._test_attention_slicing_forward_pass(test_mean_pixel_difference=A__ , expected_max_diff=3e-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def __A ( self ): self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=A__ , expected_max_diff=1e-2 ) @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def __A ( self ): pass @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def __A ( self ): pass @unittest.skip(reason="""`num_images_per_prompt` argument is not supported for this pipeline.""" ) def __A ( self ): pass def __A ( self ): return super().test_progress_bar() @slow @skip_mps class _a (unittest.TestCase ): '''simple docstring''' def __A ( self ): A__ : Union[str, Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy""" ) A__ : Tuple = TextToVideoSDPipeline.from_pretrained("""damo-vilab/text-to-video-ms-1.7b""" ) A__ : Any = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) A__ : int = pipe.to("""cuda""" ) A__ : Optional[Any] = """Spiderman is surfing""" A__ : List[str] = torch.Generator(device="""cpu""" ).manual_seed(0 ) A__ : Optional[Any] = pipe(A__ , generator=A__ , num_inference_steps=25 , output_type="""pt""" ).frames A__ : Dict = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5e-2 def __A ( self ): A__ : List[Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy""" ) A__ : Optional[int] = TextToVideoSDPipeline.from_pretrained("""damo-vilab/text-to-video-ms-1.7b""" ) A__ : List[str] = pipe.to("""cuda""" ) A__ : Dict = """Spiderman is surfing""" A__ : Union[str, Any] = torch.Generator(device="""cpu""" ).manual_seed(0 ) A__ : Optional[int] = pipe(A__ , generator=A__ , num_inference_steps=2 , output_type="""pt""" ).frames A__ : Optional[int] = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5e-2
64
1
def UpperCamelCase (lowercase_: int = 1000 ) -> int: A__ : Dict = -1 A__ : str = 0 for a in range(1 , n // 3 ): # Solving the two equations a**2+b**2=c**2 and a+b+c=N eliminating c A__ : Optional[int] = (n * n - 2 * a * n) // (2 * n - 2 * a) A__ : List[str] = n - a - b if c * c == (a * a + b * b): A__ : Tuple = a * b * c if candidate >= product: A__ : Dict = candidate return product if __name__ == "__main__": print(f'''{solution() = }''')
64
def UpperCamelCase (lowercase_: int ) -> int: if not isinstance(lowercase_ , lowercase_ ): raise TypeError("""Input value must be an 'int' type""" ) A__ : int = 0 while number: position += 1 number >>= 1 return position if __name__ == "__main__": import doctest doctest.testmod()
64
1
import enum import shutil import sys A_ , A_ : Optional[Any] = shutil.get_terminal_size() A_ : Optional[int] = {'UP': 'A', 'DOWN': 'B', 'RIGHT': 'C', 'LEFT': 'D'} class _a (enum.Enum ): '''simple docstring''' UpperCAmelCase__: str = 0 UpperCAmelCase__: Optional[Any] = 1 def UpperCamelCase (lowercase_: Optional[Any] , lowercase_: Optional[int]="" ) -> int: sys.stdout.write(str(lowercase_ ) + end ) sys.stdout.flush() def UpperCamelCase (lowercase_: int , lowercase_: List[str] , lowercase_: Optional[int]="" ) -> Optional[Any]: forceWrite(f"""\u001b[{color}m{content}\u001b[0m""" , lowercase_ ) def UpperCamelCase () -> Union[str, Any]: forceWrite("""\r""" ) def UpperCamelCase (lowercase_: int , lowercase_: str ) -> Union[str, Any]: forceWrite(f"""\033[{num_lines}{CURSOR_TO_CHAR[direction.upper()]}""" ) def UpperCamelCase () -> int: forceWrite(""" """ * TERMINAL_WIDTH ) reset_cursor() def UpperCamelCase () -> Tuple: reset_cursor() forceWrite("""-""" * TERMINAL_WIDTH )
64
from pathlib import Path import cva import numpy as np from matplotlib import pyplot as plt def UpperCamelCase (lowercase_: np.ndarray , lowercase_: np.ndarray , lowercase_: np.ndarray , lowercase_: int , lowercase_: int ) -> np.ndarray: A__ : Any = cva.getAffineTransform(lowercase_ , lowercase_ ) return cva.warpAffine(lowercase_ , lowercase_ , (rows, cols) ) if __name__ == "__main__": # read original image A_ : List[Any] = cva.imread( str(Path(__file__).resolve().parent.parent / 'image_data' / 'lena.jpg') ) # turn image in gray scale value A_ : List[Any] = cva.cvtColor(image, cva.COLOR_BGR2GRAY) # get image shape A_ , A_ : Optional[Any] = gray_img.shape # set different points to rotate image A_ : str = np.array([[50, 50], [200, 50], [50, 200]], np.floataa) A_ : Dict = np.array([[10, 100], [200, 50], [100, 250]], np.floataa) A_ : Optional[int] = np.array([[50, 50], [150, 50], [120, 200]], np.floataa) A_ : Optional[int] = np.array([[10, 100], [80, 50], [180, 250]], np.floataa) # add all rotated images in a list A_ : Dict = [ gray_img, get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), ] # plot different image rotations A_ : Union[str, Any] = plt.figure(1) A_ : Union[str, Any] = ['Original', 'Rotation 1', 'Rotation 2', 'Rotation 3'] for i, image in enumerate(images): plt.subplot(2, 2, i + 1), plt.imshow(image, 'gray') plt.title(titles[i]) plt.axis('off') plt.subplots_adjust(left=0.0, bottom=0.05, right=1.0, top=0.95) plt.show()
64
1
import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser A_ : Dict = re.compile(r'\s+') def UpperCamelCase (lowercase_: List[str] ) -> List[Any]: return {"hash": hashlib.mda(re.sub(lowercase_ , """""" , example["""content"""] ).encode("""utf-8""" ) ).hexdigest()} def UpperCamelCase (lowercase_: Union[str, Any] ) -> int: A__ : Any = [len(lowercase_ ) for line in example["""content"""].splitlines()] return {"line_mean": np.mean(lowercase_ ), "line_max": max(lowercase_ )} def UpperCamelCase (lowercase_: Dict ) -> Union[str, Any]: A__ : List[Any] = np.mean([c.isalnum() for c in example["""content"""]] ) return {"alpha_frac": alpha_frac} def UpperCamelCase (lowercase_: Optional[Any] , lowercase_: str ) -> int: if example["hash"] in uniques: uniques.remove(example["""hash"""] ) return True else: return False def UpperCamelCase (lowercase_: Union[str, Any] , lowercase_: Any=5 ) -> List[str]: A__ : Optional[int] = ["""auto-generated""", """autogenerated""", """automatically generated"""] A__ : Any = example["""content"""].splitlines() for _, line in zip(range(lowercase_ ) , lowercase_ ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def UpperCamelCase (lowercase_: Optional[int] , lowercase_: Tuple=5 , lowercase_: int=0.05 ) -> Union[str, Any]: A__ : List[Any] = ["""unit tests""", """test file""", """configuration file"""] A__ : str = example["""content"""].splitlines() A__ : Optional[int] = 0 A__ : Any = 0 # first test for _, line in zip(range(lowercase_ ) , lowercase_ ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test A__ : List[Any] = example["""content"""].count("""\n""" ) A__ : Any = int(coeff * nlines ) for line in lines: count_config += line.lower().count("""config""" ) count_test += line.lower().count("""test""" ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def UpperCamelCase (lowercase_: Optional[Any] ) -> List[str]: A__ : Union[str, Any] = ["""def """, """class """, """for """, """while """] A__ : Tuple = example["""content"""].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def UpperCamelCase (lowercase_: Any , lowercase_: str=4 ) -> List[str]: A__ : List[str] = example["""content"""].splitlines() A__ : Union[str, Any] = 0 for line in lines: counter += line.lower().count("""=""" ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def UpperCamelCase (lowercase_: int ) -> int: A__ : str = tokenizer(example["""content"""] , truncation=lowercase_ )["""input_ids"""] A__ : int = len(example["""content"""] ) / len(lowercase_ ) return {"ratio": ratio} def UpperCamelCase (lowercase_: Tuple ) -> Optional[Any]: A__ : Dict = {} results.update(get_hash(lowercase_ ) ) results.update(line_stats(lowercase_ ) ) results.update(alpha_stats(lowercase_ ) ) results.update(char_token_ratio(lowercase_ ) ) results.update(is_autogenerated(lowercase_ ) ) results.update(is_config_or_test(lowercase_ ) ) results.update(has_no_keywords(lowercase_ ) ) results.update(has_few_assignments(lowercase_ ) ) return results def UpperCamelCase (lowercase_: int , lowercase_: Any , lowercase_: Optional[int] ) -> Tuple: if not check_uniques(lowercase_ , lowercase_ ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def UpperCamelCase (lowercase_: Optional[int] ) -> Optional[int]: with open(lowercase_ , """rb""" ) as f_in: with gzip.open(str(lowercase_ ) + """.gz""" , """wb""" , compresslevel=6 ) as f_out: shutil.copyfileobj(lowercase_ , lowercase_ ) os.unlink(lowercase_ ) # Settings A_ : Any = HfArgumentParser(PreprocessingArguments) A_ : Union[str, Any] = parser.parse_args() if args.num_workers is None: A_ : Union[str, Any] = multiprocessing.cpu_count() A_ : Any = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset A_ : Union[str, Any] = time.time() A_ : int = load_dataset(args.dataset_name, split='train') print(f'''Time to load dataset: {time.time()-t_start:.2f}''') # Run preprocessing A_ : Union[str, Any] = time.time() A_ : Optional[int] = ds.map(preprocess, num_proc=args.num_workers) print(f'''Time to preprocess dataset: {time.time()-t_start:.2f}''') # Deduplicate hashes A_ : Optional[int] = set(ds.unique('hash')) A_ : Optional[Any] = len(uniques) / len(ds) print(f'''Fraction of duplicates: {1-frac:.2%}''') # Deduplicate data and apply heuristics A_ : Optional[int] = time.time() A_ : List[str] = ds.filter(filter, fn_kwargs={'uniques': uniques, 'args': args}) print(f'''Time to filter dataset: {time.time()-t_start:.2f}''') print(f'''Size of filtered dataset: {len(ds_filter)}''') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: A_ : Any = time.time() A_ , A_ : Dict = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(f'''Time to deduplicate dataset: {time.time()-t_start:.2f}''') print(f'''Size of deduplicate dataset: {len(ds_filter)}''') # Save data in batches of samples_per_file A_ : Dict = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / 'duplicate_clusters.json', 'w') as f: json.dump(duplicate_clusters, f) A_ : List[Any] = output_dir / 'data' data_dir.mkdir(exist_ok=True) A_ : Union[str, Any] = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): A_ : Any = str(data_dir / f'''file-{file_number+1:012}.json''') A_ : Union[str, Any] = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(f'''Time to save dataset: {time.time()-t_start:.2f}''')
64
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_torch_available from transformers.testing_utils import require_torch, torch_device if is_torch_available(): from transformers import PyTorchBenchmark, PyTorchBenchmarkArguments @require_torch class _a (unittest.TestCase ): '''simple docstring''' def __A ( self , A__ ): for model_result in results.values(): for batch_size, sequence_length in zip(model_result["""bs"""] , model_result["""ss"""] ): A__ : str = model_result["""result"""][batch_size][sequence_length] self.assertIsNotNone(A__ ) def __A ( self ): A__ : Dict = """sshleifer/tiny-gpt2""" A__ : Tuple = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A__ , ) A__ : int = PyTorchBenchmark(A__ ) A__ : List[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self ): A__ : Dict = """sgugger/tiny-distilbert-classification""" A__ : Dict = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A__ , only_pretrain_model=A__ , ) A__ : str = PyTorchBenchmark(A__ ) A__ : List[str] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self ): A__ : Any = """sshleifer/tiny-gpt2""" A__ : List[Any] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , torchscript=A__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A__ , ) A__ : Tuple = PyTorchBenchmark(A__ ) A__ : str = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(torch_device == """cpu""" , """Cant do half precision""" ) def __A ( self ): A__ : Optional[Any] = """sshleifer/tiny-gpt2""" A__ : Optional[Any] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , fpaa=A__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A__ , ) A__ : str = PyTorchBenchmark(A__ ) A__ : Any = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self ): A__ : Optional[Any] = """sshleifer/tiny-gpt2""" A__ : Tuple = AutoConfig.from_pretrained(A__ ) # set architectures equal to `None` A__ : List[Any] = None A__ : str = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A__ , ) A__ : List[str] = PyTorchBenchmark(A__ , configs=[config] ) A__ : Any = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self ): A__ : Optional[int] = """sshleifer/tiny-gpt2""" A__ : Optional[int] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A__ , ) A__ : Any = PyTorchBenchmark(A__ ) A__ : Dict = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) @unittest.skipIf(torch_device == """cpu""" , """Can't do half precision""" ) def __A ( self ): A__ : Optional[int] = """sshleifer/tiny-gpt2""" A__ : List[str] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , sequence_lengths=[8] , batch_sizes=[1] , fpaa=A__ , multi_process=A__ , ) A__ : Dict = PyTorchBenchmark(A__ ) A__ : Optional[int] = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def __A ( self ): A__ : int = """sshleifer/tiny-gpt2""" A__ : Optional[int] = AutoConfig.from_pretrained(A__ ) A__ : str = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A__ , ) A__ : int = PyTorchBenchmark(A__ , configs=[config] ) A__ : Tuple = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self ): A__ : List[str] = """sshleifer/tinier_bart""" A__ : List[str] = AutoConfig.from_pretrained(A__ ) A__ : List[str] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A__ , ) A__ : Union[str, Any] = PyTorchBenchmark(A__ , configs=[config] ) A__ : str = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self ): A__ : Optional[int] = """sshleifer/tiny-gpt2""" A__ : Union[str, Any] = AutoConfig.from_pretrained(A__ ) A__ : Tuple = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A__ , ) A__ : int = PyTorchBenchmark(A__ , configs=[config] ) A__ : Tuple = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def __A ( self ): A__ : Dict = """sshleifer/tinier_bart""" A__ : int = AutoConfig.from_pretrained(A__ ) A__ : Union[str, Any] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A__ , ) A__ : List[Any] = PyTorchBenchmark(A__ , configs=[config] ) A__ : Optional[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def __A ( self ): A__ : int = """sshleifer/tiny-gpt2""" with tempfile.TemporaryDirectory() as tmp_dir: A__ : Dict = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , save_to_csv=A__ , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(A__ , """inf_time.csv""" ) , train_memory_csv_file=os.path.join(A__ , """train_mem.csv""" ) , inference_memory_csv_file=os.path.join(A__ , """inf_mem.csv""" ) , train_time_csv_file=os.path.join(A__ , """train_time.csv""" ) , env_info_csv_file=os.path.join(A__ , """env.csv""" ) , multi_process=A__ , ) A__ : Optional[Any] = PyTorchBenchmark(A__ ) benchmark.run() self.assertTrue(Path(os.path.join(A__ , """inf_time.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(A__ , """train_time.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(A__ , """inf_mem.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(A__ , """train_mem.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(A__ , """env.csv""" ) ).exists() ) def __A ( self ): A__ : Optional[int] = """sshleifer/tiny-gpt2""" def _check_summary_is_not_empty(A__ ): self.assertTrue(hasattr(A__ , """sequential""" ) ) self.assertTrue(hasattr(A__ , """cumulative""" ) ) self.assertTrue(hasattr(A__ , """current""" ) ) self.assertTrue(hasattr(A__ , """total""" ) ) with tempfile.TemporaryDirectory() as tmp_dir: A__ : Dict = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(A__ , """log.txt""" ) , log_print=A__ , trace_memory_line_by_line=A__ , multi_process=A__ , ) A__ : Dict = PyTorchBenchmark(A__ ) A__ : str = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) _check_summary_is_not_empty(result.train_summary ) self.assertTrue(Path(os.path.join(A__ , """log.txt""" ) ).exists() )
64
1
from __future__ import annotations from collections.abc import Iterator from typing import Any class _a : '''simple docstring''' def __init__( self , A__ ): A__ : Any = data A__ : Node | None = None class _a : '''simple docstring''' def __init__( self ): A__ : int = None A__ : Tuple = None def __iter__( self ): A__ : Optional[int] = self.head while self.head: yield node.data A__ : List[Any] = node.next if node == self.head: break def __len__( self ): return sum(1 for _ in self ) def __repr__( self ): return "->".join(str(A__ ) for item in iter(self ) ) def __A ( self , A__ ): self.insert_nth(len(self ) , A__ ) def __A ( self , A__ ): self.insert_nth(0 , A__ ) def __A ( self , A__ , A__ ): if index < 0 or index > len(self ): raise IndexError("""list index out of range.""" ) A__ : Union[str, Any] = Node(A__ ) if self.head is None: A__ : Optional[int] = new_node # first node points itself A__ : List[Any] = new_node elif index == 0: # insert at head A__ : Tuple = self.head A__ : Tuple = new_node else: A__ : str = self.head for _ in range(index - 1 ): A__ : Optional[int] = temp.next A__ : int = temp.next A__ : int = new_node if index == len(self ) - 1: # insert at tail A__ : int = new_node def __A ( self ): return self.delete_nth(0 ) def __A ( self ): return self.delete_nth(len(self ) - 1 ) def __A ( self , A__ = 0 ): if not 0 <= index < len(self ): raise IndexError("""list index out of range.""" ) A__ : Optional[Any] = self.head if self.head == self.tail: # just one node A__ : Optional[int] = None elif index == 0: # delete head node A__ : Dict = self.tail.next.next A__ : Optional[Any] = self.head.next else: A__ : List[str] = self.head for _ in range(index - 1 ): A__ : List[str] = temp.next A__ : Tuple = temp.next A__ : List[Any] = temp.next.next if index == len(self ) - 1: # delete at tail A__ : List[str] = temp return delete_node.data def __A ( self ): return len(self ) == 0 def UpperCamelCase () -> None: A__ : int = CircularLinkedList() assert len(lowercase_ ) == 0 assert circular_linked_list.is_empty() is True assert str(lowercase_ ) == "" try: circular_linked_list.delete_front() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_tail() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_nth(-1 ) raise AssertionError except IndexError: assert True try: circular_linked_list.delete_nth(0 ) raise AssertionError except IndexError: assert True assert circular_linked_list.is_empty() is True for i in range(5 ): assert len(lowercase_ ) == i circular_linked_list.insert_nth(lowercase_ , i + 1 ) assert str(lowercase_ ) == "->".join(str(lowercase_ ) for i in range(1 , 6 ) ) circular_linked_list.insert_tail(6 ) assert str(lowercase_ ) == "->".join(str(lowercase_ ) for i in range(1 , 7 ) ) circular_linked_list.insert_head(0 ) assert str(lowercase_ ) == "->".join(str(lowercase_ ) for i in range(0 , 7 ) ) assert circular_linked_list.delete_front() == 0 assert circular_linked_list.delete_tail() == 6 assert str(lowercase_ ) == "->".join(str(lowercase_ ) for i in range(1 , 6 ) ) assert circular_linked_list.delete_nth(2 ) == 3 circular_linked_list.insert_nth(2 , 3 ) assert str(lowercase_ ) == "->".join(str(lowercase_ ) for i in range(1 , 6 ) ) assert circular_linked_list.is_empty() is False if __name__ == "__main__": import doctest doctest.testmod()
64
# tests directory-specific settings - this file is run automatically # by pytest before any tests are run import doctest import sys import warnings from os.path import abspath, dirname, join import _pytest from transformers.testing_utils import HfDoctestModule, HfDocTestParser # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. A_ : Optional[int] = abspath(join(dirname(__file__), 'src')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='ignore', category=FutureWarning) def UpperCamelCase (lowercase_: List[str] ) -> Any: config.addinivalue_line( """markers""" , """is_pt_tf_cross_test: mark test to run only when PT and TF interactions are tested""" ) config.addinivalue_line( """markers""" , """is_pt_flax_cross_test: mark test to run only when PT and FLAX interactions are tested""" ) config.addinivalue_line("""markers""" , """is_pipeline_test: mark test to run only when pipelines are tested""" ) config.addinivalue_line("""markers""" , """is_staging_test: mark test to run only in the staging environment""" ) config.addinivalue_line("""markers""" , """accelerate_tests: mark test that require accelerate""" ) config.addinivalue_line("""markers""" , """tool_tests: mark the tool tests that are run on their specific schedule""" ) def UpperCamelCase (lowercase_: Optional[int] ) -> Optional[Any]: from transformers.testing_utils import pytest_addoption_shared pytest_addoption_shared(lowercase_ ) def UpperCamelCase (lowercase_: List[str] ) -> Optional[Any]: from transformers.testing_utils import pytest_terminal_summary_main A__ : List[Any] = terminalreporter.config.getoption("""--make-reports""" ) if make_reports: pytest_terminal_summary_main(lowercase_ , id=lowercase_ ) def UpperCamelCase (lowercase_: Union[str, Any] , lowercase_: int ) -> List[str]: # If no tests are collected, pytest exists with code 5, which makes the CI fail. if exitstatus == 5: A__ : Tuple = 0 # Doctest custom flag to ignore output. A_ : Tuple = doctest.register_optionflag('IGNORE_RESULT') A_ : Dict = doctest.OutputChecker class _a (__magic_name__ ): '''simple docstring''' def __A ( self , A__ , A__ , A__ ): if IGNORE_RESULT & optionflags: return True return OutputChecker.check_output(self , A__ , A__ , A__ ) A_ : str = CustomOutputChecker A_ : Dict = HfDoctestModule A_ : Optional[int] = HfDocTestParser
64
1
def UpperCamelCase (lowercase_: Optional[int] ) -> List[Any]: A__ : Any = 0 A__ : int = len(lowercase_ ) for i in range(n - 1 ): for j in range(i + 1 , lowercase_ ): if arr[i] > arr[j]: num_inversions += 1 return num_inversions def UpperCamelCase (lowercase_: Dict ) -> Dict: if len(lowercase_ ) <= 1: return arr, 0 A__ : Optional[Any] = len(lowercase_ ) // 2 A__ : Optional[int] = arr[0:mid] A__ : int = arr[mid:] A__ , A__ : List[Any] = count_inversions_recursive(lowercase_ ) A__ , A__ : Dict = count_inversions_recursive(lowercase_ ) A__ , A__ : int = _count_cross_inversions(lowercase_ , lowercase_ ) A__ : Union[str, Any] = inversion_p + inversions_q + cross_inversions return c, num_inversions def UpperCamelCase (lowercase_: str , lowercase_: Union[str, Any] ) -> Dict: A__ : Optional[int] = [] A__ : int = 0 while i < len(lowercase_ ) and j < len(lowercase_ ): if p[i] > q[j]: # if P[1] > Q[j], then P[k] > Q[k] for all i < k <= len(P) # These are all inversions. The claim emerges from the # property that P is sorted. num_inversion += len(lowercase_ ) - i r.append(q[j] ) j += 1 else: r.append(p[i] ) i += 1 if i < len(lowercase_ ): r.extend(p[i:] ) else: r.extend(q[j:] ) return r, num_inversion def UpperCamelCase () -> Dict: A__ : List[Any] = [10, 2, 1, 5, 5, 2, 11] # this arr has 8 inversions: # (10, 2), (10, 1), (10, 5), (10, 5), (10, 2), (2, 1), (5, 2), (5, 2) A__ : int = count_inversions_bf(lowercase_ ) A__ , A__ : Tuple = count_inversions_recursive(lowercase_ ) assert num_inversions_bf == num_inversions_recursive == 8 print("""number of inversions = """ , lowercase_ ) # testing an array with zero inversion (a sorted arr_1) arr_a.sort() A__ : List[str] = count_inversions_bf(lowercase_ ) A__ , A__ : Any = count_inversions_recursive(lowercase_ ) assert num_inversions_bf == num_inversions_recursive == 0 print("""number of inversions = """ , lowercase_ ) # an empty list should also have zero inversions A__ : Optional[int] = [] A__ : Tuple = count_inversions_bf(lowercase_ ) A__ , A__ : List[str] = count_inversions_recursive(lowercase_ ) assert num_inversions_bf == num_inversions_recursive == 0 print("""number of inversions = """ , lowercase_ ) if __name__ == "__main__": main()
64
from __future__ import annotations import unittest from transformers import AutoTokenizer, PegasusConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFPegasusForConditionalGeneration, TFPegasusModel @require_tf class _a : '''simple docstring''' UpperCAmelCase__: List[Any] = PegasusConfig UpperCAmelCase__: Optional[int] = {} UpperCAmelCase__: List[str] = '''gelu''' def __init__( self , A__ , A__=13 , A__=7 , A__=True , A__=False , A__=99 , A__=32 , A__=2 , A__=4 , A__=37 , A__=0.1 , A__=0.1 , A__=40 , A__=2 , A__=1 , A__=0 , ): A__ : Dict = parent A__ : Dict = batch_size A__ : Any = seq_length A__ : Optional[Any] = is_training A__ : int = use_labels A__ : Any = vocab_size A__ : Union[str, Any] = hidden_size A__ : Tuple = num_hidden_layers A__ : Tuple = num_attention_heads A__ : List[Any] = intermediate_size A__ : Union[str, Any] = hidden_dropout_prob A__ : Optional[Any] = attention_probs_dropout_prob A__ : List[Any] = max_position_embeddings A__ : Any = eos_token_id A__ : List[Any] = pad_token_id A__ : List[Any] = bos_token_id def __A ( self ): A__ : str = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) A__ : Dict = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) A__ : List[Any] = tf.concat([input_ids, eos_tensor] , axis=1 ) A__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ : Tuple = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) A__ : str = prepare_pegasus_inputs_dict(A__ , A__ , A__ ) return config, inputs_dict def __A ( self , A__ , A__ ): A__ : int = TFPegasusModel(config=A__ ).get_decoder() A__ : List[Any] = inputs_dict["""input_ids"""] A__ : Any = input_ids[:1, :] A__ : Optional[Any] = inputs_dict["""attention_mask"""][:1, :] A__ : Optional[int] = inputs_dict["""head_mask"""] A__ : Any = 1 # first forward pass A__ : Tuple = model(A__ , attention_mask=A__ , head_mask=A__ , use_cache=A__ ) A__ , A__ : Dict = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids A__ : Tuple = ids_tensor((self.batch_size, 3) , config.vocab_size ) A__ : Optional[Any] = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and A__ : List[Any] = tf.concat([input_ids, next_tokens] , axis=-1 ) A__ : Tuple = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) A__ : Optional[Any] = model(A__ , attention_mask=A__ )[0] A__ : Any = model(A__ , attention_mask=A__ , past_key_values=A__ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice A__ : int = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) A__ : Any = output_from_no_past[:, -3:, random_slice_idx] A__ : Tuple = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(A__ , A__ , rtol=1e-3 ) def UpperCamelCase (lowercase_: Union[str, Any] , lowercase_: Dict , lowercase_: List[Any] , lowercase_: Dict=None , lowercase_: int=None , lowercase_: List[Any]=None , lowercase_: List[Any]=None , lowercase_: str=None , ) -> int: if attention_mask is None: A__ : List[str] = tf.cast(tf.math.not_equal(lowercase_ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: A__ : Dict = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: A__ : Any = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: A__ : Tuple = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: A__ : List[str] = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class _a (__magic_name__ , __magic_name__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__: List[Any] = (TFPegasusForConditionalGeneration, TFPegasusModel) if is_tf_available() else () UpperCAmelCase__: Tuple = (TFPegasusForConditionalGeneration,) if is_tf_available() else () UpperCAmelCase__: Tuple = ( { '''conversational''': TFPegasusForConditionalGeneration, '''feature-extraction''': TFPegasusModel, '''summarization''': TFPegasusForConditionalGeneration, '''text2text-generation''': TFPegasusForConditionalGeneration, '''translation''': TFPegasusForConditionalGeneration, } if is_tf_available() else {} ) UpperCAmelCase__: int = True UpperCAmelCase__: Union[str, Any] = False UpperCAmelCase__: List[str] = False def __A ( self ): A__ : Optional[Any] = TFPegasusModelTester(self ) A__ : Tuple = ConfigTester(self , config_class=A__ ) def __A ( self ): self.config_tester.run_common_tests() def __A ( self ): A__ : int = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*A__ ) @require_sentencepiece @require_tokenizers @require_tf class _a (unittest.TestCase ): '''simple docstring''' UpperCAmelCase__: Optional[int] = [ ''' PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.''', ''' The London trio are up for best UK act and best album, as well as getting two nominations in the best song category."We got told like this morning \'Oh I think you\'re nominated\'", said Dappy."And I was like \'Oh yeah, which one?\' And now we\'ve got nominated for four awards. I mean, wow!"Bandmate Fazer added: "We thought it\'s best of us to come down and mingle with everyone and say hello to the cameras. And now we find we\'ve got four nominations."The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn\'t be too disappointed if they didn\'t win this time around."At the end of the day we\'re grateful to be where we are in our careers."If it don\'t happen then it don\'t happen - live to fight another day and keep on making albums and hits for the fans."Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers\' All These Things That I\'ve Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year\'s Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border."We just done Edinburgh the other day," said Dappy."We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!" ''', ] UpperCAmelCase__: Any = [ '''California\'s largest electricity provider has cut power to hundreds of thousands of customers in an effort to''' ''' reduce the risk of wildfires.''', '''N-Dubz have revealed they\'re "grateful" to have been nominated for four Mobo Awards.''', ] # differs slightly from pytorch, likely due to numerical differences in linear layers UpperCAmelCase__: List[str] = '''google/pegasus-xsum''' @cached_property def __A ( self ): return AutoTokenizer.from_pretrained(self.model_name ) @cached_property def __A ( self ): A__ : int = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model def __A ( self , **A__ ): A__ : str = self.translate_src_text(**A__ ) assert self.expected_text == generated_words def __A ( self , **A__ ): A__ : List[str] = self.tokenizer(self.src_text , **A__ , padding=A__ , return_tensors="""tf""" ) A__ : Optional[int] = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 , use_cache=A__ , ) A__ : Dict = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=A__ ) return generated_words @slow def __A ( self ): self._assert_generated_batch_equal_expected()
64
1
from datetime import datetime as dt import os from github import Github A_ : Optional[int] = [ 'good first issue', 'good second issue', 'good difficult issue', 'feature request', 'new model', 'wip', ] def UpperCamelCase () -> Dict: A__ : Union[str, Any] = Github(os.environ["""GITHUB_TOKEN"""] ) A__ : Optional[int] = g.get_repo("""huggingface/transformers""" ) A__ : Tuple = repo.get_issues(state="""open""" ) for issue in open_issues: A__ : List[str] = sorted([comment for comment in issue.get_comments()] , key=lambda lowercase_ : i.created_at , reverse=lowercase_ ) A__ : Any = comments[0] if len(lowercase_ ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would close issue {issue.number} since it has been 7 days of inactivity since bot mention.") issue.edit(state="""closed""" ) elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would add stale comment to {issue.number}") issue.create_comment( """This issue has been automatically marked as stale because it has not had """ """recent activity. If you think this still needs to be addressed """ """please comment on this thread.\n\nPlease note that issues that do not follow the """ """[contributing guidelines](https://github.com/huggingface/transformers/blob/main/CONTRIBUTING.md) """ """are likely to be ignored.""" ) if __name__ == "__main__": main()
64
class _a : '''simple docstring''' def __init__( self ): A__ : str = """""" A__ : Any = """""" A__ : List[Any] = [] def __A ( self , A__ , A__ ): if m == -1: return n + 1 elif n == -1: return m + 1 elif self.dp[m][n] > -1: return self.dp[m][n] else: if self.worda[m] == self.worda[n]: A__ : Optional[Any] = self.__min_dist_top_down_dp(m - 1 , n - 1 ) else: A__ : Union[str, Any] = self.__min_dist_top_down_dp(A__ , n - 1 ) A__ : Union[str, Any] = self.__min_dist_top_down_dp(m - 1 , A__ ) A__ : Union[str, Any] = self.__min_dist_top_down_dp(m - 1 , n - 1 ) A__ : List[Any] = 1 + min(A__ , A__ , A__ ) return self.dp[m][n] def __A ( self , A__ , A__ ): A__ : Tuple = worda A__ : Dict = worda A__ : Optional[Any] = [[-1 for _ in range(len(A__ ) )] for _ in range(len(A__ ) )] return self.__min_dist_top_down_dp(len(A__ ) - 1 , len(A__ ) - 1 ) def __A ( self , A__ , A__ ): A__ : Optional[Any] = worda A__ : Dict = worda A__ : Union[str, Any] = len(A__ ) A__ : List[str] = len(A__ ) A__ : int = [[0 for _ in range(n + 1 )] for _ in range(m + 1 )] for i in range(m + 1 ): for j in range(n + 1 ): if i == 0: # first string is empty A__ : Tuple = j elif j == 0: # second string is empty A__ : Dict = i elif worda[i - 1] == worda[j - 1]: # last characters are equal A__ : str = self.dp[i - 1][j - 1] else: A__ : Union[str, Any] = self.dp[i][j - 1] A__ : str = self.dp[i - 1][j] A__ : Union[str, Any] = self.dp[i - 1][j - 1] A__ : Tuple = 1 + min(A__ , A__ , A__ ) return self.dp[m][n] if __name__ == "__main__": A_ : Union[str, Any] = EditDistance() print('****************** Testing Edit Distance DP Algorithm ******************') print() A_ : int = input('Enter the first string: ').strip() A_ : List[str] = input('Enter the second string: ').strip() print() print(f'''The minimum edit distance is: {solver.min_dist_top_down(Sa, Sa)}''') print(f'''The minimum edit distance is: {solver.min_dist_bottom_up(Sa, Sa)}''') print() print('*************** End of Testing Edit Distance DP Algorithm ***************')
64
1
def UpperCamelCase (lowercase_: list ) -> list: if len(lowercase_ ) <= 1: return [tuple(lowercase_ )] A__ : Optional[Any] = [] def generate(lowercase_: int , lowercase_: list ): A__ : int = [0] * n res.append(tuple(lowercase_ ) ) A__ : List[Any] = 0 while i < n: if c[i] < i: if i % 2 == 0: A__ , A__ : Optional[Any] = arr[i], arr[0] else: A__ , A__ : List[Any] = arr[i], arr[c[i]] res.append(tuple(lowercase_ ) ) c[i] += 1 A__ : List[str] = 0 else: A__ : str = 0 i += 1 generate(len(lowercase_ ) , lowercase_ ) return res if __name__ == "__main__": A_ : str = input('Enter numbers separated by a comma:\n').strip() A_ : Dict = [int(item) for item in user_input.split(',')] print(heaps(arr))
64
def UpperCamelCase (lowercase_: int , lowercase_: int ) -> int: while second != 0: A__ : int = first & second first ^= second A__ : int = c << 1 return first if __name__ == "__main__": import doctest doctest.testmod() A_ : Optional[Any] = int(input('Enter the first number: ').strip()) A_ : List[str] = int(input('Enter the second number: ').strip()) print(f'''{add(first, second) = }''')
64
1
A_ : Any = [ 'DownloadConfig', 'DownloadManager', 'DownloadMode', 'StreamingDownloadManager', ] from .download_config import DownloadConfig from .download_manager import DownloadManager, DownloadMode from .streaming_download_manager import StreamingDownloadManager
64
from __future__ import annotations from collections.abc import Callable A_ : List[Any] = list[list[float | int]] def UpperCamelCase (lowercase_: Matrix , lowercase_: Matrix ) -> Matrix: A__ : int = len(lowercase_ ) A__ : Matrix = [[0 for _ in range(size + 1 )] for _ in range(lowercase_ )] A__ : int A__ : int A__ : int A__ : int A__ : int A__ : float for row in range(lowercase_ ): for col in range(lowercase_ ): A__ : List[str] = matrix[row][col] A__ : int = vector[row][0] A__ : Optional[int] = 0 A__ : str = 0 while row < size and col < size: # pivoting A__ : int = max((abs(augmented[rowa][col] ), rowa) for rowa in range(lowercase_ , lowercase_ ) )[ 1 ] if augmented[pivot_row][col] == 0: col += 1 continue else: A__ , A__ : Union[str, Any] = augmented[pivot_row], augmented[row] for rowa in range(row + 1 , lowercase_ ): A__ : List[Any] = augmented[rowa][col] / augmented[row][col] A__ : Dict = 0 for cola in range(col + 1 , size + 1 ): augmented[rowa][cola] -= augmented[row][cola] * ratio row += 1 col += 1 # back substitution for col in range(1 , lowercase_ ): for row in range(lowercase_ ): A__ : List[str] = augmented[row][col] / augmented[col][col] for cola in range(lowercase_ , size + 1 ): augmented[row][cola] -= augmented[col][cola] * ratio # round to get rid of numbers like 2.000000000000004 return [ [round(augmented[row][size] / augmented[row][row] , 10 )] for row in range(lowercase_ ) ] def UpperCamelCase (lowercase_: list[int] ) -> Callable[[int], int]: A__ : int = len(lowercase_ ) A__ : Matrix = [[0 for _ in range(lowercase_ )] for _ in range(lowercase_ )] A__ : Matrix = [[0] for _ in range(lowercase_ )] A__ : Matrix A__ : int A__ : int A__ : int for x_val, y_val in enumerate(lowercase_ ): for col in range(lowercase_ ): A__ : Dict = (x_val + 1) ** (size - col - 1) A__ : Any = y_val A__ : Union[str, Any] = solve(lowercase_ , lowercase_ ) def interpolated_func(lowercase_: int ) -> int: return sum( round(coeffs[x_val][0] ) * (var ** (size - x_val - 1)) for x_val in range(lowercase_ ) ) return interpolated_func def UpperCamelCase (lowercase_: int ) -> int: return ( 1 - variable + variable**2 - variable**3 + variable**4 - variable**5 + variable**6 - variable**7 + variable**8 - variable**9 + variable**10 ) def UpperCamelCase (lowercase_: Callable[[int], int] = question_function , lowercase_: int = 10 ) -> int: A__ : list[int] = [func(lowercase_ ) for x_val in range(1 , order + 1 )] A__ : list[Callable[[int], int]] = [ interpolate(data_points[:max_coeff] ) for max_coeff in range(1 , order + 1 ) ] A__ : int = 0 A__ : Callable[[int], int] A__ : int for poly in polynomials: A__ : List[str] = 1 while func(lowercase_ ) == poly(lowercase_ ): x_val += 1 ret += poly(lowercase_ ) return ret if __name__ == "__main__": print(f'''{solution() = }''')
64
1
from __future__ import annotations from math import pi def UpperCamelCase (lowercase_: float , lowercase_: float , lowercase_: float ) -> dict[str, float]: if (inductance, frequency, reactance).count(0 ) != 1: raise ValueError("""One and only one argument must be 0""" ) if inductance < 0: raise ValueError("""Inductance cannot be negative""" ) if frequency < 0: raise ValueError("""Frequency cannot be negative""" ) if reactance < 0: raise ValueError("""Inductive reactance cannot be negative""" ) if inductance == 0: return {"inductance": reactance / (2 * pi * frequency)} elif frequency == 0: return {"frequency": reactance / (2 * pi * inductance)} elif reactance == 0: return {"reactance": 2 * pi * frequency * inductance} else: raise ValueError("""Exactly one argument must be 0""" ) if __name__ == "__main__": import doctest doctest.testmod()
64
from functools import lru_cache @lru_cache def UpperCamelCase (lowercase_: int ) -> int: if num < 0: raise ValueError("""Number should not be negative.""" ) return 1 if num in (0, 1) else num * factorial(num - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
64
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_tf_available, is_torch_available, ) A_ : Dict = { 'configuration_speech_to_text': ['SPEECH_TO_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Speech2TextConfig'], 'processing_speech_to_text': ['Speech2TextProcessor'], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Union[str, Any] = ['Speech2TextTokenizer'] try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Optional[int] = ['Speech2TextFeatureExtractor'] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : int = [ 'TF_SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFSpeech2TextForConditionalGeneration', 'TFSpeech2TextModel', 'TFSpeech2TextPreTrainedModel', ] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Any = [ 'SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST', 'Speech2TextForConditionalGeneration', 'Speech2TextModel', 'Speech2TextPreTrainedModel', ] if TYPE_CHECKING: from .configuration_speech_to_text import SPEECH_TO_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, SpeechaTextConfig from .processing_speech_to_text import SpeechaTextProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speech_to_text import SpeechaTextTokenizer try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_speech_to_text import SpeechaTextFeatureExtractor try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_speech_to_text import ( TF_SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, TFSpeechaTextForConditionalGeneration, TFSpeechaTextModel, TFSpeechaTextPreTrainedModel, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speech_to_text import ( SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechaTextForConditionalGeneration, SpeechaTextModel, SpeechaTextPreTrainedModel, ) else: import sys A_ : List[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
64
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class _a (datasets.BeamBasedBuilder ): '''simple docstring''' def __A ( self ): return datasets.DatasetInfo( features=datasets.Features({"""content""": datasets.Value("""string""" )} ) , supervised_keys=A__ , ) def __A ( self , A__ , A__ ): return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"""examples""": get_test_dummy_examples()} )] def __A ( self , A__ , A__ ): import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(A__ ) class _a (datasets.BeamBasedBuilder ): '''simple docstring''' def __A ( self ): return datasets.DatasetInfo( features=datasets.Features({"""a""": datasets.Sequence({"""b""": datasets.Value("""string""" )} )} ) , supervised_keys=A__ , ) def __A ( self , A__ , A__ ): return [ datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"""examples""": get_test_nested_examples()} ) ] def __A ( self , A__ , A__ ): import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(A__ ) def UpperCamelCase () -> Dict: return [(i, {"content": content}) for i, content in enumerate(["""foo""", """bar""", """foobar"""] )] def UpperCamelCase () -> Tuple: return [(i, {"a": {"b": [content]}}) for i, content in enumerate(["""foo""", """bar""", """foobar"""] )] class _a (__magic_name__ ): '''simple docstring''' @require_beam def __A ( self ): A__ : Dict = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: A__ : int = DummyBeamDataset(cache_dir=A__ , beam_runner="""DirectRunner""" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(A__ , builder.name , """default""" , """0.0.0""" , F"""{builder.name}-train.arrow""" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"""content""": datasets.Value("""string""" )} ) ) A__ : int = builder.as_dataset() self.assertEqual(dset["""train"""].num_rows , A__ ) self.assertEqual(dset["""train"""].info.splits["""train"""].num_examples , A__ ) self.assertDictEqual(dset["""train"""][0] , get_test_dummy_examples()[0][1] ) self.assertDictEqual( dset["""train"""][expected_num_examples - 1] , get_test_dummy_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(A__ , builder.name , """default""" , """0.0.0""" , """dataset_info.json""" ) ) ) del dset @require_beam def __A ( self ): import apache_beam as beam A__ : int = beam.io.parquetio.WriteToParquet A__ : List[str] = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: A__ : str = DummyBeamDataset(cache_dir=A__ , beam_runner="""DirectRunner""" ) with patch("""apache_beam.io.parquetio.WriteToParquet""" ) as write_parquet_mock: A__ : Optional[Any] = partial(A__ , num_shards=2 ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join( A__ , builder.name , """default""" , """0.0.0""" , F"""{builder.name}-train-00000-of-00002.arrow""" ) ) ) self.assertTrue( os.path.exists( os.path.join( A__ , builder.name , """default""" , """0.0.0""" , F"""{builder.name}-train-00000-of-00002.arrow""" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"""content""": datasets.Value("""string""" )} ) ) A__ : Optional[int] = builder.as_dataset() self.assertEqual(dset["""train"""].num_rows , A__ ) self.assertEqual(dset["""train"""].info.splits["""train"""].num_examples , A__ ) # Order is not preserved when sharding, so we just check that all the elements are there self.assertListEqual(sorted(dset["""train"""]["""content"""] ) , sorted(["""foo""", """bar""", """foobar"""] ) ) self.assertTrue( os.path.exists(os.path.join(A__ , builder.name , """default""" , """0.0.0""" , """dataset_info.json""" ) ) ) del dset @require_beam def __A ( self ): with tempfile.TemporaryDirectory() as tmp_cache_dir: A__ : int = DummyBeamDataset(cache_dir=A__ ) self.assertRaises(datasets.builder.MissingBeamOptions , builder.download_and_prepare ) @require_beam def __A ( self ): A__ : List[Any] = len(get_test_nested_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: A__ : Optional[int] = NestedBeamDataset(cache_dir=A__ , beam_runner="""DirectRunner""" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(A__ , builder.name , """default""" , """0.0.0""" , F"""{builder.name}-train.arrow""" ) ) ) self.assertDictEqual( builder.info.features , datasets.Features({"""a""": datasets.Sequence({"""b""": datasets.Value("""string""" )} )} ) ) A__ : Optional[int] = builder.as_dataset() self.assertEqual(dset["""train"""].num_rows , A__ ) self.assertEqual(dset["""train"""].info.splits["""train"""].num_examples , A__ ) self.assertDictEqual(dset["""train"""][0] , get_test_nested_examples()[0][1] ) self.assertDictEqual( dset["""train"""][expected_num_examples - 1] , get_test_nested_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(A__ , builder.name , """default""" , """0.0.0""" , """dataset_info.json""" ) ) ) del dset
64
1
A_ : int = tuple[float, float, float] A_ : Optional[Any] = tuple[float, float, float] def UpperCamelCase (lowercase_: Pointad , lowercase_: Pointad ) -> Vectorad: A__ : List[Any] = end_pointa[0] - end_pointa[0] A__ : Union[str, Any] = end_pointa[1] - end_pointa[1] A__ : List[str] = end_pointa[2] - end_pointa[2] return (x, y, z) def UpperCamelCase (lowercase_: Vectorad , lowercase_: Vectorad ) -> Vectorad: A__ : Optional[int] = ab[1] * ac[2] - ab[2] * ac[1] # *i A__ : Dict = (ab[0] * ac[2] - ab[2] * ac[0]) * -1 # *j A__ : Dict = ab[0] * ac[1] - ab[1] * ac[0] # *k return (x, y, z) def UpperCamelCase (lowercase_: Vectorad , lowercase_: int ) -> bool: return tuple(round(lowercase_ , lowercase_ ) for x in vector ) == (0, 0, 0) def UpperCamelCase (lowercase_: Pointad , lowercase_: Pointad , lowercase_: Pointad , lowercase_: int = 10 ) -> bool: A__ : Tuple = create_vector(lowercase_ , lowercase_ ) A__ : Dict = create_vector(lowercase_ , lowercase_ ) return is_zero_vector(get_ad_vectors_cross(lowercase_ , lowercase_ ) , lowercase_ )
64
import warnings from ...utils import logging from .image_processing_poolformer import PoolFormerImageProcessor A_ : Union[str, Any] = logging.get_logger(__name__) class _a (__magic_name__ ): '''simple docstring''' def __init__( self , *A__ , **A__ ): warnings.warn( """The class PoolFormerFeatureExtractor is deprecated and will be removed in version 5 of Transformers.""" """ Please use PoolFormerImageProcessor instead.""" , A__ , ) super().__init__(*A__ , **A__ )
64
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING A_ : Union[str, Any] = logging.get_logger(__name__) A_ : Any = { 'microsoft/table-transformer-detection': ( 'https://huggingface.co/microsoft/table-transformer-detection/resolve/main/config.json' ), } class _a (__magic_name__ ): '''simple docstring''' UpperCAmelCase__: Tuple = '''table-transformer''' UpperCAmelCase__: Optional[int] = ['''past_key_values'''] UpperCAmelCase__: List[Any] = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', } def __init__( self , A__=True , A__=None , A__=3 , A__=100 , A__=6 , A__=2048 , A__=8 , A__=6 , A__=2048 , A__=8 , A__=0.0 , A__=0.0 , A__=True , A__="relu" , A__=256 , A__=0.1 , A__=0.0 , A__=0.0 , A__=0.0_2 , A__=1.0 , A__=False , A__="sine" , A__="resnet50" , A__=True , A__=False , A__=1 , A__=5 , A__=2 , A__=1 , A__=1 , A__=5 , A__=2 , A__=0.1 , **A__ , ): if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) A__ : Dict = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(A__ , A__ ): A__ : Any = backbone_config.get("""model_type""" ) A__ : int = CONFIG_MAPPING[backbone_model_type] A__ : str = config_class.from_dict(A__ ) # set timm attributes to None A__ , A__ , A__ : Any = None, None, None A__ : Optional[int] = use_timm_backbone A__ : List[str] = backbone_config A__ : Any = num_channels A__ : Union[str, Any] = num_queries A__ : Optional[int] = d_model A__ : Union[str, Any] = encoder_ffn_dim A__ : str = encoder_layers A__ : List[Any] = encoder_attention_heads A__ : str = decoder_ffn_dim A__ : Any = decoder_layers A__ : Tuple = decoder_attention_heads A__ : List[Any] = dropout A__ : int = attention_dropout A__ : Any = activation_dropout A__ : Tuple = activation_function A__ : Tuple = init_std A__ : Union[str, Any] = init_xavier_std A__ : int = encoder_layerdrop A__ : List[str] = decoder_layerdrop A__ : Any = encoder_layers A__ : List[str] = auxiliary_loss A__ : List[str] = position_embedding_type A__ : Tuple = backbone A__ : List[Any] = use_pretrained_backbone A__ : List[Any] = dilation # Hungarian matcher A__ : Tuple = class_cost A__ : Union[str, Any] = bbox_cost A__ : List[Any] = giou_cost # Loss coefficients A__ : Any = mask_loss_coefficient A__ : int = dice_loss_coefficient A__ : Union[str, Any] = bbox_loss_coefficient A__ : str = giou_loss_coefficient A__ : int = eos_coefficient super().__init__(is_encoder_decoder=A__ , **A__ ) @property def __A ( self ): return self.encoder_attention_heads @property def __A ( self ): return self.d_model class _a (__magic_name__ ): '''simple docstring''' UpperCAmelCase__: Dict = version.parse('''1.11''' ) @property def __A ( self ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def __A ( self ): return 1e-5 @property def __A ( self ): return 12
64
import inspect import logging import os import random import shutil import tempfile import unittest import pytest import torch from torch import nn from torch.utils.data import DataLoader, TensorDataset from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_cuda from accelerate.utils import ProjectConfiguration, set_seed A_ : Any = logging.getLogger(__name__) def UpperCamelCase (lowercase_: Optional[Any]=2 , lowercase_: Union[str, Any]=3 , lowercase_: int=16 , lowercase_: int = 10 , lowercase_: int = 2 ) -> int: def get_dataset(lowercase_: Optional[int] ): A__ : Optional[Any] = torch.randn(batch_size * n_batches , 1 ) return TensorDataset(lowercase_ , a * x + b + 0.1 * torch.randn(batch_size * n_batches , 1 ) ) A__ : Dict = get_dataset(lowercase_ ) A__ : Any = get_dataset(lowercase_ ) A__ : Dict = DataLoader(lowercase_ , shuffle=lowercase_ , batch_size=lowercase_ , num_workers=4 ) A__ : Optional[Any] = DataLoader(lowercase_ , shuffle=lowercase_ , batch_size=lowercase_ , num_workers=4 ) return (train_dataloader, valid_dataloader) def UpperCamelCase (lowercase_: Optional[Any] , lowercase_: List[str] , lowercase_: int , lowercase_: int , lowercase_: List[str] , lowercase_: Dict=None ) -> List[Any]: A__ : List[Any] = [] for epoch in range(lowercase_ ): # Train quickly model.train() for batch in dataloader: A__ , A__ : Any = batch A__ : Any = model(lowercase_ ) A__ : Any = torch.nn.functional.mse_loss(lowercase_ , lowercase_ ) accelerator.backward(lowercase_ ) optimizer.step() optimizer.zero_grad() rands.append(random.random() ) # Introduce some randomness if scheduler is not None: scheduler.step() return rands class _a (nn.Module ): '''simple docstring''' def __init__( self ): super().__init__() A__ : str = nn.Parameter(torch.randn(1 ) ) A__ : Any = nn.Parameter(torch.randn(1 ) ) def __A ( self , A__ ): return x * self.a + self.b class _a (unittest.TestCase ): '''simple docstring''' def __A ( self ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) A__ : Optional[Any] = DummyModel() A__ : Optional[Any] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) A__ , A__ : str = dummy_dataloaders() A__ : Dict = ProjectConfiguration(total_limit=1 , project_dir=A__ , automatic_checkpoint_naming=A__ ) # Train baseline A__ : List[str] = Accelerator(project_config=A__ ) A__ , A__ , A__ , A__ : Any = accelerator.prepare( A__ , A__ , A__ , A__ ) # Save initial accelerator.save_state() # Save second state accelerator.save_state() self.assertEqual(len(os.listdir(accelerator.project_dir ) ) , 1 ) def __A ( self ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) A__ : str = DummyModel() A__ : Optional[int] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) A__ , A__ : int = dummy_dataloaders() # Train baseline A__ : str = Accelerator() A__ , A__ , A__ , A__ : List[str] = accelerator.prepare( A__ , A__ , A__ , A__ ) # Save initial A__ : List[Any] = os.path.join(A__ , """initial""" ) accelerator.save_state(A__ ) ((A__) , (A__)) : str = model.a.item(), model.b.item() A__ : Dict = optimizer.state_dict() A__ : List[str] = train(3 , A__ , A__ , A__ , A__ ) ((A__) , (A__)) : str = model.a.item(), model.b.item() A__ : Any = optimizer.state_dict() # Train partially set_seed(42 ) A__ : Optional[int] = DummyModel() A__ : Dict = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) A__ , A__ : Dict = dummy_dataloaders() A__ : List[str] = Accelerator() A__ , A__ , A__ , A__ : Optional[Any] = accelerator.prepare( A__ , A__ , A__ , A__ ) accelerator.load_state(A__ ) ((A__) , (A__)) : Tuple = model.a.item(), model.b.item() A__ : Union[str, Any] = optimizer.state_dict() self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) A__ : List[str] = train(2 , A__ , A__ , A__ , A__ ) # Save everything A__ : Optional[int] = os.path.join(A__ , """checkpoint""" ) accelerator.save_state(A__ ) # Load everything back in and make sure all states work accelerator.load_state(A__ ) test_rands += train(1 , A__ , A__ , A__ , A__ ) ((A__) , (A__)) : Union[str, Any] = model.a.item(), model.b.item() A__ : Optional[int] = optimizer.state_dict() self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) def __A ( self ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) A__ : int = DummyModel() A__ : Optional[int] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) A__ , A__ : List[str] = dummy_dataloaders() A__ : str = ProjectConfiguration(automatic_checkpoint_naming=A__ ) # Train baseline A__ : Any = Accelerator(project_dir=A__ , project_config=A__ ) A__ , A__ , A__ , A__ : str = accelerator.prepare( A__ , A__ , A__ , A__ ) # Save initial accelerator.save_state() ((A__) , (A__)) : Tuple = model.a.item(), model.b.item() A__ : int = optimizer.state_dict() A__ : int = train(3 , A__ , A__ , A__ , A__ ) ((A__) , (A__)) : Optional[Any] = model.a.item(), model.b.item() A__ : Any = optimizer.state_dict() # Train partially set_seed(42 ) A__ : Dict = DummyModel() A__ : List[Any] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) A__ , A__ : Union[str, Any] = dummy_dataloaders() A__ : List[Any] = ProjectConfiguration(iteration=1 , automatic_checkpoint_naming=A__ ) A__ : Dict = Accelerator(project_dir=A__ , project_config=A__ ) A__ , A__ , A__ , A__ : Union[str, Any] = accelerator.prepare( A__ , A__ , A__ , A__ ) accelerator.load_state(os.path.join(A__ , """checkpoints""" , """checkpoint_0""" ) ) ((A__) , (A__)) : Optional[int] = model.a.item(), model.b.item() A__ : Tuple = optimizer.state_dict() self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) A__ : str = train(2 , A__ , A__ , A__ , A__ ) # Save everything accelerator.save_state() # Load everything back in and make sure all states work accelerator.load_state(os.path.join(A__ , """checkpoints""" , """checkpoint_1""" ) ) test_rands += train(1 , A__ , A__ , A__ , A__ ) ((A__) , (A__)) : Optional[int] = model.a.item(), model.b.item() A__ : List[Any] = optimizer.state_dict() self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) def __A ( self ): A__ : Union[str, Any] = torch.tensor([1, 2, 3] ) A__ : int = torch.tensor([2, 3, 4] ) A__ : List[Any] = DummyModel() A__ : List[Any] = torch.optim.Adam(net.parameters() ) A__ : Tuple = Accelerator() with self.assertRaises(A__ ) as ve: accelerator.register_for_checkpointing(A__ , A__ , A__ , A__ ) A__ : Any = str(ve.exception ) self.assertTrue("""Item at index 0""" in message ) self.assertTrue("""Item at index 1""" in message ) self.assertFalse("""Item at index 2""" in message ) self.assertFalse("""Item at index 3""" in message ) def __A ( self ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) A__ : Any = DummyModel() A__ : Union[str, Any] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) A__ : Dict = torch.optim.lr_scheduler.StepLR(A__ , step_size=1 , gamma=0.9_9 ) A__ , A__ : List[Any] = dummy_dataloaders() A__ : Tuple = ProjectConfiguration(automatic_checkpoint_naming=A__ ) # Train baseline A__ : Optional[Any] = Accelerator(project_dir=A__ , project_config=A__ ) A__ , A__ , A__ , A__ , A__ : Union[str, Any] = accelerator.prepare( A__ , A__ , A__ , A__ , A__ ) # Save initial accelerator.save_state() A__ : Tuple = scheduler.state_dict() train(3 , A__ , A__ , A__ , A__ , A__ ) self.assertNotEqual(A__ , scheduler.state_dict() ) # Load everything back in and make sure all states work accelerator.load_state(os.path.join(A__ , """checkpoints""" , """checkpoint_0""" ) ) self.assertEqual(A__ , scheduler.state_dict() ) def __A ( self ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) A__ : Optional[Any] = DummyModel() A__ : int = ProjectConfiguration(automatic_checkpoint_naming=A__ , total_limit=2 ) # Train baseline A__ : List[str] = Accelerator(project_dir=A__ , project_config=A__ ) A__ : Union[str, Any] = accelerator.prepare(A__ ) # Save 3 states: for _ in range(11 ): accelerator.save_state() self.assertTrue(not os.path.exists(os.path.join(A__ , """checkpoints""" , """checkpoint_0""" ) ) ) self.assertTrue(os.path.exists(os.path.join(A__ , """checkpoints""" , """checkpoint_9""" ) ) ) self.assertTrue(os.path.exists(os.path.join(A__ , """checkpoints""" , """checkpoint_10""" ) ) ) @require_cuda def __A ( self ): A__ : Dict = ["""torchrun""", F"""--nproc_per_node={torch.cuda.device_count()}""", inspect.getfile(self.__class__ )] execute_subprocess_async(A__ , env=os.environ.copy() ) if __name__ == "__main__": A_ : List[str] = '/tmp/accelerate/state_checkpointing' A_ : Optional[Any] = DummyModel() A_ : Union[str, Any] = torch.optim.Adam(params=model.parameters(), lr=1E-3) A_ : str = torch.optim.lr_scheduler.StepLR(optimizer, step_size=1, gamma=0.99) A_ , A_ : List[Any] = dummy_dataloaders() A_ : int = ProjectConfiguration(automatic_checkpoint_naming=True) # Train baseline A_ : List[str] = Accelerator(project_dir=savedir, project_config=project_config, mixed_precision='no') if accelerator.process_index == 0: if os.path.exists(savedir): shutil.rmtree(savedir) os.makedirs(savedir) A_ , A_ , A_ , A_ , A_ : List[Any] = accelerator.prepare( model, optimizer, train_dataloader, valid_dataloader, scheduler ) A_ , A_ : Dict = accelerator.prepare(model, optimizer) train(3, model, train_dataloader, optimizer, accelerator, scheduler) # Check that the intial optimizer is loaded on the GPU for group in optimizer.param_groups: A_ : str = group['params'][0].device break assert param_device.type == accelerator.device.type A_ : Optional[Any] = model.cpu() accelerator.wait_for_everyone() accelerator.save_state() accelerator.wait_for_everyone() # Check CPU state accelerator.load_state(os.path.join(savedir, 'checkpoints', 'checkpoint_0'), map_location='cpu') for group in optimizer.param_groups: A_ : str = group['params'][0].device break assert ( param_device.type == torch.device('cpu').type ), f"Loaded optimizer states did not match, expected to be loaded on the CPU but got {param_device}" # Check device state model.to(accelerator.device) accelerator.load_state(os.path.join(savedir, 'checkpoints', 'checkpoint_0'), map_location='on_device') for group in optimizer.param_groups: A_ : Tuple = group['params'][0].device break assert ( param_device.type == accelerator.device.type ), f"Loaded optimizer states did not match, expected to be loaded on {accelerator.device} but got {param_device}" # Check error with pytest.raises(TypeError, match='Unsupported optimizer map location passed'): accelerator.load_state(os.path.join(savedir, 'checkpoints', 'checkpoint_0'), map_location='invalid') accelerator.wait_for_everyone() if accelerator.process_index == 0: shutil.rmtree(savedir) accelerator.wait_for_everyone()
64
1
import argparse import torch from safetensors.torch import load_file from diffusers import StableDiffusionPipeline def UpperCamelCase (lowercase_: Optional[int] , lowercase_: str , lowercase_: Any , lowercase_: List[Any] , lowercase_: Tuple ) -> Any: # load base model A__ : List[str] = StableDiffusionPipeline.from_pretrained(lowercase_ , torch_dtype=torch.floataa ) # load LoRA weight from .safetensors A__ : Dict = load_file(lowercase_ ) A__ : List[Any] = [] # directly update weight in diffusers model for key in state_dict: # it is suggested to print out the key, it usually will be something like below # "lora_te_text_model_encoder_layers_0_self_attn_k_proj.lora_down.weight" # as we have set the alpha beforehand, so just skip if ".alpha" in key or key in visited: continue if "text" in key: A__ : Optional[Any] = key.split(""".""" )[0].split(LORA_PREFIX_TEXT_ENCODER + """_""" )[-1].split("""_""" ) A__ : str = pipeline.text_encoder else: A__ : Any = key.split(""".""" )[0].split(LORA_PREFIX_UNET + """_""" )[-1].split("""_""" ) A__ : Union[str, Any] = pipeline.unet # find the target layer A__ : List[str] = layer_infos.pop(0 ) while len(lowercase_ ) > -1: try: A__ : Optional[Any] = curr_layer.__getattr__(lowercase_ ) if len(lowercase_ ) > 0: A__ : List[str] = layer_infos.pop(0 ) elif len(lowercase_ ) == 0: break except Exception: if len(lowercase_ ) > 0: temp_name += "_" + layer_infos.pop(0 ) else: A__ : Dict = layer_infos.pop(0 ) A__ : str = [] if "lora_down" in key: pair_keys.append(key.replace("""lora_down""" , """lora_up""" ) ) pair_keys.append(lowercase_ ) else: pair_keys.append(lowercase_ ) pair_keys.append(key.replace("""lora_up""" , """lora_down""" ) ) # update weight if len(state_dict[pair_keys[0]].shape ) == 4: A__ : List[str] = state_dict[pair_keys[0]].squeeze(3 ).squeeze(2 ).to(torch.floataa ) A__ : Optional[int] = state_dict[pair_keys[1]].squeeze(3 ).squeeze(2 ).to(torch.floataa ) curr_layer.weight.data += alpha * torch.mm(lowercase_ , lowercase_ ).unsqueeze(2 ).unsqueeze(3 ) else: A__ : str = state_dict[pair_keys[0]].to(torch.floataa ) A__ : List[Any] = state_dict[pair_keys[1]].to(torch.floataa ) curr_layer.weight.data += alpha * torch.mm(lowercase_ , lowercase_ ) # update visited list for item in pair_keys: visited.append(lowercase_ ) return pipeline if __name__ == "__main__": A_ : Dict = argparse.ArgumentParser() parser.add_argument( '--base_model_path', default=None, type=str, required=True, help='Path to the base model in diffusers format.' ) parser.add_argument( '--checkpoint_path', default=None, type=str, required=True, help='Path to the checkpoint to convert.' ) parser.add_argument('--dump_path', default=None, type=str, required=True, help='Path to the output model.') parser.add_argument( '--lora_prefix_unet', default='lora_unet', type=str, help='The prefix of UNet weight in safetensors' ) parser.add_argument( '--lora_prefix_text_encoder', default='lora_te', type=str, help='The prefix of text encoder weight in safetensors', ) parser.add_argument('--alpha', default=0.75, type=float, help='The merging ratio in W = W0 + alpha * deltaW') parser.add_argument( '--to_safetensors', action='store_true', help='Whether to store pipeline in safetensors format or not.' ) parser.add_argument('--device', type=str, help='Device to use (e.g. cpu, cuda:0, cuda:1, etc.)') A_ : str = parser.parse_args() A_ : int = args.base_model_path A_ : Any = args.checkpoint_path A_ : Optional[int] = args.dump_path A_ : Dict = args.lora_prefix_unet A_ : Optional[int] = args.lora_prefix_text_encoder A_ : List[str] = args.alpha A_ : int = convert(base_model_path, checkpoint_path, lora_prefix_unet, lora_prefix_text_encoder, alpha) A_ : int = pipe.to(args.device) pipe.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
64
def UpperCamelCase (lowercase_: str , lowercase_: str ) -> bool: A__ : Union[str, Any] = len(lowercase_ ) A__ : List[Any] = len(lowercase_ ) A__ : List[Any] = [[False for _ in range(m + 1 )] for _ in range(n + 1 )] A__ : str = True for i in range(lowercase_ ): for j in range(m + 1 ): if dp[i][j]: if j < m and a[i].upper() == b[j]: A__ : int = True if a[i].islower(): A__ : Dict = True return dp[n][m] if __name__ == "__main__": import doctest doctest.testmod()
64
1
def UpperCamelCase (lowercase_: int ) -> bool: return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
64
import itertools import random import unittest import numpy as np from transformers import ASTFeatureExtractor from transformers.testing_utils import require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin A_ : Dict = random.Random() if is_torch_available(): import torch def UpperCamelCase (lowercase_: Tuple , lowercase_: Tuple=1.0 , lowercase_: Dict=None , lowercase_: int=None ) -> str: if rng is None: A__ : Optional[Any] = global_rng A__ : List[str] = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class _a (unittest.TestCase ): '''simple docstring''' def __init__( self , A__ , A__=7 , A__=400 , A__=2000 , A__=1 , A__=0.0 , A__=1_6000 , A__=True , A__=True , ): A__ : Any = parent A__ : Optional[int] = batch_size A__ : Union[str, Any] = min_seq_length A__ : Dict = max_seq_length A__ : Tuple = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) A__ : str = feature_size A__ : Optional[int] = padding_value A__ : List[str] = sampling_rate A__ : List[str] = return_attention_mask A__ : int = do_normalize def __A ( self ): return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def __A ( self , A__=False , A__=False ): def _flatten(A__ ): return list(itertools.chain(*A__ ) ) if equal_length: A__ : Dict = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size A__ : Union[str, Any] = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: A__ : Optional[int] = [np.asarray(A__ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class _a (__magic_name__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__: int = ASTFeatureExtractor def __A ( self ): A__ : Optional[Any] = ASTFeatureExtractionTester(self ) def __A ( self ): # Tests that all call wrap to encode_plus and batch_encode_plus A__ : Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 A__ : List[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] A__ : Optional[Any] = [np.asarray(A__ ) for speech_input in speech_inputs] # Test not batched input A__ : Tuple = feat_extract(speech_inputs[0] , return_tensors="""np""" ).input_values A__ : Tuple = feat_extract(np_speech_inputs[0] , return_tensors="""np""" ).input_values self.assertTrue(np.allclose(A__ , A__ , atol=1e-3 ) ) # Test batched A__ : Tuple = feat_extract(A__ , padding=A__ , return_tensors="""np""" ).input_values A__ : Tuple = feat_extract(A__ , padding=A__ , return_tensors="""np""" ).input_values for enc_seq_a, enc_seq_a in zip(A__ , A__ ): self.assertTrue(np.allclose(A__ , A__ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. A__ : int = [floats_list((1, x) )[0] for x in (800, 800, 800)] A__ : List[str] = np.asarray(A__ ) A__ : Union[str, Any] = feat_extract(A__ , return_tensors="""np""" ).input_values A__ : Optional[Any] = feat_extract(A__ , return_tensors="""np""" ).input_values for enc_seq_a, enc_seq_a in zip(A__ , A__ ): self.assertTrue(np.allclose(A__ , A__ , atol=1e-3 ) ) @require_torch def __A ( self ): import torch A__ : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) A__ : Tuple = np.random.rand(100 ).astype(np.floataa ) A__ : Tuple = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: A__ : List[str] = feature_extractor.pad([{"""input_values""": inputs}] , return_tensors="""np""" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) A__ : Any = feature_extractor.pad([{"""input_values""": inputs}] , return_tensors="""pt""" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def __A ( self , A__ ): from datasets import load_dataset A__ : str = load_dataset("""hf-internal-testing/librispeech_asr_dummy""" , """clean""" , split="""validation""" ) # automatic decoding with librispeech A__ : str = ds.sort("""id""" ).select(range(A__ ) )[:num_samples]["""audio"""] return [x["array"] for x in speech_samples] @require_torch def __A ( self ): # fmt: off A__ : Optional[Any] = torch.tensor( [-0.9_8_9_4, -1.2_7_7_6, -0.9_0_6_6, -1.2_7_7_6, -0.9_3_4_9, -1.2_6_0_9, -1.0_3_8_6, -1.2_7_7_6, -1.1_5_6_1, -1.2_7_7_6, -1.2_0_5_2, -1.2_7_2_3, -1.2_1_9_0, -1.2_1_3_2, -1.2_7_7_6, -1.1_1_3_3, -1.1_9_5_3, -1.1_3_4_3, -1.1_5_8_4, -1.2_2_0_3, -1.1_7_7_0, -1.2_4_7_4, -1.2_3_8_1, -1.1_9_3_6, -0.9_2_7_0, -0.8_3_1_7, -0.8_0_4_9, -0.7_7_0_6, -0.7_5_6_5, -0.7_8_6_9] ) # fmt: on A__ : Any = self._load_datasamples(1 ) A__ : Tuple = ASTFeatureExtractor() A__ : Dict = feature_extractor(A__ , return_tensors="""pt""" ).input_values self.assertEquals(input_values.shape , (1, 1024, 128) ) self.assertTrue(torch.allclose(input_values[0, 0, :30] , A__ , atol=1e-4 ) )
64
1
import importlib import os import sys # This is required to make the module import works (when the python process is running from the root of the repo) sys.path.append('.') def UpperCamelCase (lowercase_: Union[str, Any] ) -> Optional[int]: A__ : Union[str, Any] = test_file.split(os.path.sep ) if components[0:2] != ["tests", "models"]: raise ValueError( """`test_file` should start with `tests/models/` (with `/` being the OS specific path separator). Got """ f"""{test_file} instead.""" ) A__ : Dict = components[-1] if not test_fn.endswith("""py""" ): raise ValueError(f"""`test_file` should be a python file. Got {test_fn} instead.""" ) if not test_fn.startswith("""test_modeling_""" ): raise ValueError( f"""`test_file` should point to a file name of the form `test_modeling_*.py`. Got {test_fn} instead.""" ) A__ : int = components[:-1] + [test_fn.replace(""".py""" , """""" )] A__ : Tuple = """.""".join(lowercase_ ) return test_module_path def UpperCamelCase (lowercase_: Any ) -> List[str]: A__ : Optional[Any] = get_module_path(lowercase_ ) A__ : int = importlib.import_module(lowercase_ ) return test_module def UpperCamelCase (lowercase_: Optional[Any] ) -> Union[str, Any]: A__ : List[Any] = [] A__ : List[str] = get_test_module(lowercase_ ) for attr in dir(lowercase_ ): if attr.endswith("""ModelTester""" ): tester_classes.append(getattr(lowercase_ , lowercase_ ) ) # sort with class names return sorted(lowercase_ , key=lambda lowercase_ : x.__name__ ) def UpperCamelCase (lowercase_: Tuple ) -> List[str]: A__ : List[str] = [] A__ : Tuple = get_test_module(lowercase_ ) for attr in dir(lowercase_ ): A__ : List[str] = getattr(lowercase_ , lowercase_ ) # (TF/Flax)ModelTesterMixin is also an attribute in specific model test module. Let's exclude them by checking # `all_model_classes` is not empty (which also excludes other special classes). A__ : List[Any] = getattr(lowercase_ , """all_model_classes""" , [] ) if len(lowercase_ ) > 0: test_classes.append(lowercase_ ) # sort with class names return sorted(lowercase_ , key=lambda lowercase_ : x.__name__ ) def UpperCamelCase (lowercase_: Optional[Any] ) -> int: A__ : List[str] = get_test_classes(lowercase_ ) A__ : Optional[Any] = set() for test_class in test_classes: model_classes.update(test_class.all_model_classes ) # sort with class names return sorted(lowercase_ , key=lambda lowercase_ : x.__name__ ) def UpperCamelCase (lowercase_: Union[str, Any] ) -> int: A__ : Any = test_class() if hasattr(lowercase_ , """setUp""" ): test.setUp() A__ : Any = None if hasattr(lowercase_ , """model_tester""" ): # `(TF/Flax)ModelTesterMixin` has this attribute default to `None`. Let's skip this case. if test.model_tester is not None: A__ : Tuple = test.model_tester.__class__ return model_tester def UpperCamelCase (lowercase_: int , lowercase_: Union[str, Any] ) -> Union[str, Any]: A__ : Any = get_test_classes(lowercase_ ) A__ : Optional[Any] = [] for test_class in test_classes: if model_class in test_class.all_model_classes: target_test_classes.append(lowercase_ ) # sort with class names return sorted(lowercase_ , key=lambda lowercase_ : x.__name__ ) def UpperCamelCase (lowercase_: List[Any] , lowercase_: Dict ) -> int: A__ : Dict = get_test_classes_for_model(lowercase_ , lowercase_ ) A__ : Optional[Any] = [] for test_class in test_classes: A__ : Union[str, Any] = get_model_tester_from_test_class(lowercase_ ) if tester_class is not None: tester_classes.append(lowercase_ ) # sort with class names return sorted(lowercase_ , key=lambda lowercase_ : x.__name__ ) def UpperCamelCase (lowercase_: Union[str, Any] ) -> int: A__ : Optional[int] = get_test_classes(lowercase_ ) A__ : Tuple = {test_class: get_model_tester_from_test_class(lowercase_ ) for test_class in test_classes} return test_tester_mapping def UpperCamelCase (lowercase_: int ) -> str: A__ : Union[str, Any] = get_model_classes(lowercase_ ) A__ : List[str] = { model_class: get_test_classes_for_model(lowercase_ , lowercase_ ) for model_class in model_classes } return model_test_mapping def UpperCamelCase (lowercase_: Union[str, Any] ) -> Optional[int]: A__ : Optional[int] = get_model_classes(lowercase_ ) A__ : Optional[Any] = { model_class: get_tester_classes_for_model(lowercase_ , lowercase_ ) for model_class in model_classes } return model_to_tester_mapping def UpperCamelCase (lowercase_: Tuple ) -> str: if isinstance(lowercase_ , lowercase_ ): return o elif isinstance(lowercase_ , lowercase_ ): return o.__name__ elif isinstance(lowercase_ , (list, tuple) ): return [to_json(lowercase_ ) for x in o] elif isinstance(lowercase_ , lowercase_ ): return {to_json(lowercase_ ): to_json(lowercase_ ) for k, v in o.items()} else: return o
64
from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class _a (__magic_name__ , __magic_name__ , __magic_name__ ): '''simple docstring''' UpperCAmelCase__: str = [r'''h\.\d+\.attn\.bias''', r'''h\.\d+\.attn\.masked_bias'''] @register_to_config def __init__( self , A__ , A__ , A__ = None , A__ = 5_0257 , A__ = 1024 , A__ = 768 , A__ = 12 , A__ = 12 , A__ = None , A__ = "gelu_new" , A__ = 0.1 , A__ = 0.1 , A__ = 0.1 , A__ = 1e-5 , A__ = 0.0_2 , A__ = True , A__ = True , A__ = False , A__ = False , ): super().__init__() A__ : Union[str, Any] = prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( F"""`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and""" F""" `n_embd`: {n_embd} are not equal.""" ) A__ : str = prefix_inner_dim A__ : Optional[Any] = prefix_hidden_dim A__ : Tuple = ( nn.Linear(self.prefix_inner_dim , self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) A__ : int = ( nn.Linear(self.prefix_hidden_dim , A__ ) if self.prefix_hidden_dim is not None else nn.Identity() ) A__ : Tuple = GPTaConfig( vocab_size=A__ , n_positions=A__ , n_embd=A__ , n_layer=A__ , n_head=A__ , n_inner=A__ , activation_function=A__ , resid_pdrop=A__ , embd_pdrop=A__ , attn_pdrop=A__ , layer_norm_epsilon=A__ , initializer_range=A__ , scale_attn_weights=A__ , use_cache=A__ , scale_attn_by_inverse_layer_idx=A__ , reorder_and_upcast_attn=A__ , ) A__ : int = GPTaLMHeadModel(A__ ) def __A ( self , A__ , A__ , A__ = None , A__ = None , ): A__ : List[str] = self.transformer.transformer.wte(A__ ) A__ : int = self.encode_prefix(A__ ) A__ : int = self.decode_prefix(A__ ) A__ : Optional[Any] = torch.cat((prefix_embeds, embedding_text) , dim=1 ) if labels is not None: A__ : Any = self.get_dummy_token(input_ids.shape[0] , input_ids.device ) A__ : List[Any] = torch.cat((dummy_token, input_ids) , dim=1 ) A__ : List[str] = self.transformer(inputs_embeds=A__ , labels=A__ , attention_mask=A__ ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def __A ( self , A__ , A__ ): return torch.zeros(A__ , self.prefix_length , dtype=torch.intaa , device=A__ ) def __A ( self , A__ ): return self.encode_prefix(A__ ) @torch.no_grad() def __A ( self , A__ , A__ , A__ ): A__ : List[Any] = torch.split(A__ , 1 , dim=0 ) A__ : Optional[int] = [] A__ : str = [] for feature in features: A__ : Dict = self.decode_prefix(feature.to(A__ ) ) # back to the clip feature # Only support beam search for now A__ , A__ : Union[str, Any] = self.generate_beam( input_embeds=A__ , device=A__ , eos_token_id=A__ ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) A__ : int = torch.stack(A__ ) A__ : List[Any] = torch.stack(A__ ) return generated_tokens, generated_seq_lengths @torch.no_grad() def __A ( self , A__=None , A__=None , A__=None , A__ = 5 , A__ = 67 , A__ = 1.0 , A__ = None , ): A__ : Any = eos_token_id A__ : Any = None A__ : Optional[int] = None A__ : Optional[Any] = torch.ones(A__ , device=A__ , dtype=torch.int ) A__ : Any = torch.zeros(A__ , device=A__ , dtype=torch.bool ) if input_embeds is not None: A__ : Dict = input_embeds else: A__ : str = self.transformer.transformer.wte(A__ ) for i in range(A__ ): A__ : Dict = self.transformer(inputs_embeds=A__ ) A__ : str = outputs.logits A__ : Union[str, Any] = logits[:, -1, :] / (temperature if temperature > 0 else 1.0) A__ : Any = logits.softmax(-1 ).log() if scores is None: A__ , A__ : Optional[int] = logits.topk(A__ , -1 ) A__ : List[Any] = generated.expand(A__ , *generated.shape[1:] ) A__ , A__ : List[Any] = next_tokens.permute(1 , 0 ), scores.squeeze(0 ) if tokens is None: A__ : Optional[Any] = next_tokens else: A__ : List[Any] = tokens.expand(A__ , *tokens.shape[1:] ) A__ : int = torch.cat((tokens, next_tokens) , dim=1 ) else: A__ : Optional[int] = -float(np.inf ) A__ : List[Any] = 0 A__ : str = scores[:, None] + logits seq_lengths[~is_stopped] += 1 A__ : Dict = scores_sum / seq_lengths[:, None] A__ , A__ : List[Any] = scores_sum_average.view(-1 ).topk(A__ , -1 ) A__ : Tuple = next_tokens // scores_sum.shape[1] A__ : Optional[Any] = seq_lengths[next_tokens_source] A__ : List[str] = next_tokens % scores_sum.shape[1] A__ : Optional[int] = next_tokens.unsqueeze(1 ) A__ : int = tokens[next_tokens_source] A__ : List[Any] = torch.cat((tokens, next_tokens) , dim=1 ) A__ : str = generated[next_tokens_source] A__ : Optional[Any] = scores_sum_average * seq_lengths A__ : Union[str, Any] = is_stopped[next_tokens_source] A__ : str = self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] , 1 , -1 ) A__ : Optional[int] = torch.cat((generated, next_token_embed) , dim=1 ) A__ : List[str] = is_stopped + next_tokens.eq(A__ ).squeeze() if is_stopped.all(): break A__ : Dict = scores / seq_lengths A__ : Dict = scores.argsort(descending=A__ ) # tokens tensors are already padded to max_seq_length A__ : Union[str, Any] = [tokens[i] for i in order] A__ : Any = torch.stack(A__ , dim=0 ) A__ : Dict = torch.tensor([seq_lengths[i] for i in order] , dtype=seq_lengths.dtype ) return output_texts, seq_lengths
64
1
def UpperCamelCase (lowercase_: int = 10 , lowercase_: int = 22 ) -> int: A__ : Any = range(1 , lowercase_ ) A__ : str = range(1 , lowercase_ ) return sum( 1 for power in powers for base in bases if len(str(base**power ) ) == power ) if __name__ == "__main__": print(f'''{solution(10, 22) = }''')
64
import io import itertools import json from dataclasses import dataclass from typing import Optional import pyarrow as pa import pyarrow.json as paj import datasets from datasets.table import table_cast from datasets.utils.file_utils import readline A_ : Tuple = datasets.utils.logging.get_logger(__name__) @dataclass class _a (datasets.BuilderConfig ): '''simple docstring''' UpperCAmelCase__: Optional[datasets.Features] = None UpperCAmelCase__: str = "utf-8" UpperCAmelCase__: Optional[str] = None UpperCAmelCase__: Optional[str] = None UpperCAmelCase__: bool = True # deprecated UpperCAmelCase__: Optional[int] = None # deprecated UpperCAmelCase__: int = 10 << 20 # 10MB UpperCAmelCase__: Optional[bool] = None class _a (datasets.ArrowBasedBuilder ): '''simple docstring''' UpperCAmelCase__: List[str] = JsonConfig def __A ( self ): if self.config.block_size is not None: logger.warning("""The JSON loader parameter `block_size` is deprecated. Please use `chunksize` instead""" ) A__ : Union[str, Any] = self.config.block_size if self.config.use_threads is not True: logger.warning( """The JSON loader parameter `use_threads` is deprecated and doesn't have any effect anymore.""" ) if self.config.newlines_in_values is not None: raise ValueError("""The JSON loader parameter `newlines_in_values` is no longer supported""" ) return datasets.DatasetInfo(features=self.config.features ) def __A ( self , A__ ): if not self.config.data_files: raise ValueError(F"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) A__ : int = dl_manager.download_and_extract(self.config.data_files ) if isinstance(A__ , (str, list, tuple) ): A__ : Optional[Any] = data_files if isinstance(A__ , A__ ): A__ : List[str] = [files] A__ : int = [dl_manager.iter_files(A__ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"""files""": files} )] A__ : List[str] = [] for split_name, files in data_files.items(): if isinstance(A__ , A__ ): A__ : Optional[int] = [files] A__ : Optional[int] = [dl_manager.iter_files(A__ ) for file in files] splits.append(datasets.SplitGenerator(name=A__ , gen_kwargs={"""files""": files} ) ) return splits def __A ( self , A__ ): if self.config.features is not None: # adding missing columns for column_name in set(self.config.features ) - set(pa_table.column_names ): A__ : Optional[Any] = self.config.features.arrow_schema.field(A__ ).type A__ : str = pa_table.append_column(A__ , pa.array([None] * len(A__ ) , type=A__ ) ) # more expensive cast to support nested structures with keys in a different order # allows str <-> int/float or str to Audio for example A__ : Optional[int] = table_cast(A__ , self.config.features.arrow_schema ) return pa_table def __A ( self , A__ ): for file_idx, file in enumerate(itertools.chain.from_iterable(A__ ) ): # If the file is one json object and if we need to look at the list of items in one specific field if self.config.field is not None: with open(A__ , encoding=self.config.encoding , errors=self.config.encoding_errors ) as f: A__ : Optional[Any] = json.load(A__ ) # We keep only the field we are interested in A__ : Optional[int] = dataset[self.config.field] # We accept two format: a list of dicts or a dict of lists if isinstance(A__ , (list, tuple) ): A__ : Union[str, Any] = set().union(*[row.keys() for row in dataset] ) A__ : Any = {col: [row.get(A__ ) for row in dataset] for col in keys} else: A__ : Any = dataset A__ : Any = pa.Table.from_pydict(A__ ) yield file_idx, self._cast_table(A__ ) # If the file has one json object per line else: with open(A__ , """rb""" ) as f: A__ : List[str] = 0 # Use block_size equal to the chunk size divided by 32 to leverage multithreading # Set a default minimum value of 16kB if the chunk size is really small A__ : List[str] = max(self.config.chunksize // 32 , 16 << 10 ) A__ : Any = ( self.config.encoding_errors if self.config.encoding_errors is not None else """strict""" ) while True: A__ : Dict = f.read(self.config.chunksize ) if not batch: break # Finish current line try: batch += f.readline() except (AttributeError, io.UnsupportedOperation): batch += readline(A__ ) # PyArrow only accepts utf-8 encoded bytes if self.config.encoding != "utf-8": A__ : List[Any] = batch.decode(self.config.encoding , errors=A__ ).encode("""utf-8""" ) try: while True: try: A__ : str = paj.read_json( io.BytesIO(A__ ) , read_options=paj.ReadOptions(block_size=A__ ) ) break except (pa.ArrowInvalid, pa.ArrowNotImplementedError) as e: if ( isinstance(A__ , pa.ArrowInvalid ) and "straddling" not in str(A__ ) or block_size > len(A__ ) ): raise else: # Increase the block size in case it was too small. # The block size will be reset for the next file. logger.debug( F"""Batch of {len(A__ )} bytes couldn't be parsed with block_size={block_size}. Retrying with block_size={block_size * 2}.""" ) block_size *= 2 except pa.ArrowInvalid as e: try: with open( A__ , encoding=self.config.encoding , errors=self.config.encoding_errors ) as f: A__ : Optional[Any] = json.load(A__ ) except json.JSONDecodeError: logger.error(F"""Failed to read file '{file}' with error {type(A__ )}: {e}""" ) raise e # If possible, parse the file as a list of json objects and exit the loop if isinstance(A__ , A__ ): # list is the only sequence type supported in JSON try: A__ : str = set().union(*[row.keys() for row in dataset] ) A__ : List[str] = {col: [row.get(A__ ) for row in dataset] for col in keys} A__ : int = pa.Table.from_pydict(A__ ) except (pa.ArrowInvalid, AttributeError) as e: logger.error(F"""Failed to read file '{file}' with error {type(A__ )}: {e}""" ) raise ValueError(F"""Not able to read records in the JSON file at {file}.""" ) from None yield file_idx, self._cast_table(A__ ) break else: logger.error(F"""Failed to read file '{file}' with error {type(A__ )}: {e}""" ) raise ValueError( F"""Not able to read records in the JSON file at {file}. """ F"""You should probably indicate the field of the JSON file containing your records. """ F"""This JSON file contain the following fields: {str(list(dataset.keys() ) )}. """ F"""Select the correct one and provide it as `field='XXX'` to the dataset loading method. """ ) from None # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(A__ ) batch_idx += 1
64
1
import os import pytest from attr import dataclass A_ : Union[str, Any] = 'us-east-1' # defaults region @dataclass class _a : '''simple docstring''' UpperCAmelCase__: str UpperCAmelCase__: Any = '''arn:aws:iam::558105141721:role/sagemaker_execution_role''' UpperCAmelCase__: Tuple = { '''task_name''': '''mnli''', '''per_device_train_batch_size''': 16, '''per_device_eval_batch_size''': 16, '''do_train''': True, '''do_eval''': True, '''do_predict''': True, '''output_dir''': '''/opt/ml/model''', '''overwrite_output_dir''': True, '''max_steps''': 5_00, '''save_steps''': 55_00, } UpperCAmelCase__: str = {**hyperparameters, '''max_steps''': 10_00} @property def __A ( self ): if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def __A ( self ): return F"""{self.framework}-transfromers-test""" @property def __A ( self ): return F"""./tests/sagemaker/scripts/{self.framework}""" @property def __A ( self ): if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope="""class""" ) def UpperCamelCase (lowercase_: Tuple ) -> List[str]: A__ : str = SageMakerTestEnvironment(framework=request.cls.framework )
64
import os try: from .build_directory_md import good_file_paths except ImportError: from build_directory_md import good_file_paths # type: ignore A_ : Dict = list(good_file_paths()) assert filepaths, "good_file_paths() failed!" A_ : Optional[Any] = [file for file in filepaths if file != file.lower()] if upper_files: print(f'''{len(upper_files)} files contain uppercase characters:''') print('\n'.join(upper_files) + '\n') A_ : Tuple = [file for file in filepaths if ' ' in file] if space_files: print(f'''{len(space_files)} files contain space characters:''') print('\n'.join(space_files) + '\n') A_ : Any = [file for file in filepaths if '-' in file] if hyphen_files: print(f'''{len(hyphen_files)} files contain hyphen characters:''') print('\n'.join(hyphen_files) + '\n') A_ : List[str] = [file for file in filepaths if os.sep not in file] if nodir_files: print(f'''{len(nodir_files)} files are not in a directory:''') print('\n'.join(nodir_files) + '\n') A_ : Any = len(upper_files + space_files + hyphen_files + nodir_files) if bad_files: import sys sys.exit(bad_files)
64
1
from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, flip_channel_order, get_resize_output_image_size, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_torch_tensor, is_vision_available, logging if is_vision_available(): import PIL if is_torch_available(): import torch A_ : Tuple = logging.get_logger(__name__) class _a (__magic_name__ ): '''simple docstring''' UpperCAmelCase__: List[str] = ['''pixel_values'''] def __init__( self , A__ = True , A__ = None , A__ = PILImageResampling.BILINEAR , A__ = True , A__ = 1 / 255 , A__ = True , A__ = None , A__ = True , **A__ , ): super().__init__(**A__ ) A__ : Dict = size if size is not None else {"""shortest_edge""": 224} A__ : Dict = get_size_dict(A__ , default_to_square=A__ ) A__ : List[str] = crop_size if crop_size is not None else {"""height""": 256, """width""": 256} A__ : List[str] = get_size_dict(A__ , param_name="""crop_size""" ) A__ : str = do_resize A__ : str = size A__ : Optional[Any] = resample A__ : Optional[Any] = do_rescale A__ : str = rescale_factor A__ : Optional[Any] = do_center_crop A__ : Dict = crop_size A__ : Optional[Any] = do_flip_channel_order def __A ( self , A__ , A__ , A__ = PIL.Image.BILINEAR , A__ = None , **A__ , ): A__ : Any = get_size_dict(A__ , default_to_square=A__ ) if "shortest_edge" not in size: raise ValueError(F"""The `size` dictionary must contain the key `shortest_edge`. Got {size.keys()}""" ) A__ : Union[str, Any] = get_resize_output_image_size(A__ , size=size["""shortest_edge"""] , default_to_square=A__ ) return resize(A__ , size=A__ , resample=A__ , data_format=A__ , **A__ ) def __A ( self , A__ , A__ , A__ = None , **A__ , ): A__ : Optional[Any] = get_size_dict(A__ ) if "height" not in size or "width" not in size: raise ValueError(F"""The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}""" ) return center_crop(A__ , size=(size["""height"""], size["""width"""]) , data_format=A__ , **A__ ) def __A ( self , A__ , A__ , A__ = None , **A__ , ): return rescale(A__ , scale=A__ , data_format=A__ , **A__ ) def __A ( self , A__ , A__ = None ): return flip_channel_order(A__ , data_format=A__ ) def __A ( self , A__ , A__ = None , A__ = None , A__ = None , A__ = None , A__ = None , A__ = None , A__ = None , A__ = None , A__ = None , A__ = ChannelDimension.FIRST , **A__ , ): A__ : List[str] = do_resize if do_resize is not None else self.do_resize A__ : Union[str, Any] = resample if resample is not None else self.resample A__ : Optional[Any] = do_rescale if do_rescale is not None else self.do_rescale A__ : int = rescale_factor if rescale_factor is not None else self.rescale_factor A__ : int = do_center_crop if do_center_crop is not None else self.do_center_crop A__ : Union[str, Any] = ( do_flip_channel_order if do_flip_channel_order is not None else self.do_flip_channel_order ) A__ : Union[str, Any] = size if size is not None else self.size A__ : Any = get_size_dict(A__ , default_to_square=A__ ) A__ : Optional[Any] = crop_size if crop_size is not None else self.crop_size A__ : List[Any] = get_size_dict(A__ , param_name="""crop_size""" ) A__ : Dict = make_list_of_images(A__ ) if not valid_images(A__ ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None: raise ValueError("""Size must be specified if do_resize is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) # All transformations expect numpy arrays. A__ : str = [to_numpy_array(A__ ) for image in images] if do_resize: A__ : Optional[Any] = [self.resize(image=A__ , size=A__ , resample=A__ ) for image in images] if do_center_crop: A__ : Dict = [self.center_crop(image=A__ , size=A__ ) for image in images] if do_rescale: A__ : int = [self.rescale(image=A__ , scale=A__ ) for image in images] # the pretrained checkpoints assume images are BGR, not RGB if do_flip_channel_order: A__ : List[Any] = [self.flip_channel_order(image=A__ ) for image in images] A__ : Any = [to_channel_dimension_format(A__ , A__ ) for image in images] A__ : Any = {"""pixel_values""": images} return BatchFeature(data=A__ , tensor_type=A__ ) def __A ( self , A__ , A__ = None ): A__ : Union[str, Any] = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(A__ ) != len(A__ ): raise ValueError( """Make sure that you pass in as many target sizes as the batch dimension of the logits""" ) if is_torch_tensor(A__ ): A__ : Tuple = target_sizes.numpy() A__ : List[Any] = [] for idx in range(len(A__ ) ): A__ : Optional[Any] = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode="""bilinear""" , align_corners=A__ ) A__ : List[Any] = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(A__ ) else: A__ : int = logits.argmax(dim=1 ) A__ : Any = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
64
import inspect import warnings from typing import Any, Dict, Optional, Union from packaging import version def UpperCamelCase (*lowercase_: Optional[int] , lowercase_: Optional[Union[Dict, Any]] = None , lowercase_: Dict=True , lowercase_: Tuple=2 ) -> Dict: from .. import __version__ A__ : Dict = take_from A__ : str = () if not isinstance(args[0] , lowercase_ ): A__ : int = (args,) for attribute, version_name, message in args: if version.parse(version.parse(lowercase_ ).base_version ) >= version.parse(lowercase_ ): raise ValueError( f"""The deprecation tuple {(attribute, version_name, message)} should be removed since diffusers'""" f""" version {__version__} is >= {version_name}""" ) A__ : Any = None if isinstance(lowercase_ , lowercase_ ) and attribute in deprecated_kwargs: values += (deprecated_kwargs.pop(lowercase_ ),) A__ : List[str] = f"""The `{attribute}` argument is deprecated and will be removed in version {version_name}.""" elif hasattr(lowercase_ , lowercase_ ): values += (getattr(lowercase_ , lowercase_ ),) A__ : Optional[Any] = f"""The `{attribute}` attribute is deprecated and will be removed in version {version_name}.""" elif deprecated_kwargs is None: A__ : int = f"""`{attribute}` is deprecated and will be removed in version {version_name}.""" if warning is not None: A__ : int = warning + """ """ if standard_warn else """""" warnings.warn(warning + message , lowercase_ , stacklevel=lowercase_ ) if isinstance(lowercase_ , lowercase_ ) and len(lowercase_ ) > 0: A__ : Union[str, Any] = inspect.getouterframes(inspect.currentframe() )[1] A__ : Optional[Any] = call_frame.filename A__ : Optional[int] = call_frame.lineno A__ : Any = call_frame.function A__ , A__ : List[str] = next(iter(deprecated_kwargs.items() ) ) raise TypeError(f"""{function} in {filename} line {line_number-1} got an unexpected keyword argument `{key}`""" ) if len(lowercase_ ) == 0: return elif len(lowercase_ ) == 1: return values[0] return values
64
1
import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class _a (__magic_name__ ): '''simple docstring''' @require_torch def __A ( self ): # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched A__ : str = """ from transformers import BertConfig, BertModel, BertTokenizer, pipeline """ A__ : Tuple = """ mname = \"hf-internal-testing/tiny-random-bert\" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) BertTokenizer.from_pretrained(mname) pipe = pipeline(task=\"fill-mask\", model=mname) print(\"success\") """ A__ : List[str] = """ import socket def offline_socket(*args, **kwargs): raise RuntimeError(\"Offline mode is enabled, we shouldn't access internet\") socket.socket = offline_socket """ # Force fetching the files so that we can use the cache A__ : Optional[Any] = """hf-internal-testing/tiny-random-bert""" BertConfig.from_pretrained(A__ ) BertModel.from_pretrained(A__ ) BertTokenizer.from_pretrained(A__ ) pipeline(task="""fill-mask""" , model=A__ ) # baseline - just load from_pretrained with normal network A__ : Any = [sys.executable, """-c""", """\n""".join([load, run, mock] )] # should succeed A__ : Dict = self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files A__ : Dict = """1""" A__ : int = subprocess.run(A__ , env=A__ , check=A__ , capture_output=A__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() ) @require_torch def __A ( self ): # python one-liner segments # this must be loaded before socket.socket is monkey-patched A__ : Optional[Any] = """ from transformers import BertConfig, BertModel, BertTokenizer, pipeline """ A__ : Optional[Any] = """ mname = \"hf-internal-testing/tiny-random-bert\" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) BertTokenizer.from_pretrained(mname) pipe = pipeline(task=\"fill-mask\", model=mname) print(\"success\") """ A__ : int = """ import socket def offline_socket(*args, **kwargs): raise socket.error(\"Faking flaky internet\") socket.socket = offline_socket """ # Force fetching the files so that we can use the cache A__ : int = """hf-internal-testing/tiny-random-bert""" BertConfig.from_pretrained(A__ ) BertModel.from_pretrained(A__ ) BertTokenizer.from_pretrained(A__ ) pipeline(task="""fill-mask""" , model=A__ ) # baseline - just load from_pretrained with normal network A__ : List[str] = [sys.executable, """-c""", """\n""".join([load, run, mock] )] # should succeed A__ : Dict = self.get_env() A__ : Dict = subprocess.run(A__ , env=A__ , check=A__ , capture_output=A__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() ) @require_torch def __A ( self ): # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched A__ : Any = """ from transformers import BertConfig, BertModel, BertTokenizer """ A__ : str = """ mname = \"hf-internal-testing/tiny-random-bert-sharded\" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) print(\"success\") """ A__ : str = """ import socket def offline_socket(*args, **kwargs): raise ValueError(\"Offline mode is enabled\") socket.socket = offline_socket """ # baseline - just load from_pretrained with normal network A__ : Union[str, Any] = [sys.executable, """-c""", """\n""".join([load, run] )] # should succeed A__ : List[Any] = self.get_env() A__ : str = subprocess.run(A__ , env=A__ , check=A__ , capture_output=A__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() ) # next emulate no network A__ : List[Any] = [sys.executable, """-c""", """\n""".join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files A__ : List[str] = """1""" A__ : Tuple = subprocess.run(A__ , env=A__ , check=A__ , capture_output=A__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() ) @require_torch def __A ( self ): A__ : Dict = """ from transformers import pipeline """ A__ : List[Any] = """ mname = \"hf-internal-testing/tiny-random-bert\" pipe = pipeline(model=mname) """ A__ : List[Any] = """ import socket def offline_socket(*args, **kwargs): raise socket.error(\"Offline mode is enabled\") socket.socket = offline_socket """ A__ : Any = self.get_env() A__ : Optional[Any] = """1""" A__ : int = [sys.executable, """-c""", """\n""".join([load, mock, run] )] A__ : List[Any] = subprocess.run(A__ , env=A__ , check=A__ , capture_output=A__ ) self.assertEqual(result.returncode , 1 , result.stderr ) self.assertIn( """You cannot infer task automatically within `pipeline` when using offline mode""" , result.stderr.decode().replace("""\n""" , """""" ) , ) @require_torch def __A ( self ): A__ : Optional[int] = """ from transformers import AutoModel """ A__ : Any = """ mname = \"hf-internal-testing/test_dynamic_model\" AutoModel.from_pretrained(mname, trust_remote_code=True) print(\"success\") """ # baseline - just load from_pretrained with normal network A__ : List[Any] = [sys.executable, """-c""", """\n""".join([load, run] )] # should succeed A__ : Dict = self.get_env() A__ : int = subprocess.run(A__ , env=A__ , check=A__ , capture_output=A__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files A__ : Tuple = """1""" A__ : Optional[Any] = subprocess.run(A__ , env=A__ , check=A__ , capture_output=A__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() )
64
import argparse import json import os import tensorstore as ts import torch from flax import serialization from flax.traverse_util import flatten_dict, unflatten_dict from tensorflow.io import gfile from transformers.modeling_utils import dtype_byte_size from transformers.models.switch_transformers.convert_switch_transformers_original_flax_checkpoint_to_pytorch import ( rename_keys, ) from transformers.utils import WEIGHTS_INDEX_NAME, WEIGHTS_NAME from transformers.utils.hub import convert_file_size_to_int def UpperCamelCase (lowercase_: List[str] , lowercase_: str ) -> Optional[Any]: if flax_key_tuple[-1] == "kernel" and flax_tensor.ndim == 3: # expert layer A__ : Union[str, Any] = flax_key_tuple[:-1] + ("""weight""",) A__ : Optional[int] = torch.permute(lowercase_ , (0, 2, 1) ) elif flax_key_tuple[-1] == "kernel" and ".".join(lowercase_ ): # linear layer A__ : Optional[Any] = flax_key_tuple[:-1] + ("""weight""",) A__ : int = flax_tensor.T elif flax_key_tuple[-1] in ["scale", "embedding"]: A__ : Optional[int] = flax_key_tuple[:-1] + ("""weight""",) return flax_key_tuple, flax_tensor def UpperCamelCase (lowercase_: Tuple , lowercase_: Optional[int] , lowercase_: str ) -> Union[str, Any]: if "metadata" in layer: A__ : Tuple = layer.split("""metadata""" ) A__ : Optional[Any] = """""".join(split_layer[0] )[:-1] A__ : Optional[Any] = [tuple(("""metadata""" + split_layer[1]).split("""/""" ) )] elif "kvstore" in layer: A__ : str = layer.split("""kvstore""" ) A__ : int = """""".join(split_layer[0] )[:-1] A__ : Optional[int] = [tuple(("""kvstore""" + split_layer[1]).split("""/""" ) )] else: A__ : Any = layer.split("""/""" ) A__ : int = """/""".join(split_layer[:-1] ) A__ : str = (split_layer[-1],) if "kvstore/path" in layer: A__ : Dict = f"""{switch_checkpoint_path}/{checkpoint_info[layer]}""" elif "kvstore/driver" in layer: A__ : Optional[int] = """file""" else: A__ : str = checkpoint_info[layer] return curr_real_layer_name, split_layer, content def UpperCamelCase (lowercase_: str , lowercase_: List[Any] ) -> int: A__ : int = rename_keys(lowercase_ ) A__ : Any = {} for k, v in current_block.items(): A__ : Dict = v A__ : str = new_current_block torch.save(lowercase_ , lowercase_ ) def UpperCamelCase (lowercase_: Dict , lowercase_: Optional[Any] , lowercase_: Optional[Any] , lowercase_: Optional[int] , lowercase_: str = WEIGHTS_NAME ) -> Tuple: A__ : Optional[int] = convert_file_size_to_int(lowercase_ ) A__ : List[Any] = [] A__ : int = {} A__ : List[str] = 0 A__ : Any = 0 os.makedirs(lowercase_ , exist_ok=lowercase_ ) with gfile.GFile(switch_checkpoint_path + """/checkpoint""" , """rb""" ) as fp: A__ : Optional[Any] = serialization.msgpack_restore(fp.read() )["""optimizer"""]["""target"""] A__ : Dict = flatten_dict(lowercase_ , sep="""/""" ) A__ : Any = {} for layer in checkpoint_info.keys(): A__ , A__ , A__ : Union[str, Any] = get_key_and_tensorstore_dict( lowercase_ , lowercase_ , lowercase_ ) if curr_real_layer_name in all_layers: A__ : Optional[int] = content else: A__ : List[Any] = {split_layer[-1]: content} for key in all_layers.keys(): # open tensorstore file A__ : Optional[Any] = ts.open(unflatten_dict(all_layers[key] ) ).result().read().result() A__ : List[Any] = torch.tensor(lowercase_ ) A__ : List[Any] = raw_weights.numel() * dtype_byte_size(raw_weights.dtype ) # use the renaming pattern from the small conversion scripts A__ , A__ : Any = rename_base_flax_keys(tuple(key.split("""/""" ) ) , lowercase_ ) A__ : Any = """/""".join(lowercase_ ) # If this weight is going to tip up over the maximal size, we split. if current_block_size + weight_size > max_shard_size: A__ : List[Any] = os.path.join( lowercase_ , weights_name.replace(""".bin""" , f"""-{len(lowercase_ )+1:05d}-of-???.bin""" ) ) rename_and_save_block(lowercase_ , lowercase_ ) sharded_state_dicts.append(current_block.keys() ) del current_block A__ : Any = {} A__ : str = 0 A__ : List[str] = raw_weights.to(getattr(lowercase_ , lowercase_ ) ) current_block_size += weight_size total_size += weight_size # Add the last block A__ : Union[str, Any] = os.path.join(lowercase_ , weights_name.replace(""".bin""" , f"""-{len(lowercase_ )+1:05d}-of-???.bin""" ) ) rename_and_save_block(lowercase_ , lowercase_ ) sharded_state_dicts.append(current_block.keys() ) # If we only have one shard, we return it if len(lowercase_ ) == 1: return {weights_name: sharded_state_dicts[0]}, None # Otherwise, let's build the index A__ : str = {} A__ : Any = {} for idx, shard in enumerate(lowercase_ ): A__ : Any = weights_name.replace( """.bin""" , f"""-{idx+1:05d}-of-{len(lowercase_ ):05d}.bin""" ) # len(sharded_state_dicts):05d} A__ : Dict = os.path.join(lowercase_ , weights_name.replace(""".bin""" , f"""-{idx+1:05d}-of-???.bin""" ) ) os.rename(lowercase_ , os.path.join(lowercase_ , lowercase_ ) ) A__ : str = shard for key in shard: A__ : Any = shard_file # Add the metadata A__ : Tuple = {"""total_size""": total_size} A__ : Union[str, Any] = {"""metadata""": metadata, """weight_map""": weight_map} with open(os.path.join(lowercase_ , lowercase_ ) , """w""" , encoding="""utf-8""" ) as f: A__ : Dict = json.dumps(lowercase_ , indent=2 , sort_keys=lowercase_ ) + """\n""" f.write(lowercase_ ) return metadata, index if __name__ == "__main__": A_ : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--switch_t5x_checkpoint_path', default='/mnt/disks/disk_switch/original_checkpoints/switch-xxl-128/checkpoint_634600', type=str, required=False, help='Path to a directory containing a folder per layer. Follows the original Google format.', ) parser.add_argument('--max_shard_size', default='10GB', required=False, help='Max shard size') parser.add_argument('--dtype', default='bfloat16', type=str, required=False, help='dtype of the saved model') parser.add_argument( '--pytorch_dump_folder_path', default='/mnt/disks/disk_switch/original_checkpoints/switch-xxl-128-converted', type=str, required=False, help='Path to the output pytorch model.', ) A_ : Dict = parser.parse_args() shard_on_the_fly( args.switch_tax_checkpoint_path, args.pytorch_dump_folder_path, args.max_shard_size, args.dtype, ) def UpperCamelCase () -> int: from transformers import SwitchTransformersConfig, SwitchTransformersForConditionalGeneration, TaTokenizer A__ : str = SwitchTransformersConfig.from_pretrained("""google/switch-base-8""" ) config.save_pretrained("""/home/arthur_huggingface_co/transformers/switch_converted""" ) A__ : str = SwitchTransformersForConditionalGeneration.from_pretrained( """/home/arthur_huggingface_co/transformers/switch_converted""" , device_map="""auto""" ) A__ : Tuple = TaTokenizer.from_pretrained("""t5-small""" ) A__ : Dict = """A <extra_id_0> walks into a bar a orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.""" A__ : Union[str, Any] = tokenizer(lowercase_ , return_tensors="""pt""" ).input_ids A__ : Tuple = model.generate(lowercase_ , decoder_start_token_id=0 ) print(tokenizer.decode(out[0] ) )
64
1
from __future__ import annotations from typing import Generic, TypeVar A_ : Optional[Any] = TypeVar('T') class _a (Generic[T] ): '''simple docstring''' def __init__( self , A__ ): A__ : int = data A__ : Optional[int] = self A__ : int = 0 class _a (Generic[T] ): '''simple docstring''' def __init__( self ): # map from node name to the node object A__ : dict[T, DisjointSetTreeNode[T]] = {} def __A ( self , A__ ): # create a new set with x as its member A__ : List[str] = DisjointSetTreeNode(A__ ) def __A ( self , A__ ): # find the set x belongs to (with path-compression) A__ : Any = self.map[data] if elem_ref != elem_ref.parent: A__ : str = self.find_set(elem_ref.parent.data ) return elem_ref.parent def __A ( self , A__ , A__ ): # helper function for union operation if nodea.rank > nodea.rank: A__ : Tuple = nodea else: A__ : str = nodea if nodea.rank == nodea.rank: nodea.rank += 1 def __A ( self , A__ , A__ ): # merge 2 disjoint sets self.link(self.find_set(A__ ) , self.find_set(A__ ) ) class _a (Generic[T] ): '''simple docstring''' def __init__( self ): # connections: map from the node to the neighbouring nodes (with weights) A__ : dict[T, dict[T, int]] = {} def __A ( self , A__ ): # add a node ONLY if its not present in the graph if node not in self.connections: A__ : Dict = {} def __A ( self , A__ , A__ , A__ ): # add an edge with the given weight self.add_node(A__ ) self.add_node(A__ ) A__ : int = weight A__ : int = weight def __A ( self ): A__ : Any = [] A__ : str = set() for start in self.connections: for end in self.connections[start]: if (start, end) not in seen: seen.add((end, start) ) edges.append((start, end, self.connections[start][end]) ) edges.sort(key=lambda A__ : x[2] ) # creating the disjoint set A__ : List[Any] = DisjointSetTree[T]() for node in self.connections: disjoint_set.make_set(A__ ) # MST generation A__ : Union[str, Any] = 0 A__ : int = 0 A__ : Dict = GraphUndirectedWeighted[T]() while num_edges < len(self.connections ) - 1: A__ , A__ , A__ : Union[str, Any] = edges[index] index += 1 A__ : Tuple = disjoint_set.find_set(A__ ) A__ : Union[str, Any] = disjoint_set.find_set(A__ ) if parent_u != parent_v: num_edges += 1 graph.add_edge(A__ , A__ , A__ ) disjoint_set.union(A__ , A__ ) return graph
64
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available A_ : Optional[Any] = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : str = ['BartphoTokenizer'] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bartpho import BartphoTokenizer else: import sys A_ : Optional[int] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
64
1
from ...configuration_utils import PretrainedConfig from ...utils import logging A_ : int = logging.get_logger(__name__) A_ : Optional[Any] = { 'facebook/vit-mae-base': 'https://huggingface.co/facebook/vit-mae-base/resolve/main/config.json', # See all ViT MAE models at https://huggingface.co/models?filter=vit-mae } class _a (__magic_name__ ): '''simple docstring''' UpperCAmelCase__: List[Any] = '''vit_mae''' def __init__( self , A__=768 , A__=12 , A__=12 , A__=3072 , A__="gelu" , A__=0.0 , A__=0.0 , A__=0.0_2 , A__=1e-12 , A__=224 , A__=16 , A__=3 , A__=True , A__=16 , A__=512 , A__=8 , A__=2048 , A__=0.7_5 , A__=False , **A__ , ): super().__init__(**A__ ) A__ : Dict = hidden_size A__ : Optional[Any] = num_hidden_layers A__ : List[str] = num_attention_heads A__ : Any = intermediate_size A__ : Union[str, Any] = hidden_act A__ : List[str] = hidden_dropout_prob A__ : Union[str, Any] = attention_probs_dropout_prob A__ : Optional[int] = initializer_range A__ : Optional[Any] = layer_norm_eps A__ : Tuple = image_size A__ : Dict = patch_size A__ : Optional[int] = num_channels A__ : List[str] = qkv_bias A__ : Tuple = decoder_num_attention_heads A__ : Any = decoder_hidden_size A__ : Optional[int] = decoder_num_hidden_layers A__ : Tuple = decoder_intermediate_size A__ : int = mask_ratio A__ : Union[str, Any] = norm_pix_loss
64
import argparse import hashlib import os import urllib import warnings import torch from torch import nn from tqdm import tqdm from transformers import WhisperConfig, WhisperForConditionalGeneration A_ : Dict = { 'tiny.en': 'https://openaipublic.azureedge.net/main/whisper/models/d3dd57d32accea0b295c96e26691aa14d8822fac7d9d27d5dc00b4ca2826dd03/tiny.en.pt', 'tiny': 'https://openaipublic.azureedge.net/main/whisper/models/65147644a518d12f04e32d6f3b26facc3f8dd46e5390956a9424a650c0ce22b9/tiny.pt', 'base.en': 'https://openaipublic.azureedge.net/main/whisper/models/25a8566e1d0c1e2231d1c762132cd20e0f96a85d16145c3a00adf5d1ac670ead/base.en.pt', 'base': 'https://openaipublic.azureedge.net/main/whisper/models/ed3a0b6b1c0edf879ad9b11b1af5a0e6ab5db9205f891f668f8b0e6c6326e34e/base.pt', 'small.en': 'https://openaipublic.azureedge.net/main/whisper/models/f953ad0fd29cacd07d5a9eda5624af0f6bcf2258be67c92b79389873d91e0872/small.en.pt', 'small': 'https://openaipublic.azureedge.net/main/whisper/models/9ecf779972d90ba49c06d968637d720dd632c55bbf19d441fb42bf17a411e794/small.pt', 'medium.en': 'https://openaipublic.azureedge.net/main/whisper/models/d7440d1dc186f76616474e0ff0b3b6b879abc9d1a4926b7adfa41db2d497ab4f/medium.en.pt', 'medium': 'https://openaipublic.azureedge.net/main/whisper/models/345ae4da62f9b3d59415adc60127b97c714f32e89e936602e85993674d08dcb1/medium.pt', 'large': 'https://openaipublic.azureedge.net/main/whisper/models/e4b87e7e0bf463eb8e6956e646f1e277e901512310def2c24bf0e11bd3c28e9a/large.pt', 'large-v2': 'https://openaipublic.azureedge.net/main/whisper/models/81f7c96c852ee8fc832187b0132e569d6c3065a3252ed18e56effd0b6a73e524/large-v2.pt', } def UpperCamelCase (lowercase_: Optional[Any] ) -> Optional[int]: A__ : List[Any] = ["""layers""", """blocks"""] for k in ignore_keys: state_dict.pop(lowercase_ , lowercase_ ) A_ : Any = { 'blocks': 'layers', 'mlp.0': 'fc1', 'mlp.2': 'fc2', 'mlp_ln': 'final_layer_norm', '.attn.query': '.self_attn.q_proj', '.attn.key': '.self_attn.k_proj', '.attn.value': '.self_attn.v_proj', '.attn_ln': '.self_attn_layer_norm', '.attn.out': '.self_attn.out_proj', '.cross_attn.query': '.encoder_attn.q_proj', '.cross_attn.key': '.encoder_attn.k_proj', '.cross_attn.value': '.encoder_attn.v_proj', '.cross_attn_ln': '.encoder_attn_layer_norm', '.cross_attn.out': '.encoder_attn.out_proj', 'decoder.ln.': 'decoder.layer_norm.', 'encoder.ln.': 'encoder.layer_norm.', 'token_embedding': 'embed_tokens', 'encoder.positional_embedding': 'encoder.embed_positions.weight', 'decoder.positional_embedding': 'decoder.embed_positions.weight', 'ln_post': 'layer_norm', } def UpperCamelCase (lowercase_: str ) -> Any: A__ : Dict = list(s_dict.keys() ) for key in keys: A__ : List[str] = key for k, v in WHISPER_MAPPING.items(): if k in key: A__ : List[Any] = new_key.replace(lowercase_ , lowercase_ ) print(f"""{key} -> {new_key}""" ) A__ : Tuple = s_dict.pop(lowercase_ ) return s_dict def UpperCamelCase (lowercase_: Tuple ) -> Optional[int]: A__ , A__ : Any = emb.weight.shape A__ : str = nn.Linear(lowercase_ , lowercase_ , bias=lowercase_ ) A__ : Union[str, Any] = emb.weight.data return lin_layer def UpperCamelCase (lowercase_: str , lowercase_: str ) -> bytes: os.makedirs(lowercase_ , exist_ok=lowercase_ ) A__ : Tuple = os.path.basename(lowercase_ ) A__ : int = url.split("""/""" )[-2] A__ : Dict = os.path.join(lowercase_ , lowercase_ ) if os.path.exists(lowercase_ ) and not os.path.isfile(lowercase_ ): raise RuntimeError(f"""{download_target} exists and is not a regular file""" ) if os.path.isfile(lowercase_ ): A__ : Optional[Any] = open(lowercase_ , """rb""" ).read() if hashlib.shaaaa(lowercase_ ).hexdigest() == expected_shaaaa: return model_bytes else: warnings.warn(f"""{download_target} exists, but the SHA256 checksum does not match; re-downloading the file""" ) with urllib.request.urlopen(lowercase_ ) as source, open(lowercase_ , """wb""" ) as output: with tqdm( total=int(source.info().get("""Content-Length""" ) ) , ncols=80 , unit="""iB""" , unit_scale=lowercase_ , unit_divisor=1024 ) as loop: while True: A__ : Any = source.read(8192 ) if not buffer: break output.write(lowercase_ ) loop.update(len(lowercase_ ) ) A__ : Dict = open(lowercase_ , """rb""" ).read() if hashlib.shaaaa(lowercase_ ).hexdigest() != expected_shaaaa: raise RuntimeError( """Model has been downloaded but the SHA256 checksum does not not match. Please retry loading the model.""" ) return model_bytes def UpperCamelCase (lowercase_: List[Any] , lowercase_: Tuple ) -> Optional[Any]: if ".pt" not in checkpoint_path: A__ : Tuple = _download(_MODELS[checkpoint_path] ) else: A__ : Optional[int] = torch.load(lowercase_ , map_location="""cpu""" ) A__ : str = original_checkpoint["""dims"""] A__ : List[Any] = original_checkpoint["""model_state_dict"""] A__ : Optional[Any] = state_dict["""decoder.token_embedding.weight"""] remove_ignore_keys_(lowercase_ ) rename_keys(lowercase_ ) A__ : List[str] = True A__ : Optional[Any] = state_dict["""decoder.layers.0.fc1.weight"""].shape[0] A__ : List[Any] = WhisperConfig( vocab_size=dimensions["""n_vocab"""] , encoder_ffn_dim=lowercase_ , decoder_ffn_dim=lowercase_ , num_mel_bins=dimensions["""n_mels"""] , d_model=dimensions["""n_audio_state"""] , max_target_positions=dimensions["""n_text_ctx"""] , encoder_layers=dimensions["""n_audio_layer"""] , encoder_attention_heads=dimensions["""n_audio_head"""] , decoder_layers=dimensions["""n_text_layer"""] , decoder_attention_heads=dimensions["""n_text_state"""] , max_source_positions=dimensions["""n_audio_ctx"""] , ) A__ : Optional[Any] = WhisperForConditionalGeneration(lowercase_ ) A__ , A__ : List[Any] = model.model.load_state_dict(lowercase_ , strict=lowercase_ ) if len(lowercase_ ) > 0 and not set(lowercase_ ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( """Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing,""" f""" but all the following weights are missing {missing}""" ) if tie_embeds: A__ : Any = make_linear_from_emb(model.model.decoder.embed_tokens ) else: A__ : str = proj_out_weights model.save_pretrained(lowercase_ ) if __name__ == "__main__": A_ : Any = argparse.ArgumentParser() # # Required parameters parser.add_argument('--checkpoint_path', type=str, help='Patht to the downloaded checkpoints') parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') A_ : Tuple = parser.parse_args() convert_openai_whisper_to_tfms(args.checkpoint_path, args.pytorch_dump_folder_path)
64
1
from __future__ import annotations import inspect import unittest from math import floor import numpy as np from transformers import CvtConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFCvtForImageClassification, TFCvtModel from transformers.models.cvt.modeling_tf_cvt import TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _a (__magic_name__ ): '''simple docstring''' def __A ( self ): A__ : Any = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(A__ , """embed_dim""" ) ) self.parent.assertTrue(hasattr(A__ , """num_heads""" ) ) class _a : '''simple docstring''' def __init__( self , A__ , A__=13 , A__=64 , A__=3 , A__=[16, 48, 96] , A__=[1, 3, 6] , A__=[1, 2, 10] , A__=[7, 3, 3] , A__=[4, 2, 2] , A__=[2, 1, 1] , A__=[2, 2, 2] , A__=[False, False, True] , A__=[0.0, 0.0, 0.0] , A__=0.0_2 , A__=1e-12 , A__=True , A__=True , A__=2 , ): A__ : List[Any] = parent A__ : str = batch_size A__ : Dict = image_size A__ : int = patch_sizes A__ : Optional[int] = patch_stride A__ : int = patch_padding A__ : str = is_training A__ : Dict = use_labels A__ : str = num_labels A__ : Any = num_channels A__ : Dict = embed_dim A__ : int = num_heads A__ : Tuple = stride_kv A__ : List[Any] = depth A__ : Dict = cls_token A__ : Optional[Any] = attention_drop_rate A__ : Tuple = initializer_range A__ : List[str] = layer_norm_eps def __A ( self ): A__ : Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A__ : List[Any] = None if self.use_labels: # create a random int32 tensor of given shape A__ : List[Any] = ids_tensor([self.batch_size] , self.num_labels ) A__ : Tuple = self.get_config() return config, pixel_values, labels def __A ( self ): return CvtConfig( image_size=self.image_size , num_labels=self.num_labels , num_channels=self.num_channels , embed_dim=self.embed_dim , num_heads=self.num_heads , patch_sizes=self.patch_sizes , patch_padding=self.patch_padding , patch_stride=self.patch_stride , stride_kv=self.stride_kv , depth=self.depth , cls_token=self.cls_token , attention_drop_rate=self.attention_drop_rate , initializer_range=self.initializer_range , ) def __A ( self , A__ , A__ , A__ ): A__ : Tuple = TFCvtModel(config=A__ ) A__ : Tuple = model(A__ , training=A__ ) A__ : Dict = (self.image_size, self.image_size) A__ , A__ : int = image_size[0], image_size[1] for i in range(len(self.depth ) ): A__ : List[Any] = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) A__ : Any = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dim[-1], height, width) ) def __A ( self , A__ , A__ , A__ ): A__ : Optional[Any] = self.num_labels A__ : Optional[Any] = TFCvtForImageClassification(A__ ) A__ : str = model(A__ , labels=A__ , training=A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __A ( self ): A__ : Union[str, Any] = self.prepare_config_and_inputs() A__ , A__ , A__ : Optional[int] = config_and_inputs A__ : Union[str, Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class _a (__magic_name__ , __magic_name__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__: str = (TFCvtModel, TFCvtForImageClassification) if is_tf_available() else () UpperCAmelCase__: List[Any] = ( {'''feature-extraction''': TFCvtModel, '''image-classification''': TFCvtForImageClassification} if is_tf_available() else {} ) UpperCAmelCase__: str = False UpperCAmelCase__: Any = False UpperCAmelCase__: Union[str, Any] = False UpperCAmelCase__: int = False UpperCAmelCase__: Tuple = False def __A ( self ): A__ : List[str] = TFCvtModelTester(self ) A__ : Any = TFCvtConfigTester(self , config_class=A__ , has_text_modality=A__ , hidden_size=37 ) def __A ( self ): self.config_tester.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() @unittest.skip(reason="""Cvt does not output attentions""" ) def __A ( self ): pass @unittest.skip(reason="""Cvt does not use inputs_embeds""" ) def __A ( self ): pass @unittest.skip(reason="""Cvt does not support input and output embeddings""" ) def __A ( self ): pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("""GPU""" ) ) == 0 , reason="""TF does not support backprop for grouped convolutions on CPU.""" , ) def __A ( self ): super().test_dataset_conversion() @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("""GPU""" ) ) == 0 , reason="""TF does not support backprop for grouped convolutions on CPU.""" , ) @slow def __A ( self ): super().test_keras_fit() @unittest.skip(reason="""Get `Failed to determine best cudnn convolution algo.` error after using TF 2.12+cuda 11.8""" ) def __A ( self ): A__ : Any = tf.keras.mixed_precision.Policy("""mixed_float16""" ) tf.keras.mixed_precision.set_global_policy(A__ ) super().test_keras_fit() tf.keras.mixed_precision.set_global_policy("""float32""" ) def __A ( self ): A__ , A__ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ : Any = model_class(A__ ) A__ : Optional[Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A__ : Dict = [*signature.parameters.keys()] A__ : Tuple = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , A__ ) def __A ( self ): def check_hidden_states_output(A__ , A__ , A__ ): A__ : Optional[int] = model_class(A__ ) A__ : Optional[Any] = model(**self._prepare_for_class(A__ , A__ ) ) A__ : Dict = outputs.hidden_states A__ : Any = len(self.model_tester.depth ) self.assertEqual(len(A__ ) , A__ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) A__ , A__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ : Optional[Any] = True check_hidden_states_output(A__ , A__ , A__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A__ : str = True check_hidden_states_output(A__ , A__ , A__ ) def __A ( self ): A__ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A__ ) def __A ( self ): A__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A__ ) @slow def __A ( self ): for model_name in TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ : List[Any] = TFCvtModel.from_pretrained(A__ ) self.assertIsNotNone(A__ ) def UpperCamelCase () -> Union[str, Any]: A__ : Dict = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class _a (unittest.TestCase ): '''simple docstring''' @cached_property def __A ( self ): return AutoImageProcessor.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def __A ( self ): A__ : Any = TFCvtForImageClassification.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) A__ : str = self.default_image_processor A__ : Union[str, Any] = prepare_img() A__ : Tuple = image_processor(images=A__ , return_tensors="""tf""" ) # forward pass A__ : Tuple = model(**A__ ) # verify the logits A__ : Optional[Any] = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , A__ ) A__ : Any = tf.constant([0.9_2_8_5, 0.9_0_1_5, -0.3_1_5_0] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , A__ , atol=1e-4 ) )
64
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class _a (__magic_name__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__: Any = TextToVideoSDPipeline UpperCAmelCase__: Any = TEXT_TO_IMAGE_PARAMS UpperCAmelCase__: Optional[Any] = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. UpperCAmelCase__: Optional[int] = frozenset( [ '''num_inference_steps''', '''generator''', '''latents''', '''return_dict''', '''callback''', '''callback_steps''', ] ) def __A ( self ): torch.manual_seed(0 ) A__ : Optional[int] = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """DownBlock3D""") , up_block_types=("""UpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""") , cross_attention_dim=32 , attention_head_dim=4 , ) A__ : Optional[int] = DDIMScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule="""scaled_linear""" , clip_sample=A__ , set_alpha_to_one=A__ , ) torch.manual_seed(0 ) A__ : Union[str, Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) A__ : Union[str, Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act="""gelu""" , projection_dim=512 , ) A__ : Union[str, Any] = CLIPTextModel(A__ ) A__ : Tuple = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) A__ : Dict = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, } return components def __A ( self , A__ , A__=0 ): if str(A__ ).startswith("""mps""" ): A__ : Tuple = torch.manual_seed(A__ ) else: A__ : List[str] = torch.Generator(device=A__ ).manual_seed(A__ ) A__ : List[str] = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """pt""", } return inputs def __A ( self ): A__ : List[str] = """cpu""" # ensure determinism for the device-dependent torch.Generator A__ : Union[str, Any] = self.get_dummy_components() A__ : Union[str, Any] = TextToVideoSDPipeline(**A__ ) A__ : int = sd_pipe.to(A__ ) sd_pipe.set_progress_bar_config(disable=A__ ) A__ : int = self.get_dummy_inputs(A__ ) A__ : int = """np""" A__ : Any = sd_pipe(**A__ ).frames A__ : Dict = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) A__ : Optional[Any] = np.array([1_5_8.0, 1_6_0.0, 1_5_3.0, 1_2_5.0, 1_0_0.0, 1_2_1.0, 1_1_1.0, 9_3.0, 1_1_3.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def __A ( self ): self._test_attention_slicing_forward_pass(test_mean_pixel_difference=A__ , expected_max_diff=3e-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def __A ( self ): self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=A__ , expected_max_diff=1e-2 ) @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def __A ( self ): pass @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def __A ( self ): pass @unittest.skip(reason="""`num_images_per_prompt` argument is not supported for this pipeline.""" ) def __A ( self ): pass def __A ( self ): return super().test_progress_bar() @slow @skip_mps class _a (unittest.TestCase ): '''simple docstring''' def __A ( self ): A__ : Union[str, Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy""" ) A__ : Tuple = TextToVideoSDPipeline.from_pretrained("""damo-vilab/text-to-video-ms-1.7b""" ) A__ : Any = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) A__ : int = pipe.to("""cuda""" ) A__ : Optional[Any] = """Spiderman is surfing""" A__ : List[str] = torch.Generator(device="""cpu""" ).manual_seed(0 ) A__ : Optional[Any] = pipe(A__ , generator=A__ , num_inference_steps=25 , output_type="""pt""" ).frames A__ : Dict = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5e-2 def __A ( self ): A__ : List[Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy""" ) A__ : Optional[int] = TextToVideoSDPipeline.from_pretrained("""damo-vilab/text-to-video-ms-1.7b""" ) A__ : List[str] = pipe.to("""cuda""" ) A__ : Dict = """Spiderman is surfing""" A__ : Union[str, Any] = torch.Generator(device="""cpu""" ).manual_seed(0 ) A__ : Optional[int] = pipe(A__ , generator=A__ , num_inference_steps=2 , output_type="""pt""" ).frames A__ : Optional[int] = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5e-2
64
1
import gc import random import tempfile import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.stable_diffusion_safe import StableDiffusionPipelineSafe as StableDiffusionPipeline from diffusers.utils import floats_tensor, nightly, torch_device from diffusers.utils.testing_utils import require_torch_gpu class _a (unittest.TestCase ): '''simple docstring''' def __A ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def __A ( self ): A__ : Optional[Any] = 1 A__ : List[Any] = 3 A__ : Optional[Any] = (32, 32) A__ : str = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(A__ ) return image @property def __A ( self ): torch.manual_seed(0 ) A__ : Tuple = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) return model @property def __A ( self ): torch.manual_seed(0 ) A__ : List[str] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) return model @property def __A ( self ): torch.manual_seed(0 ) A__ : Dict = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModel(A__ ) @property def __A ( self ): def extract(*A__ , **A__ ): class _a : '''simple docstring''' def __init__( self ): A__ : int = torch.ones([0] ) def __A ( self , A__ ): self.pixel_values.to(A__ ) return self return Out() return extract def __A ( self ): A__ : Tuple = """cpu""" # ensure determinism for the device-dependent torch.Generator A__ : Union[str, Any] = self.dummy_cond_unet A__ : List[Any] = DDIMScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule="""scaled_linear""" , clip_sample=A__ , set_alpha_to_one=A__ , ) A__ : Union[str, Any] = self.dummy_vae A__ : List[str] = self.dummy_text_encoder A__ : List[str] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) # make sure here that pndm scheduler skips prk A__ : Optional[int] = StableDiffusionPipeline( unet=A__ , scheduler=A__ , vae=A__ , text_encoder=A__ , tokenizer=A__ , safety_checker=A__ , feature_extractor=self.dummy_extractor , ) A__ : List[str] = sd_pipe.to(A__ ) sd_pipe.set_progress_bar_config(disable=A__ ) A__ : Optional[int] = """A painting of a squirrel eating a burger""" A__ : Tuple = torch.Generator(device=A__ ).manual_seed(0 ) A__ : Optional[int] = sd_pipe([prompt] , generator=A__ , guidance_scale=6.0 , num_inference_steps=2 , output_type="""np""" ) A__ : List[Any] = output.images A__ : List[str] = torch.Generator(device=A__ ).manual_seed(0 ) A__ : Optional[Any] = sd_pipe( [prompt] , generator=A__ , guidance_scale=6.0 , num_inference_steps=2 , output_type="""np""" , return_dict=A__ , )[0] A__ : Optional[Any] = image[0, -3:, -3:, -1] A__ : Tuple = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) A__ : int = np.array([0.5_7_5_6, 0.6_1_1_8, 0.5_0_0_5, 0.5_0_4_1, 0.5_4_7_1, 0.4_7_2_6, 0.4_9_7_6, 0.4_8_6_5, 0.4_8_6_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def __A ( self ): A__ : Optional[int] = """cpu""" # ensure determinism for the device-dependent torch.Generator A__ : List[Any] = self.dummy_cond_unet A__ : List[Any] = PNDMScheduler(skip_prk_steps=A__ ) A__ : Any = self.dummy_vae A__ : Union[str, Any] = self.dummy_text_encoder A__ : int = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) # make sure here that pndm scheduler skips prk A__ : str = StableDiffusionPipeline( unet=A__ , scheduler=A__ , vae=A__ , text_encoder=A__ , tokenizer=A__ , safety_checker=A__ , feature_extractor=self.dummy_extractor , ) A__ : Union[str, Any] = sd_pipe.to(A__ ) sd_pipe.set_progress_bar_config(disable=A__ ) A__ : int = """A painting of a squirrel eating a burger""" A__ : Dict = torch.Generator(device=A__ ).manual_seed(0 ) A__ : Any = sd_pipe([prompt] , generator=A__ , guidance_scale=6.0 , num_inference_steps=2 , output_type="""np""" ) A__ : Union[str, Any] = output.images A__ : int = torch.Generator(device=A__ ).manual_seed(0 ) A__ : Dict = sd_pipe( [prompt] , generator=A__ , guidance_scale=6.0 , num_inference_steps=2 , output_type="""np""" , return_dict=A__ , )[0] A__ : List[str] = image[0, -3:, -3:, -1] A__ : List[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) A__ : str = np.array([0.5_1_2_5, 0.5_7_1_6, 0.4_8_2_8, 0.5_0_6_0, 0.5_6_5_0, 0.4_7_6_8, 0.5_1_8_5, 0.4_8_9_5, 0.4_9_9_3] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def __A ( self ): A__ : Optional[Any] = StableDiffusionPipeline.from_pretrained( """hf-internal-testing/tiny-stable-diffusion-lms-pipe""" , safety_checker=A__ ) assert isinstance(A__ , A__ ) assert isinstance(pipe.scheduler , A__ ) assert pipe.safety_checker is None A__ : Dict = pipe("""example prompt""" , num_inference_steps=2 ).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(A__ ) A__ : Dict = StableDiffusionPipeline.from_pretrained(A__ ) # sanity check that the pipeline still works assert pipe.safety_checker is None A__ : Union[str, Any] = pipe("""example prompt""" , num_inference_steps=2 ).images[0] assert image is not None @unittest.skipIf(torch_device != """cuda""" , """This test requires a GPU""" ) def __A ( self ): A__ : str = self.dummy_cond_unet A__ : List[str] = PNDMScheduler(skip_prk_steps=A__ ) A__ : Union[str, Any] = self.dummy_vae A__ : List[str] = self.dummy_text_encoder A__ : List[Any] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) # put models in fp16 A__ : Optional[Any] = unet.half() A__ : int = vae.half() A__ : Optional[Any] = bert.half() # make sure here that pndm scheduler skips prk A__ : List[str] = StableDiffusionPipeline( unet=A__ , scheduler=A__ , vae=A__ , text_encoder=A__ , tokenizer=A__ , safety_checker=A__ , feature_extractor=self.dummy_extractor , ) A__ : Union[str, Any] = sd_pipe.to(A__ ) sd_pipe.set_progress_bar_config(disable=A__ ) A__ : Optional[Any] = """A painting of a squirrel eating a burger""" A__ : str = sd_pipe([prompt] , num_inference_steps=2 , output_type="""np""" ).images assert image.shape == (1, 64, 64, 3) @nightly @require_torch_gpu class _a (unittest.TestCase ): '''simple docstring''' def __A ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __A ( self ): A__ : Optional[int] = StableDiffusionPipeline.from_pretrained("""runwayml/stable-diffusion-v1-5""" , safety_checker=A__ ) A__ : List[str] = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config ) A__ : Tuple = sd_pipe.to(A__ ) sd_pipe.set_progress_bar_config(disable=A__ ) A__ : Any = ( """portrait of girl with smokey eyes makeup in abandoned hotel, grange clothes, redshift, wide high angle""" """ coloured polaroid photograph with flash, kodak film, hyper real, stunning moody cinematography, with""" """ anamorphic lenses, by maripol, fallen angels by wong kar - wai, style of suspiria and neon demon and""" """ children from bahnhof zoo, detailed """ ) A__ : Optional[int] = 40_0366_0346 A__ : Optional[Any] = 7 # without safety guidance (sld_guidance_scale = 0) A__ : Tuple = torch.manual_seed(A__ ) A__ : Tuple = sd_pipe( [prompt] , generator=A__ , guidance_scale=A__ , num_inference_steps=50 , output_type="""np""" , width=512 , height=512 , sld_guidance_scale=0 , ) A__ : Tuple = output.images A__ : List[Any] = image[0, -3:, -3:, -1] A__ : Tuple = [0.2_2_7_8, 0.2_2_3_1, 0.2_2_4_9, 0.2_3_3_3, 0.2_3_0_3, 0.1_8_8_5, 0.2_2_7_3, 0.2_1_4_4, 0.2_1_7_6] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 # without safety guidance (strong configuration) A__ : Union[str, Any] = torch.manual_seed(A__ ) A__ : Any = sd_pipe( [prompt] , generator=A__ , guidance_scale=A__ , num_inference_steps=50 , output_type="""np""" , width=512 , height=512 , sld_guidance_scale=2000 , sld_warmup_steps=7 , sld_threshold=0.0_2_5 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) A__ : Union[str, Any] = output.images A__ : List[Any] = image[0, -3:, -3:, -1] A__ : Optional[Any] = [0.2_3_8_3, 0.2_2_7_6, 0.2_3_6, 0.2_1_9_2, 0.2_1_8_6, 0.2_0_5_3, 0.1_9_7_1, 0.1_9_0_1, 0.1_7_1_9] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def __A ( self ): A__ : Any = StableDiffusionPipeline.from_pretrained("""runwayml/stable-diffusion-v1-5""" , safety_checker=A__ ) A__ : Tuple = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config ) A__ : Tuple = sd_pipe.to(A__ ) sd_pipe.set_progress_bar_config(disable=A__ ) A__ : Dict = """padme amidala taking a bath artwork, safe for work, no nudity""" A__ : List[str] = 27_3497_1755 A__ : Optional[Any] = 7 A__ : Union[str, Any] = torch.manual_seed(A__ ) A__ : List[str] = sd_pipe( [prompt] , generator=A__ , guidance_scale=A__ , num_inference_steps=50 , output_type="""np""" , width=512 , height=512 , sld_guidance_scale=0 , ) A__ : Union[str, Any] = output.images A__ : str = image[0, -3:, -3:, -1] A__ : Dict = [0.3_5_0_2, 0.3_6_2_2, 0.3_3_9_6, 0.3_6_4_2, 0.3_4_7_8, 0.3_3_1_8, 0.3_5, 0.3_3_4_8, 0.3_2_9_7] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 A__ : str = torch.manual_seed(A__ ) A__ : List[str] = sd_pipe( [prompt] , generator=A__ , guidance_scale=A__ , num_inference_steps=50 , output_type="""np""" , width=512 , height=512 , sld_guidance_scale=2000 , sld_warmup_steps=7 , sld_threshold=0.0_2_5 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) A__ : Optional[int] = output.images A__ : List[Any] = image[0, -3:, -3:, -1] A__ : int = [0.5_5_3_1, 0.5_2_0_6, 0.4_8_9_5, 0.5_1_5_6, 0.5_1_8_2, 0.4_7_5_1, 0.4_8_0_2, 0.4_8_0_3, 0.4_4_4_3] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def __A ( self ): A__ : List[str] = StableDiffusionPipeline.from_pretrained("""runwayml/stable-diffusion-v1-5""" ) A__ : List[Any] = sd_pipe.to(A__ ) sd_pipe.set_progress_bar_config(disable=A__ ) A__ : str = ( """the four horsewomen of the apocalypse, painting by tom of finland, gaston bussiere, craig mullins, j. c.""" """ leyendecker""" ) A__ : Optional[int] = 10_4435_5234 A__ : Optional[Any] = 12 A__ : Optional[int] = torch.manual_seed(A__ ) A__ : Dict = sd_pipe( [prompt] , generator=A__ , guidance_scale=A__ , num_inference_steps=50 , output_type="""np""" , width=512 , height=512 , sld_guidance_scale=0 , ) A__ : List[Any] = output.images A__ : Tuple = image[0, -3:, -3:, -1] A__ : str = np.array([0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0] ) assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-7 A__ : Union[str, Any] = torch.manual_seed(A__ ) A__ : List[Any] = sd_pipe( [prompt] , generator=A__ , guidance_scale=A__ , num_inference_steps=50 , output_type="""np""" , width=512 , height=512 , sld_guidance_scale=2000 , sld_warmup_steps=7 , sld_threshold=0.0_2_5 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) A__ : Any = output.images A__ : Dict = image[0, -3:, -3:, -1] A__ : Optional[int] = np.array([0.5_8_1_8, 0.6_2_8_5, 0.6_8_3_5, 0.6_0_1_9, 0.6_2_5, 0.6_7_5_4, 0.6_0_9_6, 0.6_3_3_4, 0.6_5_6_1] ) assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
64
def UpperCamelCase (lowercase_: int ) -> int: if not isinstance(lowercase_ , lowercase_ ): raise TypeError("""Input value must be an 'int' type""" ) A__ : int = 0 while number: position += 1 number >>= 1 return position if __name__ == "__main__": import doctest doctest.testmod()
64
1
from __future__ import annotations from collections.abc import Generator import requests from bsa import BeautifulSoup A_ : Union[str, Any] = 'https://www.indeed.co.in/jobs?q=mobile+app+development&l=' def UpperCamelCase (lowercase_: str = "mumbai" ) -> Generator[tuple[str, str], None, None]: A__ : Union[str, Any] = BeautifulSoup(requests.get(url + location ).content , """html.parser""" ) # This attribute finds out all the specifics listed in a job for job in soup.find_all("""div""" , attrs={"""data-tn-component""": """organicJob"""} ): A__ : str = job.find("""a""" , attrs={"""data-tn-element""": """jobTitle"""} ).text.strip() A__ : Optional[Any] = job.find("""span""" , {"""class""": """company"""} ).text.strip() yield job_title, company_name if __name__ == "__main__": for i, job in enumerate(fetch_jobs('Bangalore'), 1): print(f'''Job {i:>2} is {job[0]} at {job[1]}''')
64
from pathlib import Path import cva import numpy as np from matplotlib import pyplot as plt def UpperCamelCase (lowercase_: np.ndarray , lowercase_: np.ndarray , lowercase_: np.ndarray , lowercase_: int , lowercase_: int ) -> np.ndarray: A__ : Any = cva.getAffineTransform(lowercase_ , lowercase_ ) return cva.warpAffine(lowercase_ , lowercase_ , (rows, cols) ) if __name__ == "__main__": # read original image A_ : List[Any] = cva.imread( str(Path(__file__).resolve().parent.parent / 'image_data' / 'lena.jpg') ) # turn image in gray scale value A_ : List[Any] = cva.cvtColor(image, cva.COLOR_BGR2GRAY) # get image shape A_ , A_ : Optional[Any] = gray_img.shape # set different points to rotate image A_ : str = np.array([[50, 50], [200, 50], [50, 200]], np.floataa) A_ : Dict = np.array([[10, 100], [200, 50], [100, 250]], np.floataa) A_ : Optional[int] = np.array([[50, 50], [150, 50], [120, 200]], np.floataa) A_ : Optional[int] = np.array([[10, 100], [80, 50], [180, 250]], np.floataa) # add all rotated images in a list A_ : Dict = [ gray_img, get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), ] # plot different image rotations A_ : Union[str, Any] = plt.figure(1) A_ : Union[str, Any] = ['Original', 'Rotation 1', 'Rotation 2', 'Rotation 3'] for i, image in enumerate(images): plt.subplot(2, 2, i + 1), plt.imshow(image, 'gray') plt.title(titles[i]) plt.axis('off') plt.subplots_adjust(left=0.0, bottom=0.05, right=1.0, top=0.95) plt.show()
64
1
from ..utils import DummyObject, requires_backends class _a (metaclass=__magic_name__ ): '''simple docstring''' UpperCAmelCase__: List[str] = ['''torch''', '''transformers''', '''onnx'''] def __init__( self , *A__ , **A__ ): requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def __A ( cls , *A__ , **A__ ): requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def __A ( cls , *A__ , **A__ ): requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) class _a (metaclass=__magic_name__ ): '''simple docstring''' UpperCAmelCase__: int = ['''torch''', '''transformers''', '''onnx'''] def __init__( self , *A__ , **A__ ): requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def __A ( cls , *A__ , **A__ ): requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def __A ( cls , *A__ , **A__ ): requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) class _a (metaclass=__magic_name__ ): '''simple docstring''' UpperCAmelCase__: int = ['''torch''', '''transformers''', '''onnx'''] def __init__( self , *A__ , **A__ ): requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def __A ( cls , *A__ , **A__ ): requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def __A ( cls , *A__ , **A__ ): requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) class _a (metaclass=__magic_name__ ): '''simple docstring''' UpperCAmelCase__: List[str] = ['''torch''', '''transformers''', '''onnx'''] def __init__( self , *A__ , **A__ ): requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def __A ( cls , *A__ , **A__ ): requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def __A ( cls , *A__ , **A__ ): requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) class _a (metaclass=__magic_name__ ): '''simple docstring''' UpperCAmelCase__: Union[str, Any] = ['''torch''', '''transformers''', '''onnx'''] def __init__( self , *A__ , **A__ ): requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def __A ( cls , *A__ , **A__ ): requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def __A ( cls , *A__ , **A__ ): requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) class _a (metaclass=__magic_name__ ): '''simple docstring''' UpperCAmelCase__: str = ['''torch''', '''transformers''', '''onnx'''] def __init__( self , *A__ , **A__ ): requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def __A ( cls , *A__ , **A__ ): requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def __A ( cls , *A__ , **A__ ): requires_backends(cls , ["""torch""", """transformers""", """onnx"""] )
64
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_torch_available from transformers.testing_utils import require_torch, torch_device if is_torch_available(): from transformers import PyTorchBenchmark, PyTorchBenchmarkArguments @require_torch class _a (unittest.TestCase ): '''simple docstring''' def __A ( self , A__ ): for model_result in results.values(): for batch_size, sequence_length in zip(model_result["""bs"""] , model_result["""ss"""] ): A__ : str = model_result["""result"""][batch_size][sequence_length] self.assertIsNotNone(A__ ) def __A ( self ): A__ : Dict = """sshleifer/tiny-gpt2""" A__ : Tuple = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A__ , ) A__ : int = PyTorchBenchmark(A__ ) A__ : List[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self ): A__ : Dict = """sgugger/tiny-distilbert-classification""" A__ : Dict = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A__ , only_pretrain_model=A__ , ) A__ : str = PyTorchBenchmark(A__ ) A__ : List[str] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self ): A__ : Any = """sshleifer/tiny-gpt2""" A__ : List[Any] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , torchscript=A__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A__ , ) A__ : Tuple = PyTorchBenchmark(A__ ) A__ : str = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(torch_device == """cpu""" , """Cant do half precision""" ) def __A ( self ): A__ : Optional[Any] = """sshleifer/tiny-gpt2""" A__ : Optional[Any] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , fpaa=A__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A__ , ) A__ : str = PyTorchBenchmark(A__ ) A__ : Any = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self ): A__ : Optional[Any] = """sshleifer/tiny-gpt2""" A__ : Tuple = AutoConfig.from_pretrained(A__ ) # set architectures equal to `None` A__ : List[Any] = None A__ : str = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A__ , ) A__ : List[str] = PyTorchBenchmark(A__ , configs=[config] ) A__ : Any = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self ): A__ : Optional[int] = """sshleifer/tiny-gpt2""" A__ : Optional[int] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A__ , ) A__ : Any = PyTorchBenchmark(A__ ) A__ : Dict = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) @unittest.skipIf(torch_device == """cpu""" , """Can't do half precision""" ) def __A ( self ): A__ : Optional[int] = """sshleifer/tiny-gpt2""" A__ : List[str] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , sequence_lengths=[8] , batch_sizes=[1] , fpaa=A__ , multi_process=A__ , ) A__ : Dict = PyTorchBenchmark(A__ ) A__ : Optional[int] = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def __A ( self ): A__ : int = """sshleifer/tiny-gpt2""" A__ : Optional[int] = AutoConfig.from_pretrained(A__ ) A__ : str = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A__ , ) A__ : int = PyTorchBenchmark(A__ , configs=[config] ) A__ : Tuple = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self ): A__ : List[str] = """sshleifer/tinier_bart""" A__ : List[str] = AutoConfig.from_pretrained(A__ ) A__ : List[str] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A__ , ) A__ : Union[str, Any] = PyTorchBenchmark(A__ , configs=[config] ) A__ : str = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self ): A__ : Optional[int] = """sshleifer/tiny-gpt2""" A__ : Union[str, Any] = AutoConfig.from_pretrained(A__ ) A__ : Tuple = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A__ , ) A__ : int = PyTorchBenchmark(A__ , configs=[config] ) A__ : Tuple = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def __A ( self ): A__ : Dict = """sshleifer/tinier_bart""" A__ : int = AutoConfig.from_pretrained(A__ ) A__ : Union[str, Any] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A__ , ) A__ : List[Any] = PyTorchBenchmark(A__ , configs=[config] ) A__ : Optional[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def __A ( self ): A__ : int = """sshleifer/tiny-gpt2""" with tempfile.TemporaryDirectory() as tmp_dir: A__ : Dict = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , save_to_csv=A__ , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(A__ , """inf_time.csv""" ) , train_memory_csv_file=os.path.join(A__ , """train_mem.csv""" ) , inference_memory_csv_file=os.path.join(A__ , """inf_mem.csv""" ) , train_time_csv_file=os.path.join(A__ , """train_time.csv""" ) , env_info_csv_file=os.path.join(A__ , """env.csv""" ) , multi_process=A__ , ) A__ : Optional[Any] = PyTorchBenchmark(A__ ) benchmark.run() self.assertTrue(Path(os.path.join(A__ , """inf_time.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(A__ , """train_time.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(A__ , """inf_mem.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(A__ , """train_mem.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(A__ , """env.csv""" ) ).exists() ) def __A ( self ): A__ : Optional[int] = """sshleifer/tiny-gpt2""" def _check_summary_is_not_empty(A__ ): self.assertTrue(hasattr(A__ , """sequential""" ) ) self.assertTrue(hasattr(A__ , """cumulative""" ) ) self.assertTrue(hasattr(A__ , """current""" ) ) self.assertTrue(hasattr(A__ , """total""" ) ) with tempfile.TemporaryDirectory() as tmp_dir: A__ : Dict = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A__ , inference=A__ , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(A__ , """log.txt""" ) , log_print=A__ , trace_memory_line_by_line=A__ , multi_process=A__ , ) A__ : Dict = PyTorchBenchmark(A__ ) A__ : str = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) _check_summary_is_not_empty(result.train_summary ) self.assertTrue(Path(os.path.join(A__ , """log.txt""" ) ).exists() )
64
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available A_ : int = {'configuration_glpn': ['GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GLPNConfig']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Optional[int] = ['GLPNFeatureExtractor'] A_ : List[str] = ['GLPNImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Tuple = [ 'GLPN_PRETRAINED_MODEL_ARCHIVE_LIST', 'GLPNForDepthEstimation', 'GLPNLayer', 'GLPNModel', 'GLPNPreTrainedModel', ] if TYPE_CHECKING: from .configuration_glpn import GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP, GLPNConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_glpn import GLPNFeatureExtractor from .image_processing_glpn import GLPNImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_glpn import ( GLPN_PRETRAINED_MODEL_ARCHIVE_LIST, GLPNForDepthEstimation, GLPNLayer, GLPNModel, GLPNPreTrainedModel, ) else: import sys A_ : int = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
64
# tests directory-specific settings - this file is run automatically # by pytest before any tests are run import doctest import sys import warnings from os.path import abspath, dirname, join import _pytest from transformers.testing_utils import HfDoctestModule, HfDocTestParser # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. A_ : Optional[int] = abspath(join(dirname(__file__), 'src')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='ignore', category=FutureWarning) def UpperCamelCase (lowercase_: List[str] ) -> Any: config.addinivalue_line( """markers""" , """is_pt_tf_cross_test: mark test to run only when PT and TF interactions are tested""" ) config.addinivalue_line( """markers""" , """is_pt_flax_cross_test: mark test to run only when PT and FLAX interactions are tested""" ) config.addinivalue_line("""markers""" , """is_pipeline_test: mark test to run only when pipelines are tested""" ) config.addinivalue_line("""markers""" , """is_staging_test: mark test to run only in the staging environment""" ) config.addinivalue_line("""markers""" , """accelerate_tests: mark test that require accelerate""" ) config.addinivalue_line("""markers""" , """tool_tests: mark the tool tests that are run on their specific schedule""" ) def UpperCamelCase (lowercase_: Optional[int] ) -> Optional[Any]: from transformers.testing_utils import pytest_addoption_shared pytest_addoption_shared(lowercase_ ) def UpperCamelCase (lowercase_: List[str] ) -> Optional[Any]: from transformers.testing_utils import pytest_terminal_summary_main A__ : List[Any] = terminalreporter.config.getoption("""--make-reports""" ) if make_reports: pytest_terminal_summary_main(lowercase_ , id=lowercase_ ) def UpperCamelCase (lowercase_: Union[str, Any] , lowercase_: int ) -> List[str]: # If no tests are collected, pytest exists with code 5, which makes the CI fail. if exitstatus == 5: A__ : Tuple = 0 # Doctest custom flag to ignore output. A_ : Tuple = doctest.register_optionflag('IGNORE_RESULT') A_ : Dict = doctest.OutputChecker class _a (__magic_name__ ): '''simple docstring''' def __A ( self , A__ , A__ , A__ ): if IGNORE_RESULT & optionflags: return True return OutputChecker.check_output(self , A__ , A__ , A__ ) A_ : str = CustomOutputChecker A_ : Dict = HfDoctestModule A_ : Optional[int] = HfDocTestParser
64
1
import argparse import torch from transformers import LxmertConfig, LxmertForPreTraining, load_tf_weights_in_lxmert from transformers.utils import logging logging.set_verbosity_info() def UpperCamelCase (lowercase_: Optional[int] , lowercase_: Union[str, Any] , lowercase_: Optional[int] ) -> Optional[Any]: # Initialise PyTorch model A__ : str = LxmertConfig.from_json_file(lowercase_ ) print(f"""Building PyTorch model from configuration: {config}""" ) A__ : Union[str, Any] = LxmertForPreTraining(lowercase_ ) # Load weights from tf checkpoint load_tf_weights_in_lxmert(lowercase_ , lowercase_ , lowercase_ ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , lowercase_ ) if __name__ == "__main__": A_ : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help='The config json file corresponding to the pre-trained model. \nThis specifies the model architecture.', ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) A_ : List[Any] = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
64
from __future__ import annotations import unittest from transformers import AutoTokenizer, PegasusConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFPegasusForConditionalGeneration, TFPegasusModel @require_tf class _a : '''simple docstring''' UpperCAmelCase__: List[Any] = PegasusConfig UpperCAmelCase__: Optional[int] = {} UpperCAmelCase__: List[str] = '''gelu''' def __init__( self , A__ , A__=13 , A__=7 , A__=True , A__=False , A__=99 , A__=32 , A__=2 , A__=4 , A__=37 , A__=0.1 , A__=0.1 , A__=40 , A__=2 , A__=1 , A__=0 , ): A__ : Dict = parent A__ : Dict = batch_size A__ : Any = seq_length A__ : Optional[Any] = is_training A__ : int = use_labels A__ : Any = vocab_size A__ : Union[str, Any] = hidden_size A__ : Tuple = num_hidden_layers A__ : Tuple = num_attention_heads A__ : List[Any] = intermediate_size A__ : Union[str, Any] = hidden_dropout_prob A__ : Optional[Any] = attention_probs_dropout_prob A__ : List[Any] = max_position_embeddings A__ : Any = eos_token_id A__ : List[Any] = pad_token_id A__ : List[Any] = bos_token_id def __A ( self ): A__ : str = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) A__ : Dict = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) A__ : List[Any] = tf.concat([input_ids, eos_tensor] , axis=1 ) A__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ : Tuple = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) A__ : str = prepare_pegasus_inputs_dict(A__ , A__ , A__ ) return config, inputs_dict def __A ( self , A__ , A__ ): A__ : int = TFPegasusModel(config=A__ ).get_decoder() A__ : List[Any] = inputs_dict["""input_ids"""] A__ : Any = input_ids[:1, :] A__ : Optional[Any] = inputs_dict["""attention_mask"""][:1, :] A__ : Optional[int] = inputs_dict["""head_mask"""] A__ : Any = 1 # first forward pass A__ : Tuple = model(A__ , attention_mask=A__ , head_mask=A__ , use_cache=A__ ) A__ , A__ : Dict = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids A__ : Tuple = ids_tensor((self.batch_size, 3) , config.vocab_size ) A__ : Optional[Any] = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and A__ : List[Any] = tf.concat([input_ids, next_tokens] , axis=-1 ) A__ : Tuple = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) A__ : Optional[Any] = model(A__ , attention_mask=A__ )[0] A__ : Any = model(A__ , attention_mask=A__ , past_key_values=A__ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice A__ : int = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) A__ : Any = output_from_no_past[:, -3:, random_slice_idx] A__ : Tuple = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(A__ , A__ , rtol=1e-3 ) def UpperCamelCase (lowercase_: Union[str, Any] , lowercase_: Dict , lowercase_: List[Any] , lowercase_: Dict=None , lowercase_: int=None , lowercase_: List[Any]=None , lowercase_: List[Any]=None , lowercase_: str=None , ) -> int: if attention_mask is None: A__ : List[str] = tf.cast(tf.math.not_equal(lowercase_ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: A__ : Dict = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: A__ : Any = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: A__ : Tuple = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: A__ : List[str] = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class _a (__magic_name__ , __magic_name__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__: List[Any] = (TFPegasusForConditionalGeneration, TFPegasusModel) if is_tf_available() else () UpperCAmelCase__: Tuple = (TFPegasusForConditionalGeneration,) if is_tf_available() else () UpperCAmelCase__: Tuple = ( { '''conversational''': TFPegasusForConditionalGeneration, '''feature-extraction''': TFPegasusModel, '''summarization''': TFPegasusForConditionalGeneration, '''text2text-generation''': TFPegasusForConditionalGeneration, '''translation''': TFPegasusForConditionalGeneration, } if is_tf_available() else {} ) UpperCAmelCase__: int = True UpperCAmelCase__: Union[str, Any] = False UpperCAmelCase__: List[str] = False def __A ( self ): A__ : Optional[Any] = TFPegasusModelTester(self ) A__ : Tuple = ConfigTester(self , config_class=A__ ) def __A ( self ): self.config_tester.run_common_tests() def __A ( self ): A__ : int = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*A__ ) @require_sentencepiece @require_tokenizers @require_tf class _a (unittest.TestCase ): '''simple docstring''' UpperCAmelCase__: Optional[int] = [ ''' PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.''', ''' The London trio are up for best UK act and best album, as well as getting two nominations in the best song category."We got told like this morning \'Oh I think you\'re nominated\'", said Dappy."And I was like \'Oh yeah, which one?\' And now we\'ve got nominated for four awards. I mean, wow!"Bandmate Fazer added: "We thought it\'s best of us to come down and mingle with everyone and say hello to the cameras. And now we find we\'ve got four nominations."The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn\'t be too disappointed if they didn\'t win this time around."At the end of the day we\'re grateful to be where we are in our careers."If it don\'t happen then it don\'t happen - live to fight another day and keep on making albums and hits for the fans."Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers\' All These Things That I\'ve Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year\'s Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border."We just done Edinburgh the other day," said Dappy."We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!" ''', ] UpperCAmelCase__: Any = [ '''California\'s largest electricity provider has cut power to hundreds of thousands of customers in an effort to''' ''' reduce the risk of wildfires.''', '''N-Dubz have revealed they\'re "grateful" to have been nominated for four Mobo Awards.''', ] # differs slightly from pytorch, likely due to numerical differences in linear layers UpperCAmelCase__: List[str] = '''google/pegasus-xsum''' @cached_property def __A ( self ): return AutoTokenizer.from_pretrained(self.model_name ) @cached_property def __A ( self ): A__ : int = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model def __A ( self , **A__ ): A__ : str = self.translate_src_text(**A__ ) assert self.expected_text == generated_words def __A ( self , **A__ ): A__ : List[str] = self.tokenizer(self.src_text , **A__ , padding=A__ , return_tensors="""tf""" ) A__ : Optional[int] = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 , use_cache=A__ , ) A__ : Dict = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=A__ ) return generated_words @slow def __A ( self ): self._assert_generated_batch_equal_expected()
64
1
import darl # noqa import gym import tqdm from diffusers.experimental import ValueGuidedRLPipeline A_ : List[Any] = { 'n_samples': 64, 'horizon': 32, 'num_inference_steps': 20, 'n_guide_steps': 2, # can set to 0 for faster sampling, does not use value network 'scale_grad_by_std': True, 'scale': 0.1, 'eta': 0.0, 't_grad_cutoff': 2, 'device': 'cpu', } if __name__ == "__main__": A_ : str = 'hopper-medium-v2' A_ : Dict = gym.make(env_name) A_ : str = ValueGuidedRLPipeline.from_pretrained( 'bglick13/hopper-medium-v2-value-function-hor32', env=env, ) env.seed(0) A_ : List[Any] = env.reset() A_ : str = 0 A_ : List[str] = 0 A_ : Any = 1000 A_ : Optional[Any] = [obs.copy()] try: for t in tqdm.tqdm(range(T)): # call the policy A_ : int = pipeline(obs, planning_horizon=32) # execute action in environment A_ , A_ , A_ , A_ : Any = env.step(denorm_actions) A_ : Any = env.get_normalized_score(total_reward) # update return total_reward += reward total_score += score print( f'''Step: {t}, Reward: {reward}, Total Reward: {total_reward}, Score: {score}, Total Score:''' f''' {total_score}''' ) # save observations for rendering rollout.append(next_observation.copy()) A_ : List[Any] = next_observation except KeyboardInterrupt: pass print(f'''Total reward: {total_reward}''')
64
class _a : '''simple docstring''' def __init__( self ): A__ : str = """""" A__ : Any = """""" A__ : List[Any] = [] def __A ( self , A__ , A__ ): if m == -1: return n + 1 elif n == -1: return m + 1 elif self.dp[m][n] > -1: return self.dp[m][n] else: if self.worda[m] == self.worda[n]: A__ : Optional[Any] = self.__min_dist_top_down_dp(m - 1 , n - 1 ) else: A__ : Union[str, Any] = self.__min_dist_top_down_dp(A__ , n - 1 ) A__ : Union[str, Any] = self.__min_dist_top_down_dp(m - 1 , A__ ) A__ : Union[str, Any] = self.__min_dist_top_down_dp(m - 1 , n - 1 ) A__ : List[Any] = 1 + min(A__ , A__ , A__ ) return self.dp[m][n] def __A ( self , A__ , A__ ): A__ : Tuple = worda A__ : Dict = worda A__ : Optional[Any] = [[-1 for _ in range(len(A__ ) )] for _ in range(len(A__ ) )] return self.__min_dist_top_down_dp(len(A__ ) - 1 , len(A__ ) - 1 ) def __A ( self , A__ , A__ ): A__ : Optional[Any] = worda A__ : Dict = worda A__ : Union[str, Any] = len(A__ ) A__ : List[str] = len(A__ ) A__ : int = [[0 for _ in range(n + 1 )] for _ in range(m + 1 )] for i in range(m + 1 ): for j in range(n + 1 ): if i == 0: # first string is empty A__ : Tuple = j elif j == 0: # second string is empty A__ : Dict = i elif worda[i - 1] == worda[j - 1]: # last characters are equal A__ : str = self.dp[i - 1][j - 1] else: A__ : Union[str, Any] = self.dp[i][j - 1] A__ : str = self.dp[i - 1][j] A__ : Union[str, Any] = self.dp[i - 1][j - 1] A__ : Tuple = 1 + min(A__ , A__ , A__ ) return self.dp[m][n] if __name__ == "__main__": A_ : Union[str, Any] = EditDistance() print('****************** Testing Edit Distance DP Algorithm ******************') print() A_ : int = input('Enter the first string: ').strip() A_ : List[str] = input('Enter the second string: ').strip() print() print(f'''The minimum edit distance is: {solver.min_dist_top_down(Sa, Sa)}''') print(f'''The minimum edit distance is: {solver.min_dist_bottom_up(Sa, Sa)}''') print() print('*************** End of Testing Edit Distance DP Algorithm ***************')
64
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) A_ : Optional[int] = { 'configuration_vision_encoder_decoder': ['VisionEncoderDecoderConfig', 'VisionEncoderDecoderOnnxConfig'] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Tuple = ['VisionEncoderDecoderModel'] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Tuple = ['TFVisionEncoderDecoderModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Optional[int] = ['FlaxVisionEncoderDecoderModel'] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys A_ : Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
64
def UpperCamelCase (lowercase_: int , lowercase_: int ) -> int: while second != 0: A__ : int = first & second first ^= second A__ : int = c << 1 return first if __name__ == "__main__": import doctest doctest.testmod() A_ : Optional[Any] = int(input('Enter the first number: ').strip()) A_ : List[str] = int(input('Enter the second number: ').strip()) print(f'''{add(first, second) = }''')
64
1
from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging A_ : List[str] = logging.get_logger(__name__) class _a (__magic_name__ ): '''simple docstring''' UpperCAmelCase__: Any = ['''audio_values''', '''audio_mask'''] def __init__( self , A__=2048 , A__=1 , A__=[16, 16] , A__=128 , A__=4_4100 , A__=86 , A__=2048 , A__=0.0 , **A__ , ): super().__init__( feature_size=A__ , sampling_rate=A__ , padding_value=A__ , **A__ , ) A__ : int = spectrogram_length A__ : List[str] = num_channels A__ : Tuple = patch_size A__ : Any = feature_size // self.patch_size[1] A__ : Tuple = n_fft A__ : Any = sampling_rate // hop_length_to_sampling_rate A__ : Dict = sampling_rate A__ : Dict = padding_value A__ : Dict = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=A__ , min_frequency=0.0 , max_frequency=2_2_0_5_0.0 , sampling_rate=A__ , norm="""slaney""" , mel_scale="""slaney""" , ).T def __A ( self , A__ ): A__ : List[Any] = spectrogram( A__ , window_function(self.n_fft , """hann""" ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel="""dB""" , db_range=8_0.0 , ) A__ : List[Any] = log_spec[:, :-1] A__ : Dict = log_spec - 2_0.0 A__ : List[str] = np.clip(log_spec / 4_0.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__( self , A__ , A__ = None , A__ = True , A__ = None , A__ = False , A__ = False , **A__ , ): if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( """This feature extractor is set to support sampling rate""" F""" of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled""" F""" with {self.sampling_rate} and not {sampling_rate}.""" ) else: logger.warning( """It is strongly recommended to pass the `sampling_rate` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) A__ : Optional[int] = isinstance(A__ , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F"""Only mono-channel audio is supported for input to {self}""" ) A__ : Optional[int] = is_batched_numpy or ( isinstance(A__ , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: A__ : int = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(A__ , np.ndarray ): A__ : List[Any] = np.asarray(A__ , dtype=np.floataa ) elif isinstance(A__ , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): A__ : int = raw_speech.astype(np.floataa ) # always return batch if not is_batched: A__ : Dict = [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis A__ : List[str] = [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , A__ ): A__ : Dict = [np.asarray(A__ , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask A__ : Optional[int] = max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: A__ : str = [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] A__ : Tuple = np.array(A__ ).astype(np.floataa ) # convert into correct format for padding A__ : Tuple = max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch A__ : str = np.ones([len(A__ ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) A__ : Tuple = padded_audio_features * self.padding_value for i in range(len(A__ ) ): A__ : List[Any] = audio_features[i] A__ : List[str] = feature # return as BatchFeature if return_attention_mask: A__ : Union[str, Any] = {"""audio_values""": padded_audio_features, """audio_mask""": audio_mask} else: A__ : Optional[int] = {"""audio_values""": padded_audio_features} A__ : List[Any] = BatchFeature(data=A__ , tensor_type=A__ ) return encoded_inputs
64
from __future__ import annotations from collections.abc import Callable A_ : List[Any] = list[list[float | int]] def UpperCamelCase (lowercase_: Matrix , lowercase_: Matrix ) -> Matrix: A__ : int = len(lowercase_ ) A__ : Matrix = [[0 for _ in range(size + 1 )] for _ in range(lowercase_ )] A__ : int A__ : int A__ : int A__ : int A__ : int A__ : float for row in range(lowercase_ ): for col in range(lowercase_ ): A__ : List[str] = matrix[row][col] A__ : int = vector[row][0] A__ : Optional[int] = 0 A__ : str = 0 while row < size and col < size: # pivoting A__ : int = max((abs(augmented[rowa][col] ), rowa) for rowa in range(lowercase_ , lowercase_ ) )[ 1 ] if augmented[pivot_row][col] == 0: col += 1 continue else: A__ , A__ : Union[str, Any] = augmented[pivot_row], augmented[row] for rowa in range(row + 1 , lowercase_ ): A__ : List[Any] = augmented[rowa][col] / augmented[row][col] A__ : Dict = 0 for cola in range(col + 1 , size + 1 ): augmented[rowa][cola] -= augmented[row][cola] * ratio row += 1 col += 1 # back substitution for col in range(1 , lowercase_ ): for row in range(lowercase_ ): A__ : List[str] = augmented[row][col] / augmented[col][col] for cola in range(lowercase_ , size + 1 ): augmented[row][cola] -= augmented[col][cola] * ratio # round to get rid of numbers like 2.000000000000004 return [ [round(augmented[row][size] / augmented[row][row] , 10 )] for row in range(lowercase_ ) ] def UpperCamelCase (lowercase_: list[int] ) -> Callable[[int], int]: A__ : int = len(lowercase_ ) A__ : Matrix = [[0 for _ in range(lowercase_ )] for _ in range(lowercase_ )] A__ : Matrix = [[0] for _ in range(lowercase_ )] A__ : Matrix A__ : int A__ : int A__ : int for x_val, y_val in enumerate(lowercase_ ): for col in range(lowercase_ ): A__ : Dict = (x_val + 1) ** (size - col - 1) A__ : Any = y_val A__ : Union[str, Any] = solve(lowercase_ , lowercase_ ) def interpolated_func(lowercase_: int ) -> int: return sum( round(coeffs[x_val][0] ) * (var ** (size - x_val - 1)) for x_val in range(lowercase_ ) ) return interpolated_func def UpperCamelCase (lowercase_: int ) -> int: return ( 1 - variable + variable**2 - variable**3 + variable**4 - variable**5 + variable**6 - variable**7 + variable**8 - variable**9 + variable**10 ) def UpperCamelCase (lowercase_: Callable[[int], int] = question_function , lowercase_: int = 10 ) -> int: A__ : list[int] = [func(lowercase_ ) for x_val in range(1 , order + 1 )] A__ : list[Callable[[int], int]] = [ interpolate(data_points[:max_coeff] ) for max_coeff in range(1 , order + 1 ) ] A__ : int = 0 A__ : Callable[[int], int] A__ : int for poly in polynomials: A__ : List[str] = 1 while func(lowercase_ ) == poly(lowercase_ ): x_val += 1 ret += poly(lowercase_ ) return ret if __name__ == "__main__": print(f'''{solution() = }''')
64
1
def UpperCamelCase (lowercase_: list[int] , lowercase_: list[int] , lowercase_: int ) -> bool: return not any( neighbour == 1 and colored_vertices[i] == color for i, neighbour in enumerate(lowercase_ ) ) def UpperCamelCase (lowercase_: list[list[int]] , lowercase_: int , lowercase_: list[int] , lowercase_: int ) -> bool: # Base Case if index == len(lowercase_ ): return True # Recursive Step for i in range(lowercase_ ): if valid_coloring(graph[index] , lowercase_ , lowercase_ ): # Color current vertex A__ : Optional[Any] = i # Validate coloring if util_color(lowercase_ , lowercase_ , lowercase_ , index + 1 ): return True # Backtrack A__ : str = -1 return False def UpperCamelCase (lowercase_: list[list[int]] , lowercase_: int ) -> list[int]: A__ : Dict = [-1] * len(lowercase_ ) if util_color(lowercase_ , lowercase_ , lowercase_ , 0 ): return colored_vertices return []
64
from functools import lru_cache @lru_cache def UpperCamelCase (lowercase_: int ) -> int: if num < 0: raise ValueError("""Number should not be negative.""" ) return 1 if num in (0, 1) else num * factorial(num - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
64
1
# tests directory-specific settings - this file is run automatically # by pytest before any tests are run import doctest import sys import warnings from os.path import abspath, dirname, join import _pytest from transformers.testing_utils import HfDoctestModule, HfDocTestParser # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. A_ : Optional[int] = abspath(join(dirname(__file__), 'src')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='ignore', category=FutureWarning) def UpperCamelCase (lowercase_: List[str] ) -> Any: config.addinivalue_line( """markers""" , """is_pt_tf_cross_test: mark test to run only when PT and TF interactions are tested""" ) config.addinivalue_line( """markers""" , """is_pt_flax_cross_test: mark test to run only when PT and FLAX interactions are tested""" ) config.addinivalue_line("""markers""" , """is_pipeline_test: mark test to run only when pipelines are tested""" ) config.addinivalue_line("""markers""" , """is_staging_test: mark test to run only in the staging environment""" ) config.addinivalue_line("""markers""" , """accelerate_tests: mark test that require accelerate""" ) config.addinivalue_line("""markers""" , """tool_tests: mark the tool tests that are run on their specific schedule""" ) def UpperCamelCase (lowercase_: Optional[int] ) -> Optional[Any]: from transformers.testing_utils import pytest_addoption_shared pytest_addoption_shared(lowercase_ ) def UpperCamelCase (lowercase_: List[str] ) -> Optional[Any]: from transformers.testing_utils import pytest_terminal_summary_main A__ : List[Any] = terminalreporter.config.getoption("""--make-reports""" ) if make_reports: pytest_terminal_summary_main(lowercase_ , id=lowercase_ ) def UpperCamelCase (lowercase_: Union[str, Any] , lowercase_: int ) -> List[str]: # If no tests are collected, pytest exists with code 5, which makes the CI fail. if exitstatus == 5: A__ : Tuple = 0 # Doctest custom flag to ignore output. A_ : Tuple = doctest.register_optionflag('IGNORE_RESULT') A_ : Dict = doctest.OutputChecker class _a (__magic_name__ ): '''simple docstring''' def __A ( self , A__ , A__ , A__ ): if IGNORE_RESULT & optionflags: return True return OutputChecker.check_output(self , A__ , A__ , A__ ) A_ : str = CustomOutputChecker A_ : Dict = HfDoctestModule A_ : Optional[int] = HfDocTestParser
64
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class _a (datasets.BeamBasedBuilder ): '''simple docstring''' def __A ( self ): return datasets.DatasetInfo( features=datasets.Features({"""content""": datasets.Value("""string""" )} ) , supervised_keys=A__ , ) def __A ( self , A__ , A__ ): return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"""examples""": get_test_dummy_examples()} )] def __A ( self , A__ , A__ ): import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(A__ ) class _a (datasets.BeamBasedBuilder ): '''simple docstring''' def __A ( self ): return datasets.DatasetInfo( features=datasets.Features({"""a""": datasets.Sequence({"""b""": datasets.Value("""string""" )} )} ) , supervised_keys=A__ , ) def __A ( self , A__ , A__ ): return [ datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"""examples""": get_test_nested_examples()} ) ] def __A ( self , A__ , A__ ): import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(A__ ) def UpperCamelCase () -> Dict: return [(i, {"content": content}) for i, content in enumerate(["""foo""", """bar""", """foobar"""] )] def UpperCamelCase () -> Tuple: return [(i, {"a": {"b": [content]}}) for i, content in enumerate(["""foo""", """bar""", """foobar"""] )] class _a (__magic_name__ ): '''simple docstring''' @require_beam def __A ( self ): A__ : Dict = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: A__ : int = DummyBeamDataset(cache_dir=A__ , beam_runner="""DirectRunner""" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(A__ , builder.name , """default""" , """0.0.0""" , F"""{builder.name}-train.arrow""" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"""content""": datasets.Value("""string""" )} ) ) A__ : int = builder.as_dataset() self.assertEqual(dset["""train"""].num_rows , A__ ) self.assertEqual(dset["""train"""].info.splits["""train"""].num_examples , A__ ) self.assertDictEqual(dset["""train"""][0] , get_test_dummy_examples()[0][1] ) self.assertDictEqual( dset["""train"""][expected_num_examples - 1] , get_test_dummy_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(A__ , builder.name , """default""" , """0.0.0""" , """dataset_info.json""" ) ) ) del dset @require_beam def __A ( self ): import apache_beam as beam A__ : int = beam.io.parquetio.WriteToParquet A__ : List[str] = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: A__ : str = DummyBeamDataset(cache_dir=A__ , beam_runner="""DirectRunner""" ) with patch("""apache_beam.io.parquetio.WriteToParquet""" ) as write_parquet_mock: A__ : Optional[Any] = partial(A__ , num_shards=2 ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join( A__ , builder.name , """default""" , """0.0.0""" , F"""{builder.name}-train-00000-of-00002.arrow""" ) ) ) self.assertTrue( os.path.exists( os.path.join( A__ , builder.name , """default""" , """0.0.0""" , F"""{builder.name}-train-00000-of-00002.arrow""" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"""content""": datasets.Value("""string""" )} ) ) A__ : Optional[int] = builder.as_dataset() self.assertEqual(dset["""train"""].num_rows , A__ ) self.assertEqual(dset["""train"""].info.splits["""train"""].num_examples , A__ ) # Order is not preserved when sharding, so we just check that all the elements are there self.assertListEqual(sorted(dset["""train"""]["""content"""] ) , sorted(["""foo""", """bar""", """foobar"""] ) ) self.assertTrue( os.path.exists(os.path.join(A__ , builder.name , """default""" , """0.0.0""" , """dataset_info.json""" ) ) ) del dset @require_beam def __A ( self ): with tempfile.TemporaryDirectory() as tmp_cache_dir: A__ : int = DummyBeamDataset(cache_dir=A__ ) self.assertRaises(datasets.builder.MissingBeamOptions , builder.download_and_prepare ) @require_beam def __A ( self ): A__ : List[Any] = len(get_test_nested_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: A__ : Optional[int] = NestedBeamDataset(cache_dir=A__ , beam_runner="""DirectRunner""" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(A__ , builder.name , """default""" , """0.0.0""" , F"""{builder.name}-train.arrow""" ) ) ) self.assertDictEqual( builder.info.features , datasets.Features({"""a""": datasets.Sequence({"""b""": datasets.Value("""string""" )} )} ) ) A__ : Optional[int] = builder.as_dataset() self.assertEqual(dset["""train"""].num_rows , A__ ) self.assertEqual(dset["""train"""].info.splits["""train"""].num_examples , A__ ) self.assertDictEqual(dset["""train"""][0] , get_test_nested_examples()[0][1] ) self.assertDictEqual( dset["""train"""][expected_num_examples - 1] , get_test_nested_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(A__ , builder.name , """default""" , """0.0.0""" , """dataset_info.json""" ) ) ) del dset
64
1
from PIL import Image def UpperCamelCase (lowercase_: Image ) -> Image: A__ , A__ : List[Any] = image.size A__ : Optional[int] = 0 A__ : Tuple = image.load() for i in range(lowercase_ ): for j in range(lowercase_ ): A__ : int = pixels[j, i] mean += pixel mean //= width * height for j in range(lowercase_ ): for i in range(lowercase_ ): A__ : List[Any] = 255 if pixels[i, j] > mean else 0 return image if __name__ == "__main__": A_ : Dict = mean_threshold(Image.open('path_to_image').convert('L')) image.save('output_image_path')
64
import warnings from ...utils import logging from .image_processing_poolformer import PoolFormerImageProcessor A_ : Union[str, Any] = logging.get_logger(__name__) class _a (__magic_name__ ): '''simple docstring''' def __init__( self , *A__ , **A__ ): warnings.warn( """The class PoolFormerFeatureExtractor is deprecated and will be removed in version 5 of Transformers.""" """ Please use PoolFormerImageProcessor instead.""" , A__ , ) super().__init__(*A__ , **A__ )
64
1
import logging import os from typing import List, Tuple import numpy as np import psutil import torch import torch.distributed as dist from transformers import RagRetriever A_ : List[Any] = logging.getLogger(__name__) class _a (__magic_name__ ): '''simple docstring''' def __init__( self , A__ , A__ , A__ , A__=None ): super().__init__( A__ , question_encoder_tokenizer=A__ , generator_tokenizer=A__ , index=A__ , init_retrieval=A__ , ) A__ : List[str] = None def __A ( self , A__ ): logger.info("""initializing retrieval""" ) # initializing a separate process group for retrieval as the default # nccl backend doesn't support gather/scatter operations while gloo # is too slow to replace nccl for the core gpu communication if dist.is_initialized(): logger.info("""dist initialized""" ) # needs to be set manually A__ : List[Any] = self._infer_socket_ifname() # avoid clash with the NCCL port A__ : Tuple = str(distributed_port + 1 ) A__ : Dict = dist.new_group(ranks=A__ , backend="""gloo""" ) # initialize retriever only on the main worker if not dist.is_initialized() or self._is_main(): logger.info("""dist not initialized / main""" ) self.index.init_index() # all processes wait untill the retriever is initialized by the main process if dist.is_initialized(): torch.distributed.barrier(group=self.process_group ) def __A ( self ): return dist.get_rank(group=self.process_group ) == 0 def __A ( self , A__ , A__ , A__=torch.floataa ): A__ : Tuple = torch.empty(A__ , dtype=A__ ) dist.scatter(A__ , src=0 , scatter_list=A__ , group=self.process_group ) return target_tensor def __A ( self ): A__ : Union[str, Any] = psutil.net_if_addrs() # a hacky way to deal with varying network interface names A__ : Optional[int] = next((addr for addr in addrs if addr.startswith("""e""" )) , A__ ) return ifname def __A ( self , A__ , A__ ): # single GPU training if not dist.is_initialized(): A__ , A__ : str = self._main_retrieve(A__ , A__ ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(A__ ) # distributed training A__ : str = dist.get_world_size(group=self.process_group ) # gather logic A__ : Dict = None if self._is_main(): A__ : Any = [torch.empty(question_hidden_states.shape , dtype=torch.floataa ) for _ in range(A__ )] dist.gather(torch.tensor(A__ ) , dst=0 , gather_list=A__ , group=self.process_group ) # scatter logic A__ : Union[str, Any] = question_hidden_states.shape[0] A__ : List[str] = [] A__ : Union[str, Any] = [] if self._is_main(): assert len(A__ ) == world_size A__ , A__ : Any = self._main_retrieve(torch.cat(A__ ).numpy() , A__ ) A__ , A__ : int = torch.tensor(A__ ), torch.tensor(A__ ) A__ : int = self._chunk_tensor(A__ , A__ ) A__ : Dict = self._chunk_tensor(A__ , A__ ) A__ : Optional[int] = self._scattered(A__ , [n_queries, n_docs] , target_type=torch.intaa ) A__ : Optional[int] = self._scattered(A__ , [n_queries, n_docs, question_hidden_states.shape[1]] ) return retrieved_doc_embeds.numpy(), doc_ids.numpy(), self.index.get_doc_dicts(A__ )
64
import inspect import logging import os import random import shutil import tempfile import unittest import pytest import torch from torch import nn from torch.utils.data import DataLoader, TensorDataset from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_cuda from accelerate.utils import ProjectConfiguration, set_seed A_ : Any = logging.getLogger(__name__) def UpperCamelCase (lowercase_: Optional[Any]=2 , lowercase_: Union[str, Any]=3 , lowercase_: int=16 , lowercase_: int = 10 , lowercase_: int = 2 ) -> int: def get_dataset(lowercase_: Optional[int] ): A__ : Optional[Any] = torch.randn(batch_size * n_batches , 1 ) return TensorDataset(lowercase_ , a * x + b + 0.1 * torch.randn(batch_size * n_batches , 1 ) ) A__ : Dict = get_dataset(lowercase_ ) A__ : Any = get_dataset(lowercase_ ) A__ : Dict = DataLoader(lowercase_ , shuffle=lowercase_ , batch_size=lowercase_ , num_workers=4 ) A__ : Optional[Any] = DataLoader(lowercase_ , shuffle=lowercase_ , batch_size=lowercase_ , num_workers=4 ) return (train_dataloader, valid_dataloader) def UpperCamelCase (lowercase_: Optional[Any] , lowercase_: List[str] , lowercase_: int , lowercase_: int , lowercase_: List[str] , lowercase_: Dict=None ) -> List[Any]: A__ : List[Any] = [] for epoch in range(lowercase_ ): # Train quickly model.train() for batch in dataloader: A__ , A__ : Any = batch A__ : Any = model(lowercase_ ) A__ : Any = torch.nn.functional.mse_loss(lowercase_ , lowercase_ ) accelerator.backward(lowercase_ ) optimizer.step() optimizer.zero_grad() rands.append(random.random() ) # Introduce some randomness if scheduler is not None: scheduler.step() return rands class _a (nn.Module ): '''simple docstring''' def __init__( self ): super().__init__() A__ : str = nn.Parameter(torch.randn(1 ) ) A__ : Any = nn.Parameter(torch.randn(1 ) ) def __A ( self , A__ ): return x * self.a + self.b class _a (unittest.TestCase ): '''simple docstring''' def __A ( self ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) A__ : Optional[Any] = DummyModel() A__ : Optional[Any] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) A__ , A__ : str = dummy_dataloaders() A__ : Dict = ProjectConfiguration(total_limit=1 , project_dir=A__ , automatic_checkpoint_naming=A__ ) # Train baseline A__ : List[str] = Accelerator(project_config=A__ ) A__ , A__ , A__ , A__ : Any = accelerator.prepare( A__ , A__ , A__ , A__ ) # Save initial accelerator.save_state() # Save second state accelerator.save_state() self.assertEqual(len(os.listdir(accelerator.project_dir ) ) , 1 ) def __A ( self ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) A__ : str = DummyModel() A__ : Optional[int] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) A__ , A__ : int = dummy_dataloaders() # Train baseline A__ : str = Accelerator() A__ , A__ , A__ , A__ : List[str] = accelerator.prepare( A__ , A__ , A__ , A__ ) # Save initial A__ : List[Any] = os.path.join(A__ , """initial""" ) accelerator.save_state(A__ ) ((A__) , (A__)) : str = model.a.item(), model.b.item() A__ : Dict = optimizer.state_dict() A__ : List[str] = train(3 , A__ , A__ , A__ , A__ ) ((A__) , (A__)) : str = model.a.item(), model.b.item() A__ : Any = optimizer.state_dict() # Train partially set_seed(42 ) A__ : Optional[int] = DummyModel() A__ : Dict = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) A__ , A__ : Dict = dummy_dataloaders() A__ : List[str] = Accelerator() A__ , A__ , A__ , A__ : Optional[Any] = accelerator.prepare( A__ , A__ , A__ , A__ ) accelerator.load_state(A__ ) ((A__) , (A__)) : Tuple = model.a.item(), model.b.item() A__ : Union[str, Any] = optimizer.state_dict() self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) A__ : List[str] = train(2 , A__ , A__ , A__ , A__ ) # Save everything A__ : Optional[int] = os.path.join(A__ , """checkpoint""" ) accelerator.save_state(A__ ) # Load everything back in and make sure all states work accelerator.load_state(A__ ) test_rands += train(1 , A__ , A__ , A__ , A__ ) ((A__) , (A__)) : Union[str, Any] = model.a.item(), model.b.item() A__ : Optional[int] = optimizer.state_dict() self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) def __A ( self ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) A__ : int = DummyModel() A__ : Optional[int] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) A__ , A__ : List[str] = dummy_dataloaders() A__ : str = ProjectConfiguration(automatic_checkpoint_naming=A__ ) # Train baseline A__ : Any = Accelerator(project_dir=A__ , project_config=A__ ) A__ , A__ , A__ , A__ : str = accelerator.prepare( A__ , A__ , A__ , A__ ) # Save initial accelerator.save_state() ((A__) , (A__)) : Tuple = model.a.item(), model.b.item() A__ : int = optimizer.state_dict() A__ : int = train(3 , A__ , A__ , A__ , A__ ) ((A__) , (A__)) : Optional[Any] = model.a.item(), model.b.item() A__ : Any = optimizer.state_dict() # Train partially set_seed(42 ) A__ : Dict = DummyModel() A__ : List[Any] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) A__ , A__ : Union[str, Any] = dummy_dataloaders() A__ : List[Any] = ProjectConfiguration(iteration=1 , automatic_checkpoint_naming=A__ ) A__ : Dict = Accelerator(project_dir=A__ , project_config=A__ ) A__ , A__ , A__ , A__ : Union[str, Any] = accelerator.prepare( A__ , A__ , A__ , A__ ) accelerator.load_state(os.path.join(A__ , """checkpoints""" , """checkpoint_0""" ) ) ((A__) , (A__)) : Optional[int] = model.a.item(), model.b.item() A__ : Tuple = optimizer.state_dict() self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) A__ : str = train(2 , A__ , A__ , A__ , A__ ) # Save everything accelerator.save_state() # Load everything back in and make sure all states work accelerator.load_state(os.path.join(A__ , """checkpoints""" , """checkpoint_1""" ) ) test_rands += train(1 , A__ , A__ , A__ , A__ ) ((A__) , (A__)) : Optional[int] = model.a.item(), model.b.item() A__ : List[Any] = optimizer.state_dict() self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) def __A ( self ): A__ : Union[str, Any] = torch.tensor([1, 2, 3] ) A__ : int = torch.tensor([2, 3, 4] ) A__ : List[Any] = DummyModel() A__ : List[Any] = torch.optim.Adam(net.parameters() ) A__ : Tuple = Accelerator() with self.assertRaises(A__ ) as ve: accelerator.register_for_checkpointing(A__ , A__ , A__ , A__ ) A__ : Any = str(ve.exception ) self.assertTrue("""Item at index 0""" in message ) self.assertTrue("""Item at index 1""" in message ) self.assertFalse("""Item at index 2""" in message ) self.assertFalse("""Item at index 3""" in message ) def __A ( self ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) A__ : Any = DummyModel() A__ : Union[str, Any] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) A__ : Dict = torch.optim.lr_scheduler.StepLR(A__ , step_size=1 , gamma=0.9_9 ) A__ , A__ : List[Any] = dummy_dataloaders() A__ : Tuple = ProjectConfiguration(automatic_checkpoint_naming=A__ ) # Train baseline A__ : Optional[Any] = Accelerator(project_dir=A__ , project_config=A__ ) A__ , A__ , A__ , A__ , A__ : Union[str, Any] = accelerator.prepare( A__ , A__ , A__ , A__ , A__ ) # Save initial accelerator.save_state() A__ : Tuple = scheduler.state_dict() train(3 , A__ , A__ , A__ , A__ , A__ ) self.assertNotEqual(A__ , scheduler.state_dict() ) # Load everything back in and make sure all states work accelerator.load_state(os.path.join(A__ , """checkpoints""" , """checkpoint_0""" ) ) self.assertEqual(A__ , scheduler.state_dict() ) def __A ( self ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) A__ : Optional[Any] = DummyModel() A__ : int = ProjectConfiguration(automatic_checkpoint_naming=A__ , total_limit=2 ) # Train baseline A__ : List[str] = Accelerator(project_dir=A__ , project_config=A__ ) A__ : Union[str, Any] = accelerator.prepare(A__ ) # Save 3 states: for _ in range(11 ): accelerator.save_state() self.assertTrue(not os.path.exists(os.path.join(A__ , """checkpoints""" , """checkpoint_0""" ) ) ) self.assertTrue(os.path.exists(os.path.join(A__ , """checkpoints""" , """checkpoint_9""" ) ) ) self.assertTrue(os.path.exists(os.path.join(A__ , """checkpoints""" , """checkpoint_10""" ) ) ) @require_cuda def __A ( self ): A__ : Dict = ["""torchrun""", F"""--nproc_per_node={torch.cuda.device_count()}""", inspect.getfile(self.__class__ )] execute_subprocess_async(A__ , env=os.environ.copy() ) if __name__ == "__main__": A_ : List[str] = '/tmp/accelerate/state_checkpointing' A_ : Optional[Any] = DummyModel() A_ : Union[str, Any] = torch.optim.Adam(params=model.parameters(), lr=1E-3) A_ : str = torch.optim.lr_scheduler.StepLR(optimizer, step_size=1, gamma=0.99) A_ , A_ : List[Any] = dummy_dataloaders() A_ : int = ProjectConfiguration(automatic_checkpoint_naming=True) # Train baseline A_ : List[str] = Accelerator(project_dir=savedir, project_config=project_config, mixed_precision='no') if accelerator.process_index == 0: if os.path.exists(savedir): shutil.rmtree(savedir) os.makedirs(savedir) A_ , A_ , A_ , A_ , A_ : List[Any] = accelerator.prepare( model, optimizer, train_dataloader, valid_dataloader, scheduler ) A_ , A_ : Dict = accelerator.prepare(model, optimizer) train(3, model, train_dataloader, optimizer, accelerator, scheduler) # Check that the intial optimizer is loaded on the GPU for group in optimizer.param_groups: A_ : str = group['params'][0].device break assert param_device.type == accelerator.device.type A_ : Optional[Any] = model.cpu() accelerator.wait_for_everyone() accelerator.save_state() accelerator.wait_for_everyone() # Check CPU state accelerator.load_state(os.path.join(savedir, 'checkpoints', 'checkpoint_0'), map_location='cpu') for group in optimizer.param_groups: A_ : str = group['params'][0].device break assert ( param_device.type == torch.device('cpu').type ), f"Loaded optimizer states did not match, expected to be loaded on the CPU but got {param_device}" # Check device state model.to(accelerator.device) accelerator.load_state(os.path.join(savedir, 'checkpoints', 'checkpoint_0'), map_location='on_device') for group in optimizer.param_groups: A_ : Tuple = group['params'][0].device break assert ( param_device.type == accelerator.device.type ), f"Loaded optimizer states did not match, expected to be loaded on {accelerator.device} but got {param_device}" # Check error with pytest.raises(TypeError, match='Unsupported optimizer map location passed'): accelerator.load_state(os.path.join(savedir, 'checkpoints', 'checkpoint_0'), map_location='invalid') accelerator.wait_for_everyone() if accelerator.process_index == 0: shutil.rmtree(savedir) accelerator.wait_for_everyone()
64
1
from math import factorial def UpperCamelCase (lowercase_: int , lowercase_: int ) -> int: # If either of the conditions are true, the function is being asked # to calculate a factorial of a negative number, which is not possible if n < k or k < 0: raise ValueError("""Please enter positive integers for n and k where n >= k""" ) return factorial(lowercase_ ) // (factorial(lowercase_ ) * factorial(n - k )) if __name__ == "__main__": print( 'The number of five-card hands possible from a standard', f'''fifty-two card deck is: {combinations(52, 5)}\n''', ) print( 'If a class of 40 students must be arranged into groups of', f'''4 for group projects, there are {combinations(40, 4)} ways''', 'to arrange them.\n', ) print( 'If 10 teams are competing in a Formula One race, there', f'''are {combinations(10, 3)} ways that first, second and''', 'third place can be awarded.', )
64
def UpperCamelCase (lowercase_: str , lowercase_: str ) -> bool: A__ : Union[str, Any] = len(lowercase_ ) A__ : List[Any] = len(lowercase_ ) A__ : List[Any] = [[False for _ in range(m + 1 )] for _ in range(n + 1 )] A__ : str = True for i in range(lowercase_ ): for j in range(m + 1 ): if dp[i][j]: if j < m and a[i].upper() == b[j]: A__ : int = True if a[i].islower(): A__ : Dict = True return dp[n][m] if __name__ == "__main__": import doctest doctest.testmod()
64
1