code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" import copy import os from collections import OrderedDict from typing import TYPE_CHECKING, Any, Dict, Mapping, Optional, Union if TYPE_CHECKING: from ...processing_utils import ProcessorMixin from ...utils import TensorType from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __snake_case : List[str] = logging.get_logger(__name__) __snake_case : Optional[int] = { 'google/owlvit-base-patch32': 'https://huggingface.co/google/owlvit-base-patch32/resolve/main/config.json', 'google/owlvit-base-patch16': 'https://huggingface.co/google/owlvit-base-patch16/resolve/main/config.json', 'google/owlvit-large-patch14': 'https://huggingface.co/google/owlvit-large-patch14/resolve/main/config.json', } class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'owlvit_text_model' def __init__( self: Tuple , _SCREAMING_SNAKE_CASE: Optional[int]=4_9408 , _SCREAMING_SNAKE_CASE: Dict=512 , _SCREAMING_SNAKE_CASE: List[Any]=2048 , _SCREAMING_SNAKE_CASE: int=12 , _SCREAMING_SNAKE_CASE: Any=8 , _SCREAMING_SNAKE_CASE: Optional[Any]=16 , _SCREAMING_SNAKE_CASE: Tuple="quick_gelu" , _SCREAMING_SNAKE_CASE: str=1e-5 , _SCREAMING_SNAKE_CASE: Any=0.0 , _SCREAMING_SNAKE_CASE: int=0.02 , _SCREAMING_SNAKE_CASE: int=1.0 , _SCREAMING_SNAKE_CASE: Dict=0 , _SCREAMING_SNAKE_CASE: Optional[int]=4_9406 , _SCREAMING_SNAKE_CASE: str=4_9407 , **_SCREAMING_SNAKE_CASE: int , ) -> List[str]: """simple docstring""" super().__init__(pad_token_id=_SCREAMING_SNAKE_CASE , bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = vocab_size __lowerCAmelCase : Optional[Any] = hidden_size __lowerCAmelCase : Tuple = intermediate_size __lowerCAmelCase : int = num_hidden_layers __lowerCAmelCase : Tuple = num_attention_heads __lowerCAmelCase : Any = max_position_embeddings __lowerCAmelCase : Optional[int] = hidden_act __lowerCAmelCase : Dict = layer_norm_eps __lowerCAmelCase : Tuple = attention_dropout __lowerCAmelCase : List[str] = initializer_range __lowerCAmelCase : Optional[Any] = initializer_factor @classmethod def _SCREAMING_SNAKE_CASE ( cls: Optional[int] , _SCREAMING_SNAKE_CASE: Union[str, os.PathLike] , **_SCREAMING_SNAKE_CASE: Tuple) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(_SCREAMING_SNAKE_CASE) __lowerCAmelCase , __lowerCAmelCase : Dict = cls.get_config_dict(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE) # get the text config dict if we are loading from OwlViTConfig if config_dict.get("model_type") == "owlvit": __lowerCAmelCase : str = config_dict["text_config"] if "model_type" in config_dict and hasattr(cls , "model_type") and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict['model_type']} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""") return cls.from_dict(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'owlvit_vision_model' def __init__( self: Any , _SCREAMING_SNAKE_CASE: Tuple=768 , _SCREAMING_SNAKE_CASE: List[str]=3072 , _SCREAMING_SNAKE_CASE: List[str]=12 , _SCREAMING_SNAKE_CASE: Tuple=12 , _SCREAMING_SNAKE_CASE: Dict=3 , _SCREAMING_SNAKE_CASE: int=768 , _SCREAMING_SNAKE_CASE: Optional[int]=32 , _SCREAMING_SNAKE_CASE: Union[str, Any]="quick_gelu" , _SCREAMING_SNAKE_CASE: List[str]=1e-5 , _SCREAMING_SNAKE_CASE: int=0.0 , _SCREAMING_SNAKE_CASE: Tuple=0.02 , _SCREAMING_SNAKE_CASE: Union[str, Any]=1.0 , **_SCREAMING_SNAKE_CASE: Optional[int] , ) -> Union[str, Any]: """simple docstring""" super().__init__(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = hidden_size __lowerCAmelCase : Union[str, Any] = intermediate_size __lowerCAmelCase : List[str] = num_hidden_layers __lowerCAmelCase : List[Any] = num_attention_heads __lowerCAmelCase : Optional[int] = num_channels __lowerCAmelCase : Any = image_size __lowerCAmelCase : Optional[Any] = patch_size __lowerCAmelCase : Optional[int] = hidden_act __lowerCAmelCase : Tuple = layer_norm_eps __lowerCAmelCase : List[Any] = attention_dropout __lowerCAmelCase : List[Any] = initializer_range __lowerCAmelCase : List[str] = initializer_factor @classmethod def _SCREAMING_SNAKE_CASE ( cls: Tuple , _SCREAMING_SNAKE_CASE: Union[str, os.PathLike] , **_SCREAMING_SNAKE_CASE: List[Any]) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(_SCREAMING_SNAKE_CASE) __lowerCAmelCase , __lowerCAmelCase : Union[str, Any] = cls.get_config_dict(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE) # get the vision config dict if we are loading from OwlViTConfig if config_dict.get("model_type") == "owlvit": __lowerCAmelCase : Tuple = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type") and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict['model_type']} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""") return cls.from_dict(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'owlvit' SCREAMING_SNAKE_CASE = True def __init__( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: Dict=None , _SCREAMING_SNAKE_CASE: List[Any]=None , _SCREAMING_SNAKE_CASE: Tuple=512 , _SCREAMING_SNAKE_CASE: Optional[Any]=2.6592 , _SCREAMING_SNAKE_CASE: Union[str, Any]=True , **_SCREAMING_SNAKE_CASE: Tuple , ) -> int: """simple docstring""" super().__init__(**_SCREAMING_SNAKE_CASE) if text_config is None: __lowerCAmelCase : Union[str, Any] = {} logger.info("text_config is None. Initializing the OwlViTTextConfig with default values.") if vision_config is None: __lowerCAmelCase : Optional[int] = {} logger.info("vision_config is None. initializing the OwlViTVisionConfig with default values.") __lowerCAmelCase : str = OwlViTTextConfig(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = OwlViTVisionConfig(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = projection_dim __lowerCAmelCase : Dict = logit_scale_init_value __lowerCAmelCase : Any = return_dict __lowerCAmelCase : Union[str, Any] = 1.0 @classmethod def _SCREAMING_SNAKE_CASE ( cls: str , _SCREAMING_SNAKE_CASE: Union[str, os.PathLike] , **_SCREAMING_SNAKE_CASE: List[str]) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(_SCREAMING_SNAKE_CASE) __lowerCAmelCase , __lowerCAmelCase : Tuple = cls.get_config_dict(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE) if "model_type" in config_dict and hasattr(cls , "model_type") and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict['model_type']} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""") return cls.from_dict(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE) @classmethod def _SCREAMING_SNAKE_CASE ( cls: List[str] , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Dict , **_SCREAMING_SNAKE_CASE: Optional[int]) -> Any: """simple docstring""" __lowerCAmelCase : Union[str, Any] = {} __lowerCAmelCase : Any = text_config __lowerCAmelCase : Dict = vision_config return cls.from_dict(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Any) -> Optional[int]: """simple docstring""" __lowerCAmelCase : Optional[int] = copy.deepcopy(self.__dict__) __lowerCAmelCase : str = self.text_config.to_dict() __lowerCAmelCase : int = self.vision_config.to_dict() __lowerCAmelCase : Dict = self.__class__.model_type return output class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' @property def _SCREAMING_SNAKE_CASE ( self: Dict) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ("input_ids", {0: "batch", 1: "sequence"}), ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ("attention_mask", {0: "batch", 1: "sequence"}), ]) @property def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ("logits_per_image", {0: "batch"}), ("logits_per_text", {0: "batch"}), ("text_embeds", {0: "batch"}), ("image_embeds", {0: "batch"}), ]) @property def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> float: """simple docstring""" return 1e-4 def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: "ProcessorMixin" , _SCREAMING_SNAKE_CASE: int = -1 , _SCREAMING_SNAKE_CASE: int = -1 , _SCREAMING_SNAKE_CASE: Optional["TensorType"] = None , ) -> Mapping[str, Any]: """simple docstring""" __lowerCAmelCase : str = super().generate_dummy_inputs( processor.tokenizer , batch_size=_SCREAMING_SNAKE_CASE , seq_length=_SCREAMING_SNAKE_CASE , framework=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = super().generate_dummy_inputs( processor.image_processor , batch_size=_SCREAMING_SNAKE_CASE , framework=_SCREAMING_SNAKE_CASE) return {**text_input_dict, **image_input_dict} @property def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> int: """simple docstring""" return 14
269
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxSeqaSeqConfigWithPast from ...utils import logging if TYPE_CHECKING: from ...feature_extraction_utils import FeatureExtractionMixin from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType __snake_case : Optional[Any] = logging.get_logger(__name__) __snake_case : Optional[Any] = { 'openai/whisper-base': 'https://huggingface.co/openai/whisper-base/resolve/main/config.json', } # fmt: off __snake_case : Any = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 357, 366, 438, 532, 685, 705, 796, 930, 1_058, 1_220, 1_267, 1_279, 1_303, 1_343, 1_377, 1_391, 1_635, 1_782, 1_875, 2_162, 2_361, 2_488, 3_467, 4_008, 4_211, 4_600, 4_808, 5_299, 5_855, 6_329, 7_203, 9_609, 9_959, 10_563, 10_786, 11_420, 11_709, 11_907, 13_163, 13_697, 13_700, 14_808, 15_306, 16_410, 16_791, 17_992, 19_203, 19_510, 20_724, 22_305, 22_935, 27_007, 30_109, 30_420, 33_409, 34_949, 40_283, 40_493, 40_549, 47_282, 49_146, 50_257, 50_359, 50_360, 50_361 ] __snake_case : str = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 359, 503, 522, 542, 873, 893, 902, 918, 922, 931, 1_350, 1_853, 1_982, 2_460, 2_627, 3_246, 3_253, 3_268, 3_536, 3_846, 3_961, 4_183, 4_667, 6_585, 6_647, 7_273, 9_061, 9_383, 10_428, 10_929, 11_938, 12_033, 12_331, 12_562, 13_793, 14_157, 14_635, 15_265, 15_618, 16_553, 16_604, 18_362, 18_956, 20_075, 21_675, 22_520, 26_130, 26_161, 26_435, 28_279, 29_464, 31_650, 32_302, 32_470, 36_865, 42_863, 47_425, 49_870, 50_254, 50_258, 50_360, 50_361, 50_362 ] class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'whisper' SCREAMING_SNAKE_CASE = ['past_key_values'] SCREAMING_SNAKE_CASE = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self: Optional[int] , _SCREAMING_SNAKE_CASE: Any=5_1865 , _SCREAMING_SNAKE_CASE: Optional[Any]=80 , _SCREAMING_SNAKE_CASE: Optional[int]=6 , _SCREAMING_SNAKE_CASE: Any=4 , _SCREAMING_SNAKE_CASE: Dict=6 , _SCREAMING_SNAKE_CASE: Dict=4 , _SCREAMING_SNAKE_CASE: Optional[Any]=1536 , _SCREAMING_SNAKE_CASE: List[str]=1536 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.0 , _SCREAMING_SNAKE_CASE: str=0.0 , _SCREAMING_SNAKE_CASE: List[str]=5_0257 , _SCREAMING_SNAKE_CASE: Optional[Any]=True , _SCREAMING_SNAKE_CASE: List[str]=True , _SCREAMING_SNAKE_CASE: Optional[int]="gelu" , _SCREAMING_SNAKE_CASE: Tuple=256 , _SCREAMING_SNAKE_CASE: str=0.0 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.0 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.0 , _SCREAMING_SNAKE_CASE: List[Any]=0.02 , _SCREAMING_SNAKE_CASE: Tuple=False , _SCREAMING_SNAKE_CASE: Union[str, Any]=1500 , _SCREAMING_SNAKE_CASE: str=448 , _SCREAMING_SNAKE_CASE: Any=5_0256 , _SCREAMING_SNAKE_CASE: Any=5_0256 , _SCREAMING_SNAKE_CASE: List[str]=5_0256 , _SCREAMING_SNAKE_CASE: Dict=None , _SCREAMING_SNAKE_CASE: List[Any]=[220, 5_0256] , _SCREAMING_SNAKE_CASE: Dict=False , _SCREAMING_SNAKE_CASE: str=256 , _SCREAMING_SNAKE_CASE: List[str]=False , _SCREAMING_SNAKE_CASE: Tuple=0.05 , _SCREAMING_SNAKE_CASE: List[str]=10 , _SCREAMING_SNAKE_CASE: str=2 , _SCREAMING_SNAKE_CASE: Any=0.0 , _SCREAMING_SNAKE_CASE: Optional[int]=10 , _SCREAMING_SNAKE_CASE: int=0 , _SCREAMING_SNAKE_CASE: Any=7 , **_SCREAMING_SNAKE_CASE: List[str] , ) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[int] = vocab_size __lowerCAmelCase : Optional[Any] = num_mel_bins __lowerCAmelCase : int = d_model __lowerCAmelCase : List[Any] = encoder_layers __lowerCAmelCase : List[Any] = encoder_attention_heads __lowerCAmelCase : List[str] = decoder_layers __lowerCAmelCase : Tuple = decoder_attention_heads __lowerCAmelCase : Any = decoder_ffn_dim __lowerCAmelCase : Tuple = encoder_ffn_dim __lowerCAmelCase : List[str] = dropout __lowerCAmelCase : Union[str, Any] = attention_dropout __lowerCAmelCase : Union[str, Any] = activation_dropout __lowerCAmelCase : Dict = activation_function __lowerCAmelCase : Tuple = init_std __lowerCAmelCase : str = encoder_layerdrop __lowerCAmelCase : int = decoder_layerdrop __lowerCAmelCase : Optional[int] = use_cache __lowerCAmelCase : Union[str, Any] = encoder_layers __lowerCAmelCase : List[str] = scale_embedding # scale factor will be sqrt(d_model) if True __lowerCAmelCase : int = max_source_positions __lowerCAmelCase : Any = max_target_positions # Audio Classification-specific parameters. Feel free to ignore for other classes. __lowerCAmelCase : Dict = classifier_proj_size __lowerCAmelCase : Dict = use_weighted_layer_sum # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __lowerCAmelCase : int = apply_spec_augment __lowerCAmelCase : Union[str, Any] = mask_time_prob __lowerCAmelCase : str = mask_time_length __lowerCAmelCase : int = mask_time_min_masks __lowerCAmelCase : List[Any] = mask_feature_prob __lowerCAmelCase : Tuple = mask_feature_length __lowerCAmelCase : Any = mask_feature_min_masks __lowerCAmelCase : Union[str, Any] = median_filter_width super().__init__( pad_token_id=_SCREAMING_SNAKE_CASE , bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE , is_encoder_decoder=_SCREAMING_SNAKE_CASE , decoder_start_token_id=_SCREAMING_SNAKE_CASE , suppress_tokens=_SCREAMING_SNAKE_CASE , begin_suppress_tokens=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' @property def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> Mapping[str, Mapping[int, str]]: """simple docstring""" __lowerCAmelCase : List[str] = OrderedDict( [ ("input_features", {0: "batch", 1: "feature_size", 2: "encoder_sequence"}), ]) if self.use_past: __lowerCAmelCase : Tuple = {0: "batch"} else: __lowerCAmelCase : Union[str, Any] = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(_SCREAMING_SNAKE_CASE , direction="inputs") return common_inputs def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: Union["PreTrainedTokenizerBase", "FeatureExtractionMixin"] , _SCREAMING_SNAKE_CASE: int = -1 , _SCREAMING_SNAKE_CASE: int = -1 , _SCREAMING_SNAKE_CASE: bool = False , _SCREAMING_SNAKE_CASE: Optional["TensorType"] = None , _SCREAMING_SNAKE_CASE: int = 2_2050 , _SCREAMING_SNAKE_CASE: float = 5.0 , _SCREAMING_SNAKE_CASE: int = 220 , ) -> Mapping[str, Any]: """simple docstring""" __lowerCAmelCase : int = OrderedDict() __lowerCAmelCase : Optional[Any] = OnnxConfig.generate_dummy_inputs( self , preprocessor=preprocessor.feature_extractor , batch_size=_SCREAMING_SNAKE_CASE , framework=_SCREAMING_SNAKE_CASE , sampling_rate=_SCREAMING_SNAKE_CASE , time_duration=_SCREAMING_SNAKE_CASE , frequency=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : List[str] = encoder_inputs["input_features"].shape[2] __lowerCAmelCase : List[str] = encoder_sequence_length // 2 if self.use_past else seq_length __lowerCAmelCase : List[Any] = super().generate_dummy_inputs( preprocessor.tokenizer , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = encoder_inputs.pop("input_features") __lowerCAmelCase : List[Any] = decoder_inputs.pop("decoder_input_ids") if "past_key_values" in decoder_inputs: __lowerCAmelCase : int = decoder_inputs.pop("past_key_values") return dummy_inputs @property def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> float: """simple docstring""" return 1e-3
269
1
"""simple docstring""" def _lowercase ( __snake_case = 1_000_000 ) -> int: __lowerCAmelCase : Union[str, Any] = [i - 1 for i in range(limit + 1 )] for i in range(2 ,limit + 1 ): if phi[i] == i - 1: for j in range(2 * i ,limit + 1 ,__snake_case ): phi[j] -= phi[j] // i return sum(phi[2 : limit + 1] ) if __name__ == "__main__": print(solution())
269
"""simple docstring""" import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration __snake_case : Optional[int] = 50_000 __snake_case : Dict = 5_000 __snake_case , __snake_case : Union[str, Any] = os.path.split(__file__) __snake_case : Any = os.path.join(RESULTS_BASEPATH, 'results', RESULTS_FILENAME.replace('.py', '.json')) @get_duration def _lowercase ( __snake_case ,__snake_case ) -> Dict: for i in range(__snake_case ): __lowerCAmelCase : Union[str, Any] = dataset[i] @get_duration def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> Dict: for i in range(0 ,len(__snake_case ) ,__snake_case ): __lowerCAmelCase : List[str] = dataset[i : i + batch_size] @get_duration def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> Dict: with dataset.formatted_as(type=__snake_case ): for i in range(__snake_case ): __lowerCAmelCase : Union[str, Any] = dataset[i] @get_duration def _lowercase ( __snake_case ,__snake_case ,__snake_case ,__snake_case ) -> str: with dataset.formatted_as(type=__snake_case ): for i in range(0 ,__snake_case ,__snake_case ): __lowerCAmelCase : Optional[int] = dataset[i : i + batch_size] def _lowercase ( ) -> Union[str, Any]: __lowerCAmelCase : Optional[int] = {"num examples": SPEED_TEST_N_EXAMPLES} __lowerCAmelCase : Optional[int] = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted, {"type": "pandas", "length": SMALL_TEST}), (read_formatted, {"type": "torch", "length": SMALL_TEST}), (read_formatted, {"type": "tensorflow", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] __lowerCAmelCase : Any = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print("generating dataset" ) __lowerCAmelCase : int = datasets.Features( {"list": datasets.Sequence(datasets.Value("float32" ) ), "numbers": datasets.Value("float32" )} ) __lowerCAmelCase : str = generate_example_dataset( os.path.join(__snake_case ,"dataset.arrow" ) ,__snake_case ,num_examples=__snake_case ,seq_shapes={"list": (100,)} ,) print("first set of iterations" ) for func, kwargs in functions: print(func.__name__ ,str(__snake_case ) ) __lowerCAmelCase : str = func(__snake_case ,**__snake_case ) print("shuffling dataset" ) __lowerCAmelCase : Optional[int] = dataset.shuffle() print("Second set of iterations (after shuffling" ) for func, kwargs in functions_shuffled: print("shuffled " ,func.__name__ ,str(__snake_case ) ) __lowerCAmelCase : List[str] = func( __snake_case ,**__snake_case ) with open(__snake_case ,"wb" ) as f: f.write(json.dumps(__snake_case ).encode("utf-8" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
269
1
"""simple docstring""" from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .embeddings import GaussianFourierProjection, TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin from .unet_ad_blocks import get_down_block, get_mid_block, get_out_block, get_up_block @dataclass class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 42 class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): '''simple docstring''' @register_to_config def __init__( self: Dict , _SCREAMING_SNAKE_CASE: int = 6_5536 , _SCREAMING_SNAKE_CASE: Optional[int] = None , _SCREAMING_SNAKE_CASE: int = 2 , _SCREAMING_SNAKE_CASE: int = 2 , _SCREAMING_SNAKE_CASE: int = 0 , _SCREAMING_SNAKE_CASE: str = "fourier" , _SCREAMING_SNAKE_CASE: bool = True , _SCREAMING_SNAKE_CASE: bool = False , _SCREAMING_SNAKE_CASE: float = 0.0 , _SCREAMING_SNAKE_CASE: Tuple[str] = ("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , _SCREAMING_SNAKE_CASE: Tuple[str] = ("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , _SCREAMING_SNAKE_CASE: Tuple[str] = "UNetMidBlock1D" , _SCREAMING_SNAKE_CASE: str = None , _SCREAMING_SNAKE_CASE: Tuple[int] = (32, 32, 64) , _SCREAMING_SNAKE_CASE: str = None , _SCREAMING_SNAKE_CASE: int = 8 , _SCREAMING_SNAKE_CASE: int = 1 , _SCREAMING_SNAKE_CASE: bool = False , ) -> int: """simple docstring""" super().__init__() __lowerCAmelCase : Any = sample_size # time if time_embedding_type == "fourier": __lowerCAmelCase : Union[str, Any] = GaussianFourierProjection( embedding_size=8 , set_W_to_weight=_SCREAMING_SNAKE_CASE , log=_SCREAMING_SNAKE_CASE , flip_sin_to_cos=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = 2 * block_out_channels[0] elif time_embedding_type == "positional": __lowerCAmelCase : List[Any] = Timesteps( block_out_channels[0] , flip_sin_to_cos=_SCREAMING_SNAKE_CASE , downscale_freq_shift=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = block_out_channels[0] if use_timestep_embedding: __lowerCAmelCase : str = block_out_channels[0] * 4 __lowerCAmelCase : Union[str, Any] = TimestepEmbedding( in_channels=_SCREAMING_SNAKE_CASE , time_embed_dim=_SCREAMING_SNAKE_CASE , act_fn=_SCREAMING_SNAKE_CASE , out_dim=block_out_channels[0] , ) __lowerCAmelCase : int = nn.ModuleList([]) __lowerCAmelCase : Any = None __lowerCAmelCase : Any = nn.ModuleList([]) __lowerCAmelCase : Tuple = None # down __lowerCAmelCase : List[str] = in_channels for i, down_block_type in enumerate(_SCREAMING_SNAKE_CASE): __lowerCAmelCase : Optional[Any] = output_channel __lowerCAmelCase : List[str] = block_out_channels[i] if i == 0: input_channel += extra_in_channels __lowerCAmelCase : Optional[Any] = i == len(_SCREAMING_SNAKE_CASE) - 1 __lowerCAmelCase : Any = get_down_block( _SCREAMING_SNAKE_CASE , num_layers=_SCREAMING_SNAKE_CASE , in_channels=_SCREAMING_SNAKE_CASE , out_channels=_SCREAMING_SNAKE_CASE , temb_channels=block_out_channels[0] , add_downsample=not is_final_block or downsample_each_block , ) self.down_blocks.append(_SCREAMING_SNAKE_CASE) # mid __lowerCAmelCase : Optional[int] = get_mid_block( _SCREAMING_SNAKE_CASE , in_channels=block_out_channels[-1] , mid_channels=block_out_channels[-1] , out_channels=block_out_channels[-1] , embed_dim=block_out_channels[0] , num_layers=_SCREAMING_SNAKE_CASE , add_downsample=_SCREAMING_SNAKE_CASE , ) # up __lowerCAmelCase : Optional[int] = list(reversed(_SCREAMING_SNAKE_CASE)) __lowerCAmelCase : str = reversed_block_out_channels[0] if out_block_type is None: __lowerCAmelCase : List[Any] = out_channels else: __lowerCAmelCase : Dict = block_out_channels[0] for i, up_block_type in enumerate(_SCREAMING_SNAKE_CASE): __lowerCAmelCase : Tuple = output_channel __lowerCAmelCase : str = ( reversed_block_out_channels[i + 1] if i < len(_SCREAMING_SNAKE_CASE) - 1 else final_upsample_channels ) __lowerCAmelCase : str = i == len(_SCREAMING_SNAKE_CASE) - 1 __lowerCAmelCase : List[Any] = get_up_block( _SCREAMING_SNAKE_CASE , num_layers=_SCREAMING_SNAKE_CASE , in_channels=_SCREAMING_SNAKE_CASE , out_channels=_SCREAMING_SNAKE_CASE , temb_channels=block_out_channels[0] , add_upsample=not is_final_block , ) self.up_blocks.append(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = output_channel # out __lowerCAmelCase : Any = norm_num_groups if norm_num_groups is not None else min(block_out_channels[0] // 4 , 32) __lowerCAmelCase : List[str] = get_out_block( out_block_type=_SCREAMING_SNAKE_CASE , num_groups_out=_SCREAMING_SNAKE_CASE , embed_dim=block_out_channels[0] , out_channels=_SCREAMING_SNAKE_CASE , act_fn=_SCREAMING_SNAKE_CASE , fc_dim=block_out_channels[-1] // 4 , ) def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: torch.FloatTensor , _SCREAMING_SNAKE_CASE: Union[torch.Tensor, float, int] , _SCREAMING_SNAKE_CASE: bool = True , ) -> Union[UNetaDOutput, Tuple]: """simple docstring""" __lowerCAmelCase : Any = timestep if not torch.is_tensor(_SCREAMING_SNAKE_CASE): __lowerCAmelCase : Dict = torch.tensor([timesteps] , dtype=torch.long , device=sample.device) elif torch.is_tensor(_SCREAMING_SNAKE_CASE) and len(timesteps.shape) == 0: __lowerCAmelCase : Any = timesteps[None].to(sample.device) __lowerCAmelCase : int = self.time_proj(_SCREAMING_SNAKE_CASE) if self.config.use_timestep_embedding: __lowerCAmelCase : Optional[Any] = self.time_mlp(_SCREAMING_SNAKE_CASE) else: __lowerCAmelCase : int = timestep_embed[..., None] __lowerCAmelCase : Dict = timestep_embed.repeat([1, 1, sample.shape[2]]).to(sample.dtype) __lowerCAmelCase : Dict = timestep_embed.broadcast_to((sample.shape[:1] + timestep_embed.shape[1:])) # 2. down __lowerCAmelCase : Tuple = () for downsample_block in self.down_blocks: __lowerCAmelCase , __lowerCAmelCase : str = downsample_block(hidden_states=_SCREAMING_SNAKE_CASE , temb=_SCREAMING_SNAKE_CASE) down_block_res_samples += res_samples # 3. mid if self.mid_block: __lowerCAmelCase : Dict = self.mid_block(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) # 4. up for i, upsample_block in enumerate(self.up_blocks): __lowerCAmelCase : Optional[int] = down_block_res_samples[-1:] __lowerCAmelCase : Tuple = down_block_res_samples[:-1] __lowerCAmelCase : str = upsample_block(_SCREAMING_SNAKE_CASE , res_hidden_states_tuple=_SCREAMING_SNAKE_CASE , temb=_SCREAMING_SNAKE_CASE) # 5. post-process if self.out_block: __lowerCAmelCase : Optional[int] = self.out_block(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) if not return_dict: return (sample,) return UNetaDOutput(sample=_SCREAMING_SNAKE_CASE)
269
"""simple docstring""" import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self: Tuple , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: List[str]=13 , _SCREAMING_SNAKE_CASE: Tuple=7 , _SCREAMING_SNAKE_CASE: int=True , _SCREAMING_SNAKE_CASE: Optional[Any]=True , _SCREAMING_SNAKE_CASE: str=False , _SCREAMING_SNAKE_CASE: Optional[Any]=True , _SCREAMING_SNAKE_CASE: int=99 , _SCREAMING_SNAKE_CASE: int=32 , _SCREAMING_SNAKE_CASE: List[str]=5 , _SCREAMING_SNAKE_CASE: Union[str, Any]=4 , _SCREAMING_SNAKE_CASE: int=64 , _SCREAMING_SNAKE_CASE: List[str]="gelu" , _SCREAMING_SNAKE_CASE: str=0.1 , _SCREAMING_SNAKE_CASE: Any=0.1 , _SCREAMING_SNAKE_CASE: Optional[int]=512 , _SCREAMING_SNAKE_CASE: Tuple=16 , _SCREAMING_SNAKE_CASE: Any=2 , _SCREAMING_SNAKE_CASE: List[str]=0.02 , _SCREAMING_SNAKE_CASE: Tuple=3 , _SCREAMING_SNAKE_CASE: Optional[Any]=4 , _SCREAMING_SNAKE_CASE: int=None , _SCREAMING_SNAKE_CASE: int=2 , _SCREAMING_SNAKE_CASE: str=2 , _SCREAMING_SNAKE_CASE: Union[str, Any]=2 , _SCREAMING_SNAKE_CASE: List[Any]=2 , _SCREAMING_SNAKE_CASE: int=4 , _SCREAMING_SNAKE_CASE: List[str]=1 , ) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : List[str] = parent __lowerCAmelCase : Optional[Any] = batch_size __lowerCAmelCase : Union[str, Any] = seq_length __lowerCAmelCase : Optional[Any] = is_training __lowerCAmelCase : Optional[int] = use_input_mask __lowerCAmelCase : Dict = use_token_type_ids __lowerCAmelCase : Dict = use_labels __lowerCAmelCase : Dict = vocab_size __lowerCAmelCase : Tuple = hidden_size __lowerCAmelCase : List[Any] = num_hidden_layers __lowerCAmelCase : Union[str, Any] = num_attention_heads __lowerCAmelCase : Tuple = intermediate_size __lowerCAmelCase : List[Any] = hidden_act __lowerCAmelCase : Optional[Any] = hidden_dropout_prob __lowerCAmelCase : Optional[Any] = attention_probs_dropout_prob __lowerCAmelCase : Optional[int] = max_position_embeddings __lowerCAmelCase : Union[str, Any] = type_vocab_size __lowerCAmelCase : Optional[int] = type_sequence_label_size __lowerCAmelCase : Dict = initializer_range __lowerCAmelCase : Tuple = num_labels __lowerCAmelCase : Optional[Any] = num_choices __lowerCAmelCase : Union[str, Any] = scope __lowerCAmelCase : Optional[Any] = q_groups __lowerCAmelCase : Optional[int] = k_groups __lowerCAmelCase : Any = v_groups __lowerCAmelCase : int = post_attention_groups __lowerCAmelCase : List[str] = intermediate_groups __lowerCAmelCase : Optional[Any] = output_groups def _SCREAMING_SNAKE_CASE ( self: Dict) -> List[str]: """simple docstring""" __lowerCAmelCase : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) __lowerCAmelCase : Union[str, Any] = None if self.use_input_mask: __lowerCAmelCase : Tuple = random_attention_mask([self.batch_size, self.seq_length]) __lowerCAmelCase : Optional[int] = None __lowerCAmelCase : List[Any] = None __lowerCAmelCase : str = None if self.use_labels: __lowerCAmelCase : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size) __lowerCAmelCase : str = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) __lowerCAmelCase : str = ids_tensor([self.batch_size] , self.num_choices) __lowerCAmelCase : Any = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _SCREAMING_SNAKE_CASE ( self: Tuple) -> int: """simple docstring""" return SqueezeBertConfig( embedding_size=self.hidden_size , vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , attention_probs_dropout_prob=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , q_groups=self.q_groups , k_groups=self.k_groups , v_groups=self.v_groups , post_attention_groups=self.post_attention_groups , intermediate_groups=self.intermediate_groups , output_groups=self.output_groups , ) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Tuple , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Tuple) -> Tuple: """simple docstring""" __lowerCAmelCase : List[Any] = SqueezeBertModel(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : List[Any] = model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = model(_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Tuple) -> Dict: """simple docstring""" __lowerCAmelCase : int = SqueezeBertForMaskedLM(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : Dict = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def _SCREAMING_SNAKE_CASE ( self: Optional[int] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Union[str, Any]) -> int: """simple docstring""" __lowerCAmelCase : str = SqueezeBertForQuestionAnswering(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : Union[str, Any] = model( _SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , start_positions=_SCREAMING_SNAKE_CASE , end_positions=_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: Tuple , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: Tuple) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : List[Any] = self.num_labels __lowerCAmelCase : Union[str, Any] = SqueezeBertForSequenceClassification(_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : int = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[int]) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : Dict = self.num_labels __lowerCAmelCase : Optional[int] = SqueezeBertForTokenClassification(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : List[str] = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: int) -> Tuple: """simple docstring""" __lowerCAmelCase : List[str] = self.num_choices __lowerCAmelCase : str = SqueezeBertForMultipleChoice(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : int = input_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() __lowerCAmelCase : Union[str, Any] = input_mask.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() __lowerCAmelCase : str = model( _SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices)) def _SCREAMING_SNAKE_CASE ( self: str) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[int] = self.prepare_config_and_inputs() ((__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase)) : Union[str, Any] = config_and_inputs __lowerCAmelCase : int = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) SCREAMING_SNAKE_CASE = ( { 'feature-extraction': SqueezeBertModel, 'fill-mask': SqueezeBertForMaskedLM, 'question-answering': SqueezeBertForQuestionAnswering, 'text-classification': SqueezeBertForSequenceClassification, 'token-classification': SqueezeBertForTokenClassification, 'zero-shot': SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = False def _SCREAMING_SNAKE_CASE ( self: Dict) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Any = SqueezeBertModelTester(self) __lowerCAmelCase : Optional[int] = ConfigTester(self , config_class=_SCREAMING_SNAKE_CASE , dim=37) def _SCREAMING_SNAKE_CASE ( self: Any) -> Tuple: """simple docstring""" self.config_tester.run_common_tests() def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> Any: """simple docstring""" __lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> List[Any]: """simple docstring""" __lowerCAmelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> int: """simple docstring""" __lowerCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Any) -> int: """simple docstring""" __lowerCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> str: """simple docstring""" __lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*_SCREAMING_SNAKE_CASE) @slow def _SCREAMING_SNAKE_CASE ( self: Any) -> Dict: """simple docstring""" for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase : Optional[Any] = SqueezeBertModel.from_pretrained(_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) @require_sentencepiece @require_tokenizers @require_torch class A__ ( unittest.TestCase ): '''simple docstring''' @slow def _SCREAMING_SNAKE_CASE ( self: int) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[Any] = SqueezeBertForSequenceClassification.from_pretrained("squeezebert/squeezebert-mnli") __lowerCAmelCase : List[Any] = torch.tensor([[1, 2_9414, 232, 328, 740, 1140, 1_2695, 69, 13, 1588, 2]]) __lowerCAmelCase : List[Any] = model(_SCREAMING_SNAKE_CASE)[0] __lowerCAmelCase : Any = torch.Size((1, 3)) self.assertEqual(output.shape , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = torch.tensor([[0.6401, -0.0349, -0.6041]]) self.assertTrue(torch.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=1e-4))
269
1
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import ASTFeatureExtractor from transformers.testing_utils import require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin __snake_case : Tuple = random.Random() if is_torch_available(): import torch def _lowercase ( __snake_case ,__snake_case=1.0 ,__snake_case=None ,__snake_case=None ) -> Union[str, Any]: if rng is None: __lowerCAmelCase : List[str] = global_rng __lowerCAmelCase : int = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class A__ ( unittest.TestCase ): '''simple docstring''' def __init__( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Tuple=7 , _SCREAMING_SNAKE_CASE: Dict=400 , _SCREAMING_SNAKE_CASE: Optional[int]=2000 , _SCREAMING_SNAKE_CASE: Any=1 , _SCREAMING_SNAKE_CASE: List[str]=0.0 , _SCREAMING_SNAKE_CASE: Union[str, Any]=1_6000 , _SCREAMING_SNAKE_CASE: Tuple=True , _SCREAMING_SNAKE_CASE: Union[str, Any]=True , ) -> List[Any]: """simple docstring""" __lowerCAmelCase : Tuple = parent __lowerCAmelCase : Union[str, Any] = batch_size __lowerCAmelCase : Tuple = min_seq_length __lowerCAmelCase : Optional[int] = max_seq_length __lowerCAmelCase : List[Any] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __lowerCAmelCase : Any = feature_size __lowerCAmelCase : Optional[Any] = padding_value __lowerCAmelCase : str = sampling_rate __lowerCAmelCase : Optional[Any] = return_attention_mask __lowerCAmelCase : Dict = do_normalize def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> Dict: """simple docstring""" return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def _SCREAMING_SNAKE_CASE ( self: str , _SCREAMING_SNAKE_CASE: List[str]=False , _SCREAMING_SNAKE_CASE: int=False) -> Any: """simple docstring""" def _flatten(_SCREAMING_SNAKE_CASE: Union[str, Any]): return list(itertools.chain(*_SCREAMING_SNAKE_CASE)) if equal_length: __lowerCAmelCase : Any = floats_list((self.batch_size, self.max_seq_length)) else: # make sure that inputs increase in size __lowerCAmelCase : Union[str, Any] = [ _flatten(floats_list((x, self.feature_size))) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff) ] if numpify: __lowerCAmelCase : Tuple = [np.asarray(_SCREAMING_SNAKE_CASE) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = ASTFeatureExtractor def _SCREAMING_SNAKE_CASE ( self: int) -> List[str]: """simple docstring""" __lowerCAmelCase : Optional[Any] = ASTFeatureExtractionTester(self) def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> Optional[int]: """simple docstring""" __lowerCAmelCase : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) # create three inputs of length 800, 1000, and 1200 __lowerCAmelCase : str = [floats_list((1, x))[0] for x in range(800 , 1400 , 200)] __lowerCAmelCase : Optional[int] = [np.asarray(_SCREAMING_SNAKE_CASE) for speech_input in speech_inputs] # Test not batched input __lowerCAmelCase : Optional[Any] = feat_extract(speech_inputs[0] , return_tensors="np").input_values __lowerCAmelCase : str = feat_extract(np_speech_inputs[0] , return_tensors="np").input_values self.assertTrue(np.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=1e-3)) # Test batched __lowerCAmelCase : Optional[int] = feat_extract(_SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE , return_tensors="np").input_values __lowerCAmelCase : Any = feat_extract(_SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE , return_tensors="np").input_values for enc_seq_a, enc_seq_a in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE): self.assertTrue(np.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=1e-3)) # Test 2-D numpy arrays are batched. __lowerCAmelCase : List[Any] = [floats_list((1, x))[0] for x in (800, 800, 800)] __lowerCAmelCase : Optional[int] = np.asarray(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = feat_extract(_SCREAMING_SNAKE_CASE , return_tensors="np").input_values __lowerCAmelCase : Tuple = feat_extract(_SCREAMING_SNAKE_CASE , return_tensors="np").input_values for enc_seq_a, enc_seq_a in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE): self.assertTrue(np.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=1e-3)) @require_torch def _SCREAMING_SNAKE_CASE ( self: Dict) -> Tuple: """simple docstring""" import torch __lowerCAmelCase : Optional[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) __lowerCAmelCase : List[str] = np.random.rand(100).astype(np.floataa) __lowerCAmelCase : List[str] = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __lowerCAmelCase : int = feature_extractor.pad([{"input_values": inputs}] , return_tensors="np") self.assertTrue(np_processed.input_values.dtype == np.floataa) __lowerCAmelCase : Optional[int] = feature_extractor.pad([{"input_values": inputs}] , return_tensors="pt") self.assertTrue(pt_processed.input_values.dtype == torch.floataa) def _SCREAMING_SNAKE_CASE ( self: Dict , _SCREAMING_SNAKE_CASE: Optional[int]) -> int: """simple docstring""" from datasets import load_dataset __lowerCAmelCase : List[Any] = load_dataset("hf-internal-testing/librispeech_asr_dummy" , "clean" , split="validation") # automatic decoding with librispeech __lowerCAmelCase : int = ds.sort("id").select(range(_SCREAMING_SNAKE_CASE))[:num_samples]["audio"] return [x["array"] for x in speech_samples] @require_torch def _SCREAMING_SNAKE_CASE ( self: Dict) -> Dict: """simple docstring""" __lowerCAmelCase : Dict = torch.tensor( [-0.9894, -1.2776, -0.9066, -1.2776, -0.9349, -1.2609, -1.0386, -1.2776, -1.1561, -1.2776, -1.2052, -1.2723, -1.2190, -1.2132, -1.2776, -1.1133, -1.1953, -1.1343, -1.1584, -1.2203, -1.1770, -1.2474, -1.2381, -1.1936, -0.9270, -0.8317, -0.8049, -0.7706, -0.7565, -0.7869]) # fmt: on __lowerCAmelCase : Dict = self._load_datasamples(1) __lowerCAmelCase : Any = ASTFeatureExtractor() __lowerCAmelCase : int = feature_extractor(_SCREAMING_SNAKE_CASE , return_tensors="pt").input_values self.assertEquals(input_values.shape , (1, 1024, 128)) self.assertTrue(torch.allclose(input_values[0, 0, :30] , _SCREAMING_SNAKE_CASE , atol=1e-4))
269
"""simple docstring""" import os from math import logaa def _lowercase ( __snake_case = "base_exp.txt" ) -> int: __lowerCAmelCase : float = 0 __lowerCAmelCase : Any = 0 for i, line in enumerate(open(os.path.join(os.path.dirname(__snake_case ) ,__snake_case ) ) ): __lowerCAmelCase , __lowerCAmelCase : List[str] = list(map(__snake_case ,line.split("," ) ) ) if x * logaa(__snake_case ) > largest: __lowerCAmelCase : Tuple = x * logaa(__snake_case ) __lowerCAmelCase : Optional[Any] = i + 1 return result if __name__ == "__main__": print(solution())
269
1
"""simple docstring""" import argparse import os import sys from unittest.mock import patch import pytorch_lightning as pl import timeout_decorator import torch from distillation import SummarizationDistiller, distill_main from finetune import SummarizationModule, main from transformers import MarianMTModel from transformers.file_utils import cached_path from transformers.testing_utils import TestCasePlus, require_torch_gpu, slow from utils import load_json __snake_case : Any = 'sshleifer/mar_enro_6_3_student' class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> Any: """simple docstring""" super().setUp() __lowerCAmelCase : str = cached_path( "https://cdn-datasets.huggingface.co/translation/wmt_en_ro-tr40k-va0.5k-te0.5k.tar.gz" , extract_compressed_file=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : str = F"""{data_cached}/wmt_en_ro-tr40k-va0.5k-te0.5k""" @slow @require_torch_gpu def _SCREAMING_SNAKE_CASE ( self: str) -> Optional[int]: """simple docstring""" MarianMTModel.from_pretrained(_SCREAMING_SNAKE_CASE) @slow @require_torch_gpu def _SCREAMING_SNAKE_CASE ( self: Dict) -> List[Any]: """simple docstring""" __lowerCAmelCase : Tuple = { "$MAX_LEN": 64, "$BS": 64, "$GAS": 1, "$ENRO_DIR": self.data_dir, "facebook/mbart-large-cc25": MARIAN_MODEL, # "val_check_interval=0.25": "val_check_interval=1.0", "--learning_rate=3e-5": "--learning_rate 3e-4", "--num_train_epochs 6": "--num_train_epochs 1", } # Clean up bash script __lowerCAmelCase : List[str] = (self.test_file_dir / "train_mbart_cc25_enro.sh").open().read().split("finetune.py")[1].strip() __lowerCAmelCase : List[Any] = bash_script.replace("\\\n" , "").strip().replace("\"$@\"" , "") for k, v in env_vars_to_replace.items(): __lowerCAmelCase : Optional[int] = bash_script.replace(_SCREAMING_SNAKE_CASE , str(_SCREAMING_SNAKE_CASE)) __lowerCAmelCase : int = self.get_auto_remove_tmp_dir() # bash_script = bash_script.replace("--fp16 ", "") __lowerCAmelCase : Optional[Any] = F""" --output_dir {output_dir} --tokenizer_name Helsinki-NLP/opus-mt-en-ro --sortish_sampler --do_predict --gpus 1 --freeze_encoder --n_train 40000 --n_val 500 --n_test 500 --fp16_opt_level O1 --num_sanity_val_steps 0 --eval_beams 2 """.split() # XXX: args.gpus > 1 : handle multi_gpu in the future __lowerCAmelCase : str = ["finetune.py"] + bash_script.split() + args with patch.object(_SCREAMING_SNAKE_CASE , "argv" , _SCREAMING_SNAKE_CASE): __lowerCAmelCase : int = argparse.ArgumentParser() __lowerCAmelCase : List[str] = pl.Trainer.add_argparse_args(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = SummarizationModule.add_model_specific_args(_SCREAMING_SNAKE_CASE , os.getcwd()) __lowerCAmelCase : int = parser.parse_args() __lowerCAmelCase : Any = main(_SCREAMING_SNAKE_CASE) # Check metrics __lowerCAmelCase : Any = load_json(model.metrics_save_path) __lowerCAmelCase : Optional[Any] = metrics["val"][0] __lowerCAmelCase : str = metrics["val"][-1] self.assertEqual(len(metrics["val"]) , (args.max_epochs / args.val_check_interval)) assert isinstance(last_step_stats[F"""val_avg_{model.val_metric}"""] , _SCREAMING_SNAKE_CASE) self.assertGreater(last_step_stats["val_avg_gen_time"] , 0.01) # model hanging on generate. Maybe bad config was saved. (XXX: old comment/assert?) self.assertLessEqual(last_step_stats["val_avg_gen_time"] , 1.0) # test learning requirements: # 1. BLEU improves over the course of training by more than 2 pts self.assertGreater(last_step_stats["val_avg_bleu"] - first_step_stats["val_avg_bleu"] , 2) # 2. BLEU finishes above 17 self.assertGreater(last_step_stats["val_avg_bleu"] , 17) # 3. test BLEU and val BLEU within ~1.1 pt. self.assertLess(abs(metrics["val"][-1]["val_avg_bleu"] - metrics["test"][-1]["test_avg_bleu"]) , 1.1) # check lightning ckpt can be loaded and has a reasonable statedict __lowerCAmelCase : str = os.listdir(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = [x for x in contents if x.endswith(".ckpt")][0] __lowerCAmelCase : List[Any] = os.path.join(args.output_dir , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = torch.load(_SCREAMING_SNAKE_CASE , map_location="cpu") __lowerCAmelCase : Dict = "model.model.decoder.layers.0.encoder_attn_layer_norm.weight" assert expected_key in ckpt["state_dict"] assert ckpt["state_dict"]["model.model.decoder.layers.0.encoder_attn_layer_norm.weight"].dtype == torch.floataa # TODO: turn on args.do_predict when PL bug fixed. if args.do_predict: __lowerCAmelCase : int = {os.path.basename(_SCREAMING_SNAKE_CASE) for p in contents} assert "test_generations.txt" in contents assert "test_results.txt" in contents # assert len(metrics["val"]) == desired_n_evals assert len(metrics["test"]) == 1 class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' @timeout_decorator.timeout(600) @slow @require_torch_gpu def _SCREAMING_SNAKE_CASE ( self: Any) -> Tuple: """simple docstring""" __lowerCAmelCase : List[str] = F"""{self.test_file_dir_str}/test_data/wmt_en_ro""" __lowerCAmelCase : Any = { "--fp16_opt_level=O1": "", "$MAX_LEN": 128, "$BS": 16, "$GAS": 1, "$ENRO_DIR": data_dir, "$m": "sshleifer/student_marian_en_ro_6_1", "val_check_interval=0.25": "val_check_interval=1.0", } # Clean up bash script __lowerCAmelCase : List[Any] = ( (self.test_file_dir / "distil_marian_no_teacher.sh").open().read().split("distillation.py")[1].strip() ) __lowerCAmelCase : str = bash_script.replace("\\\n" , "").strip().replace("\"$@\"" , "") __lowerCAmelCase : Union[str, Any] = bash_script.replace("--fp16 " , " ") for k, v in env_vars_to_replace.items(): __lowerCAmelCase : Tuple = bash_script.replace(_SCREAMING_SNAKE_CASE , str(_SCREAMING_SNAKE_CASE)) __lowerCAmelCase : int = self.get_auto_remove_tmp_dir() __lowerCAmelCase : int = bash_script.replace("--fp16" , "") __lowerCAmelCase : List[Any] = 6 __lowerCAmelCase : int = ( ["distillation.py"] + bash_script.split() + [ F"""--output_dir={output_dir}""", "--gpus=1", "--learning_rate=1e-3", F"""--num_train_epochs={epochs}""", "--warmup_steps=10", "--val_check_interval=1.0", "--do_predict", ] ) with patch.object(_SCREAMING_SNAKE_CASE , "argv" , _SCREAMING_SNAKE_CASE): __lowerCAmelCase : Optional[Any] = argparse.ArgumentParser() __lowerCAmelCase : Optional[int] = pl.Trainer.add_argparse_args(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = SummarizationDistiller.add_model_specific_args(_SCREAMING_SNAKE_CASE , os.getcwd()) __lowerCAmelCase : List[Any] = parser.parse_args() # assert args.gpus == gpus THIS BREAKS for multi_gpu __lowerCAmelCase : List[str] = distill_main(_SCREAMING_SNAKE_CASE) # Check metrics __lowerCAmelCase : Union[str, Any] = load_json(model.metrics_save_path) __lowerCAmelCase : Any = metrics["val"][0] __lowerCAmelCase : Union[str, Any] = metrics["val"][-1] assert len(metrics["val"]) >= (args.max_epochs / args.val_check_interval) # +1 accounts for val_sanity_check assert last_step_stats["val_avg_gen_time"] >= 0.01 assert first_step_stats["val_avg_bleu"] < last_step_stats["val_avg_bleu"] # model learned nothing assert 1.0 >= last_step_stats["val_avg_gen_time"] # model hanging on generate. Maybe bad config was saved. assert isinstance(last_step_stats[F"""val_avg_{model.val_metric}"""] , _SCREAMING_SNAKE_CASE) # check lightning ckpt can be loaded and has a reasonable statedict __lowerCAmelCase : Any = os.listdir(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = [x for x in contents if x.endswith(".ckpt")][0] __lowerCAmelCase : Optional[Any] = os.path.join(args.output_dir , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = torch.load(_SCREAMING_SNAKE_CASE , map_location="cpu") __lowerCAmelCase : Dict = "model.model.decoder.layers.0.encoder_attn_layer_norm.weight" assert expected_key in ckpt["state_dict"] assert ckpt["state_dict"]["model.model.decoder.layers.0.encoder_attn_layer_norm.weight"].dtype == torch.floataa # TODO: turn on args.do_predict when PL bug fixed. if args.do_predict: __lowerCAmelCase : Dict = {os.path.basename(_SCREAMING_SNAKE_CASE) for p in contents} assert "test_generations.txt" in contents assert "test_results.txt" in contents # assert len(metrics["val"]) == desired_n_evals assert len(metrics["test"]) == 1
269
"""simple docstring""" import collections import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_flax_cross_test, require_flax, require_torch, require_vision, slow, torch_device, ) from transformers.utils import is_flax_available, is_torch_available, is_vision_available from ...test_modeling_flax_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_flax_bert import FlaxBertModelTester from ..clip.test_modeling_flax_clip import FlaxCLIPVisionModelTester from ..vit.test_modeling_flax_vit import FlaxViTModelTester if is_flax_available(): from transformers import ( FlaxBertModel, FlaxCLIPVisionModel, FlaxVisionTextDualEncoderModel, FlaxViTModel, VisionTextDualEncoderConfig, VisionTextDualEncoderProcessor, ) from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) if is_torch_available(): import torch from transformers import VisionTextDualEncoderModel if is_vision_available(): from PIL import Image def _lowercase ( __snake_case ) -> List[str]: if isinstance(__snake_case ,collections.abc.Iterable ): return x return (x, x) @require_flax class A__ : '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Optional[Any]) -> str: """simple docstring""" pass def _SCREAMING_SNAKE_CASE ( self: str) -> int: """simple docstring""" pass def _SCREAMING_SNAKE_CASE ( self: List[str]) -> Tuple: """simple docstring""" pass def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: np.ndarray , _SCREAMING_SNAKE_CASE: np.ndarray , _SCREAMING_SNAKE_CASE: float) -> List[Any]: """simple docstring""" __lowerCAmelCase : str = np.abs((a - b)).max() self.assertLessEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , F"""Difference between torch and flax is {diff} (>= {tol}).""") def _SCREAMING_SNAKE_CASE ( self: str , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Any=None , **_SCREAMING_SNAKE_CASE: Tuple) -> List[str]: """simple docstring""" __lowerCAmelCase : Tuple = VisionTextDualEncoderConfig.from_vision_text_configs(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = FlaxVisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = model(input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], config.projection_dim)) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], config.projection_dim)) def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Optional[Any]=None , **_SCREAMING_SNAKE_CASE: List[str]) -> Dict: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : List[Any] = self.get_vision_text_model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = {"vision_model": vision_model, "text_model": text_model} __lowerCAmelCase : Optional[Any] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = model(input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], model.config.projection_dim)) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], model.config.projection_dim)) def _SCREAMING_SNAKE_CASE ( self: Dict , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: List[Any]=None , **_SCREAMING_SNAKE_CASE: Tuple) -> str: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : Optional[Any] = self.get_vision_text_model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = {"vision_model": vision_model, "text_model": text_model} __lowerCAmelCase : Any = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = model(input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = output[0] with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = FlaxVisionTextDualEncoderModel.from_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = model(input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = after_output[0] __lowerCAmelCase : Any = np.amax(np.abs(out_a - out_a)) self.assertLessEqual(_SCREAMING_SNAKE_CASE , 1e-3) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Any=None , **_SCREAMING_SNAKE_CASE: List[str]) -> Optional[int]: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : Any = self.get_vision_text_model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = {"vision_model": vision_model, "text_model": text_model} __lowerCAmelCase : Optional[Any] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = model( input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , output_attentions=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = output.vision_model_output.attentions self.assertEqual(len(_SCREAMING_SNAKE_CASE) , vision_config.num_hidden_layers) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) __lowerCAmelCase : List[str] = to_atuple(vision_model.config.image_size) __lowerCAmelCase : Any = to_atuple(vision_model.config.patch_size) __lowerCAmelCase : Union[str, Any] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) __lowerCAmelCase : Tuple = num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len)) __lowerCAmelCase : Union[str, Any] = output.text_model_output.attentions self.assertEqual(len(_SCREAMING_SNAKE_CASE) , text_config.num_hidden_layers) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: int) -> str: """simple docstring""" pt_model.to(_SCREAMING_SNAKE_CASE) pt_model.eval() # prepare inputs __lowerCAmelCase : Union[str, Any] = inputs_dict __lowerCAmelCase : Union[str, Any] = {k: torch.tensor(v.tolist()) for k, v in flax_inputs.items()} with torch.no_grad(): __lowerCAmelCase : Any = pt_model(**_SCREAMING_SNAKE_CASE).to_tuple() __lowerCAmelCase : List[Any] = fx_model(**_SCREAMING_SNAKE_CASE).to_tuple() self.assertEqual(len(_SCREAMING_SNAKE_CASE) , len(_SCREAMING_SNAKE_CASE) , "Output lengths differ between Flax and PyTorch") for fx_output, pt_output in zip(fx_outputs[:4] , pt_outputs[:4]): self.assert_almost_equals(_SCREAMING_SNAKE_CASE , pt_output.numpy() , 4e-2) # PT -> Flax with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = FlaxVisionTextDualEncoderModel.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = fx_model_loaded(**_SCREAMING_SNAKE_CASE).to_tuple() self.assertEqual(len(_SCREAMING_SNAKE_CASE) , len(_SCREAMING_SNAKE_CASE) , "Output lengths differ between Flax and PyTorch") for fx_output_loaded, pt_output in zip(fx_outputs_loaded[:4] , pt_outputs[:4]): self.assert_almost_equals(_SCREAMING_SNAKE_CASE , pt_output.numpy() , 4e-2) # Flax -> PT with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = VisionTextDualEncoderModel.from_pretrained(_SCREAMING_SNAKE_CASE , from_flax=_SCREAMING_SNAKE_CASE) pt_model_loaded.to(_SCREAMING_SNAKE_CASE) pt_model_loaded.eval() with torch.no_grad(): __lowerCAmelCase : Optional[Any] = pt_model_loaded(**_SCREAMING_SNAKE_CASE).to_tuple() self.assertEqual(len(_SCREAMING_SNAKE_CASE) , len(_SCREAMING_SNAKE_CASE) , "Output lengths differ between Flax and PyTorch") for fx_output, pt_output_loaded in zip(fx_outputs[:4] , pt_outputs_loaded[:4]): self.assert_almost_equals(_SCREAMING_SNAKE_CASE , pt_output_loaded.numpy() , 4e-2) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Dict) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Union[str, Any] = VisionTextDualEncoderConfig.from_vision_text_configs(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = VisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = FlaxVisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = fx_state self.check_pt_flax_equivalence(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Tuple , _SCREAMING_SNAKE_CASE: Dict) -> str: """simple docstring""" __lowerCAmelCase : Optional[int] = VisionTextDualEncoderConfig.from_vision_text_configs(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = VisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = FlaxVisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = load_flax_weights_in_pytorch_model(_SCREAMING_SNAKE_CASE , fx_model.params) self.check_pt_flax_equivalence(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> str: """simple docstring""" __lowerCAmelCase : List[str] = self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Dict) -> int: """simple docstring""" __lowerCAmelCase : List[Any] = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> List[Any]: """simple docstring""" __lowerCAmelCase : Dict = self.prepare_config_and_inputs() self.check_save_load(**_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: int) -> Dict: """simple docstring""" __lowerCAmelCase : List[str] = self.prepare_config_and_inputs() self.check_vision_text_output_attention(**_SCREAMING_SNAKE_CASE) @is_pt_flax_cross_test def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> Any: """simple docstring""" __lowerCAmelCase : Dict = self.prepare_config_and_inputs() __lowerCAmelCase : List[Any] = config_inputs_dict.pop("vision_config") __lowerCAmelCase : str = config_inputs_dict.pop("text_config") __lowerCAmelCase : Union[str, Any] = config_inputs_dict self.check_equivalence_pt_to_flax(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) self.check_equivalence_flax_to_pt(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) @slow def _SCREAMING_SNAKE_CASE ( self: str) -> Dict: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : Dict = self.get_pretrained_model_and_inputs() __lowerCAmelCase : Union[str, Any] = model_a(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = outputs[0] with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = FlaxVisionTextDualEncoderModel.from_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = model_a(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = after_outputs[0] __lowerCAmelCase : List[Any] = np.amax(np.abs(out_a - out_a)) self.assertLessEqual(_SCREAMING_SNAKE_CASE , 1e-5) @require_flax class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: List[str]) -> List[Any]: """simple docstring""" __lowerCAmelCase : Any = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( "hf-internal-testing/tiny-random-vit" , "hf-internal-testing/tiny-bert" , vision_from_pt=_SCREAMING_SNAKE_CASE , text_from_pt=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Union[str, Any] = 13 __lowerCAmelCase : Optional[int] = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ]) __lowerCAmelCase : List[Any] = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size) __lowerCAmelCase : List[Any] = random_attention_mask([batch_size, 4]) __lowerCAmelCase : str = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def _SCREAMING_SNAKE_CASE ( self: int , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Union[str, Any]) -> Optional[int]: """simple docstring""" __lowerCAmelCase : List[str] = FlaxViTModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = FlaxBertModel(_SCREAMING_SNAKE_CASE) return vision_model, text_model def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> int: """simple docstring""" __lowerCAmelCase : List[Any] = FlaxViTModelTester(self) __lowerCAmelCase : Optional[Any] = FlaxBertModelTester(self) __lowerCAmelCase : int = vit_model_tester.prepare_config_and_inputs() __lowerCAmelCase : List[str] = bert_model_tester.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase : Tuple = vision_config_and_inputs __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Union[str, Any] = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_torch class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Dict) -> List[str]: """simple docstring""" __lowerCAmelCase : Optional[int] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( "hf-internal-testing/tiny-random-clip" , "hf-internal-testing/tiny-bert" , vision_from_pt=_SCREAMING_SNAKE_CASE , text_from_pt=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Optional[int] = 13 __lowerCAmelCase : List[str] = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ]) __lowerCAmelCase : Any = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size) __lowerCAmelCase : str = random_attention_mask([batch_size, 4]) __lowerCAmelCase : Optional[int] = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Union[str, Any]) -> int: """simple docstring""" __lowerCAmelCase : int = FlaxCLIPVisionModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = FlaxBertModel(_SCREAMING_SNAKE_CASE) return vision_model, text_model def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : List[Any] = FlaxCLIPVisionModelTester(self) __lowerCAmelCase : str = FlaxBertModelTester(self) __lowerCAmelCase : Optional[Any] = clip_model_tester.prepare_config_and_inputs() __lowerCAmelCase : Dict = bert_model_tester.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase : Any = vision_config_and_inputs __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : List[Any] = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_flax @require_vision class A__ ( unittest.TestCase ): '''simple docstring''' @slow def _SCREAMING_SNAKE_CASE ( self: List[str]) -> List[str]: """simple docstring""" __lowerCAmelCase : Dict = FlaxVisionTextDualEncoderModel.from_pretrained("clip-italian/clip-italian" , logit_scale_init_value=1.0) __lowerCAmelCase : str = VisionTextDualEncoderProcessor.from_pretrained("clip-italian/clip-italian") __lowerCAmelCase : Tuple = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png") __lowerCAmelCase : Optional[int] = processor( text=["una foto di un gatto", "una foto di un cane"] , images=_SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE , return_tensors="np") __lowerCAmelCase : List[str] = model(**_SCREAMING_SNAKE_CASE) # verify the logits self.assertEqual(outputs.logits_per_image.shape , (inputs.pixel_values.shape[0], inputs.input_ids.shape[0])) self.assertEqual( outputs.logits_per_text.shape , (inputs.input_ids.shape[0], inputs.pixel_values.shape[0]) , ) __lowerCAmelCase : List[str] = np.array([[1.228_4727, 0.310_4122]]) self.assertTrue(np.allclose(outputs.logits_per_image , _SCREAMING_SNAKE_CASE , atol=1e-3))
269
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_videomae import VideoMAEImageProcessor __snake_case : str = logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self: Any , *_SCREAMING_SNAKE_CASE: Dict , **_SCREAMING_SNAKE_CASE: List[str]) -> None: """simple docstring""" warnings.warn( "The class VideoMAEFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use VideoMAEImageProcessor instead." , _SCREAMING_SNAKE_CASE , ) super().__init__(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE)
269
"""simple docstring""" from __future__ import annotations def _lowercase ( __snake_case ,__snake_case ,__snake_case ,__snake_case ) -> list: __lowerCAmelCase : Dict = [] __lowerCAmelCase , __lowerCAmelCase : Any = input_list[low:mid], input_list[mid : high + 1] while left and right: result.append((left if left[0] <= right[0] else right).pop(0 ) ) __lowerCAmelCase : int = result + left + right return input_list def _lowercase ( __snake_case ) -> list: if len(__snake_case ) <= 1: return input_list __lowerCAmelCase : int = list(__snake_case ) # iteration for two-way merging __lowerCAmelCase : Optional[int] = 2 while p <= len(__snake_case ): # getting low, high and middle value for merge-sort of single list for i in range(0 ,len(__snake_case ) ,__snake_case ): __lowerCAmelCase : Union[str, Any] = i __lowerCAmelCase : Tuple = i + p - 1 __lowerCAmelCase : Optional[Any] = (low + high + 1) // 2 __lowerCAmelCase : Any = merge(__snake_case ,__snake_case ,__snake_case ,__snake_case ) # final merge of last two parts if p * 2 >= len(__snake_case ): __lowerCAmelCase : Optional[Any] = i __lowerCAmelCase : Union[str, Any] = merge(__snake_case ,0 ,__snake_case ,len(__snake_case ) - 1 ) break p *= 2 return input_list if __name__ == "__main__": __snake_case : Union[str, Any] = input('Enter numbers separated by a comma:\n').strip() if user_input == "": __snake_case : Optional[int] = [] else: __snake_case : int = [int(item.strip()) for item in user_input.split(',')] print(iter_merge_sort(unsorted))
269
1
"""simple docstring""" import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = ['image_processor', 'tokenizer'] SCREAMING_SNAKE_CASE = 'LayoutLMv2ImageProcessor' SCREAMING_SNAKE_CASE = ('LayoutXLMTokenizer', 'LayoutXLMTokenizerFast') def __init__( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: Tuple=None , _SCREAMING_SNAKE_CASE: List[Any]=None , **_SCREAMING_SNAKE_CASE: List[str]) -> Any: """simple docstring""" if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , _SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Optional[int] = kwargs.pop("feature_extractor") __lowerCAmelCase : int = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`.") if tokenizer is None: raise ValueError("You need to specify a `tokenizer`.") super().__init__(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) def __call__( self: Tuple , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , _SCREAMING_SNAKE_CASE: Optional[Union[PreTokenizedInput, List[PreTokenizedInput]]] = None , _SCREAMING_SNAKE_CASE: Union[List[List[int]], List[List[List[int]]]] = None , _SCREAMING_SNAKE_CASE: Optional[Union[List[int], List[List[int]]]] = None , _SCREAMING_SNAKE_CASE: bool = True , _SCREAMING_SNAKE_CASE: Union[bool, str, PaddingStrategy] = False , _SCREAMING_SNAKE_CASE: Union[bool, str, TruncationStrategy] = None , _SCREAMING_SNAKE_CASE: Optional[int] = None , _SCREAMING_SNAKE_CASE: int = 0 , _SCREAMING_SNAKE_CASE: Optional[int] = None , _SCREAMING_SNAKE_CASE: Optional[bool] = None , _SCREAMING_SNAKE_CASE: Optional[bool] = None , _SCREAMING_SNAKE_CASE: bool = False , _SCREAMING_SNAKE_CASE: bool = False , _SCREAMING_SNAKE_CASE: bool = False , _SCREAMING_SNAKE_CASE: bool = False , _SCREAMING_SNAKE_CASE: bool = True , _SCREAMING_SNAKE_CASE: Optional[Union[str, TensorType]] = None , **_SCREAMING_SNAKE_CASE: int , ) -> BatchEncoding: """simple docstring""" if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( "You cannot provide bounding boxes " "if you initialized the image processor with apply_ocr set to True.") if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( "You cannot provide word labels if you initialized the image processor with apply_ocr set to True.") if return_overflowing_tokens is True and return_offsets_mapping is False: raise ValueError("You cannot return overflowing tokens without returning the offsets mapping.") # first, apply the image processor __lowerCAmelCase : List[Any] = self.image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors=_SCREAMING_SNAKE_CASE) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE): __lowerCAmelCase : Tuple = [text] # add batch dimension (as the image processor always adds a batch dimension) __lowerCAmelCase : Optional[Any] = features["words"] __lowerCAmelCase : Union[str, Any] = self.tokenizer( text=text if text is not None else features["words"] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features["boxes"] , word_labels=_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE , truncation=_SCREAMING_SNAKE_CASE , max_length=_SCREAMING_SNAKE_CASE , stride=_SCREAMING_SNAKE_CASE , pad_to_multiple_of=_SCREAMING_SNAKE_CASE , return_token_type_ids=_SCREAMING_SNAKE_CASE , return_attention_mask=_SCREAMING_SNAKE_CASE , return_overflowing_tokens=_SCREAMING_SNAKE_CASE , return_special_tokens_mask=_SCREAMING_SNAKE_CASE , return_offsets_mapping=_SCREAMING_SNAKE_CASE , return_length=_SCREAMING_SNAKE_CASE , verbose=_SCREAMING_SNAKE_CASE , return_tensors=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) # add pixel values __lowerCAmelCase : Dict = features.pop("pixel_values") if return_overflowing_tokens is True: __lowerCAmelCase : Dict = self.get_overflowing_images(_SCREAMING_SNAKE_CASE , encoded_inputs["overflow_to_sample_mapping"]) __lowerCAmelCase : Union[str, Any] = images return encoded_inputs def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: List[Any]) -> List[Any]: """simple docstring""" __lowerCAmelCase : str = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx]) if len(_SCREAMING_SNAKE_CASE) != len(_SCREAMING_SNAKE_CASE): raise ValueError( "Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got" F""" {len(_SCREAMING_SNAKE_CASE)} and {len(_SCREAMING_SNAKE_CASE)}""") return images_with_overflow def _SCREAMING_SNAKE_CASE ( self: int , *_SCREAMING_SNAKE_CASE: int , **_SCREAMING_SNAKE_CASE: List[Any]) -> Tuple: """simple docstring""" return self.tokenizer.batch_decode(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: int , *_SCREAMING_SNAKE_CASE: Union[str, Any] , **_SCREAMING_SNAKE_CASE: Optional[Any]) -> Optional[Any]: """simple docstring""" return self.tokenizer.decode(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE) @property def _SCREAMING_SNAKE_CASE ( self: Dict) -> List[str]: """simple docstring""" return ["input_ids", "bbox", "attention_mask", "image"] @property def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> List[str]: """simple docstring""" warnings.warn( "`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead." , _SCREAMING_SNAKE_CASE , ) return self.image_processor_class @property def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> int: """simple docstring""" warnings.warn( "`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead." , _SCREAMING_SNAKE_CASE , ) return self.image_processor
269
"""simple docstring""" import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class A__ ( unittest.TestCase ): '''simple docstring''' @slow def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> str: """simple docstring""" __lowerCAmelCase : Optional[Any] = FlaxMTaForConditionalGeneration.from_pretrained("google/mt5-small") __lowerCAmelCase : Optional[Any] = AutoTokenizer.from_pretrained("google/mt5-small") __lowerCAmelCase : Tuple = tokenizer("Hello there" , return_tensors="np").input_ids __lowerCAmelCase : Dict = tokenizer("Hi I am" , return_tensors="np").input_ids __lowerCAmelCase : str = shift_tokens_right(_SCREAMING_SNAKE_CASE , model.config.pad_token_id , model.config.decoder_start_token_id) __lowerCAmelCase : Optional[int] = model(_SCREAMING_SNAKE_CASE , decoder_input_ids=_SCREAMING_SNAKE_CASE).logits __lowerCAmelCase : int = optax.softmax_cross_entropy(_SCREAMING_SNAKE_CASE , onehot(_SCREAMING_SNAKE_CASE , logits.shape[-1])).mean() __lowerCAmelCase : List[str] = -(labels.shape[-1] * loss.item()) __lowerCAmelCase : str = -84.9127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE) < 1e-4)
269
1
"""simple docstring""" import warnings from ...utils import is_sklearn_available, requires_backends if is_sklearn_available(): from scipy.stats import pearsonr, spearmanr from sklearn.metrics import fa_score, matthews_corrcoef __snake_case : Union[str, Any] = ( 'This metric will be removed from the library soon, metrics should be handled with the 🤗 Evaluate ' 'library. You can have a look at this example script for pointers: ' 'https://github.com/huggingface/transformers/blob/main/examples/pytorch/text-classification/run_glue.py' ) def _lowercase ( __snake_case ,__snake_case ) -> List[str]: warnings.warn(__snake_case ,__snake_case ) requires_backends(__snake_case ,"sklearn" ) return (preds == labels).mean() def _lowercase ( __snake_case ,__snake_case ) -> int: warnings.warn(__snake_case ,__snake_case ) requires_backends(__snake_case ,"sklearn" ) __lowerCAmelCase : Union[str, Any] = simple_accuracy(__snake_case ,__snake_case ) __lowerCAmelCase : str = fa_score(y_true=__snake_case ,y_pred=__snake_case ) return { "acc": acc, "f1": fa, "acc_and_f1": (acc + fa) / 2, } def _lowercase ( __snake_case ,__snake_case ) -> Union[str, Any]: warnings.warn(__snake_case ,__snake_case ) requires_backends(__snake_case ,"sklearn" ) __lowerCAmelCase : Union[str, Any] = pearsonr(__snake_case ,__snake_case )[0] __lowerCAmelCase : str = spearmanr(__snake_case ,__snake_case )[0] return { "pearson": pearson_corr, "spearmanr": spearman_corr, "corr": (pearson_corr + spearman_corr) / 2, } def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> Any: warnings.warn(__snake_case ,__snake_case ) requires_backends(__snake_case ,"sklearn" ) assert len(__snake_case ) == len(__snake_case ), F"""Predictions and labels have mismatched lengths {len(__snake_case )} and {len(__snake_case )}""" if task_name == "cola": return {"mcc": matthews_corrcoef(__snake_case ,__snake_case )} elif task_name == "sst-2": return {"acc": simple_accuracy(__snake_case ,__snake_case )} elif task_name == "mrpc": return acc_and_fa(__snake_case ,__snake_case ) elif task_name == "sts-b": return pearson_and_spearman(__snake_case ,__snake_case ) elif task_name == "qqp": return acc_and_fa(__snake_case ,__snake_case ) elif task_name == "mnli": return {"mnli/acc": simple_accuracy(__snake_case ,__snake_case )} elif task_name == "mnli-mm": return {"mnli-mm/acc": simple_accuracy(__snake_case ,__snake_case )} elif task_name == "qnli": return {"acc": simple_accuracy(__snake_case ,__snake_case )} elif task_name == "rte": return {"acc": simple_accuracy(__snake_case ,__snake_case )} elif task_name == "wnli": return {"acc": simple_accuracy(__snake_case ,__snake_case )} elif task_name == "hans": return {"acc": simple_accuracy(__snake_case ,__snake_case )} else: raise KeyError(__snake_case ) def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> Union[str, Any]: warnings.warn(__snake_case ,__snake_case ) requires_backends(__snake_case ,"sklearn" ) if len(__snake_case ) != len(__snake_case ): raise ValueError(F"""Predictions and labels have mismatched lengths {len(__snake_case )} and {len(__snake_case )}""" ) if task_name == "xnli": return {"acc": simple_accuracy(__snake_case ,__snake_case )} else: raise KeyError(__snake_case )
269
"""simple docstring""" import re def _lowercase ( __snake_case ) -> str: if len(re.findall("[ATCG]" ,__snake_case ) ) != len(__snake_case ): raise ValueError("Invalid Strand" ) return dna.translate(dna.maketrans("ATCG" ,"TAGC" ) ) if __name__ == "__main__": import doctest doctest.testmod()
269
1
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import convert_to_rgb, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL __snake_case : Dict = logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = ['pixel_values'] def __init__( self: Tuple , _SCREAMING_SNAKE_CASE: bool = True , _SCREAMING_SNAKE_CASE: Dict[str, int] = None , _SCREAMING_SNAKE_CASE: PILImageResampling = PILImageResampling.BICUBIC , _SCREAMING_SNAKE_CASE: bool = True , _SCREAMING_SNAKE_CASE: Union[int, float] = 1 / 255 , _SCREAMING_SNAKE_CASE: bool = True , _SCREAMING_SNAKE_CASE: Optional[Union[float, List[float]]] = None , _SCREAMING_SNAKE_CASE: Optional[Union[float, List[float]]] = None , _SCREAMING_SNAKE_CASE: bool = True , **_SCREAMING_SNAKE_CASE: int , ) -> None: """simple docstring""" super().__init__(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = size if size is not None else {"height": 384, "width": 384} __lowerCAmelCase : Union[str, Any] = get_size_dict(_SCREAMING_SNAKE_CASE , default_to_square=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = do_resize __lowerCAmelCase : Optional[Any] = size __lowerCAmelCase : Any = resample __lowerCAmelCase : Union[str, Any] = do_rescale __lowerCAmelCase : Optional[int] = rescale_factor __lowerCAmelCase : Dict = do_normalize __lowerCAmelCase : Any = image_mean if image_mean is not None else OPENAI_CLIP_MEAN __lowerCAmelCase : Tuple = image_std if image_std is not None else OPENAI_CLIP_STD __lowerCAmelCase : List[Any] = do_convert_rgb def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: np.ndarray , _SCREAMING_SNAKE_CASE: Dict[str, int] , _SCREAMING_SNAKE_CASE: PILImageResampling = PILImageResampling.BICUBIC , _SCREAMING_SNAKE_CASE: Optional[Union[str, ChannelDimension]] = None , **_SCREAMING_SNAKE_CASE: int , ) -> np.ndarray: """simple docstring""" __lowerCAmelCase : int = get_size_dict(_SCREAMING_SNAKE_CASE , default_to_square=_SCREAMING_SNAKE_CASE) if "height" not in size or "width" not in size: raise ValueError(F"""The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}""") __lowerCAmelCase : Tuple = (size["height"], size["width"]) return resize(_SCREAMING_SNAKE_CASE , size=_SCREAMING_SNAKE_CASE , resample=_SCREAMING_SNAKE_CASE , data_format=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Dict , _SCREAMING_SNAKE_CASE: np.ndarray , _SCREAMING_SNAKE_CASE: Union[int, float] , _SCREAMING_SNAKE_CASE: Optional[Union[str, ChannelDimension]] = None , **_SCREAMING_SNAKE_CASE: List[str] , ) -> List[Any]: """simple docstring""" return rescale(_SCREAMING_SNAKE_CASE , scale=_SCREAMING_SNAKE_CASE , data_format=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: np.ndarray , _SCREAMING_SNAKE_CASE: Union[float, List[float]] , _SCREAMING_SNAKE_CASE: Union[float, List[float]] , _SCREAMING_SNAKE_CASE: Optional[Union[str, ChannelDimension]] = None , **_SCREAMING_SNAKE_CASE: int , ) -> np.ndarray: """simple docstring""" return normalize(_SCREAMING_SNAKE_CASE , mean=_SCREAMING_SNAKE_CASE , std=_SCREAMING_SNAKE_CASE , data_format=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: ImageInput , _SCREAMING_SNAKE_CASE: Optional[bool] = None , _SCREAMING_SNAKE_CASE: Optional[Dict[str, int]] = None , _SCREAMING_SNAKE_CASE: PILImageResampling = None , _SCREAMING_SNAKE_CASE: Optional[bool] = None , _SCREAMING_SNAKE_CASE: Optional[float] = None , _SCREAMING_SNAKE_CASE: Optional[bool] = None , _SCREAMING_SNAKE_CASE: Optional[Union[float, List[float]]] = None , _SCREAMING_SNAKE_CASE: Optional[Union[float, List[float]]] = None , _SCREAMING_SNAKE_CASE: Optional[Union[str, TensorType]] = None , _SCREAMING_SNAKE_CASE: bool = None , _SCREAMING_SNAKE_CASE: ChannelDimension = ChannelDimension.FIRST , **_SCREAMING_SNAKE_CASE: int , ) -> PIL.Image.Image: """simple docstring""" __lowerCAmelCase : Optional[Any] = do_resize if do_resize is not None else self.do_resize __lowerCAmelCase : Union[str, Any] = resample if resample is not None else self.resample __lowerCAmelCase : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale __lowerCAmelCase : Optional[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor __lowerCAmelCase : str = do_normalize if do_normalize is not None else self.do_normalize __lowerCAmelCase : Optional[int] = image_mean if image_mean is not None else self.image_mean __lowerCAmelCase : Optional[Any] = image_std if image_std is not None else self.image_std __lowerCAmelCase : Optional[int] = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb __lowerCAmelCase : str = size if size is not None else self.size __lowerCAmelCase : List[str] = get_size_dict(_SCREAMING_SNAKE_CASE , default_to_square=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = make_list_of_images(_SCREAMING_SNAKE_CASE) if not valid_images(_SCREAMING_SNAKE_CASE): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray.") if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True.") if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True.") if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True.") # PIL RGBA images are converted to RGB if do_convert_rgb: __lowerCAmelCase : List[Any] = [convert_to_rgb(_SCREAMING_SNAKE_CASE) for image in images] # All transformations expect numpy arrays. __lowerCAmelCase : Any = [to_numpy_array(_SCREAMING_SNAKE_CASE) for image in images] if do_resize: __lowerCAmelCase : str = [self.resize(image=_SCREAMING_SNAKE_CASE , size=_SCREAMING_SNAKE_CASE , resample=_SCREAMING_SNAKE_CASE) for image in images] if do_rescale: __lowerCAmelCase : List[Any] = [self.rescale(image=_SCREAMING_SNAKE_CASE , scale=_SCREAMING_SNAKE_CASE) for image in images] if do_normalize: __lowerCAmelCase : List[str] = [self.normalize(image=_SCREAMING_SNAKE_CASE , mean=_SCREAMING_SNAKE_CASE , std=_SCREAMING_SNAKE_CASE) for image in images] __lowerCAmelCase : Any = [to_channel_dimension_format(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) for image in images] __lowerCAmelCase : Optional[Any] = BatchFeature(data={"pixel_values": images} , tensor_type=_SCREAMING_SNAKE_CASE) return encoded_outputs
269
"""simple docstring""" import os import unittest from transformers import FunnelTokenizer, FunnelTokenizerFast from transformers.models.funnel.tokenization_funnel import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = FunnelTokenizer SCREAMING_SNAKE_CASE = FunnelTokenizerFast SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = True def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Optional[int]: """simple docstring""" super().setUp() __lowerCAmelCase : str = [ "<unk>", "<cls>", "<sep>", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] __lowerCAmelCase : int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"]) with open(self.vocab_file , "w" , encoding="utf-8") as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens])) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , **_SCREAMING_SNAKE_CASE: Union[str, Any]) -> int: """simple docstring""" return FunnelTokenizer.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Any , **_SCREAMING_SNAKE_CASE: Any) -> str: """simple docstring""" return FunnelTokenizerFast.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: str) -> Any: """simple docstring""" __lowerCAmelCase : Union[str, Any] = "UNwant\u00E9d,running" __lowerCAmelCase : str = "unwanted, running" return input_text, output_text def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> List[str]: """simple docstring""" __lowerCAmelCase : Any = self.tokenizer_class(self.vocab_file) __lowerCAmelCase : Any = tokenizer.tokenize("UNwant\u00E9d,running") self.assertListEqual(_SCREAMING_SNAKE_CASE , ["un", "##want", "##ed", ",", "runn", "##ing"]) self.assertListEqual(tokenizer.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE) , [7, 4, 5, 10, 8, 9]) def _SCREAMING_SNAKE_CASE ( self: List[str]) -> List[str]: """simple docstring""" __lowerCAmelCase : Optional[Any] = self.get_tokenizers(do_lower_case=_SCREAMING_SNAKE_CASE) for tokenizer in tokenizers: __lowerCAmelCase : List[str] = tokenizer("UNwant\u00E9d,running") __lowerCAmelCase : Optional[int] = len(inputs["input_ids"]) - 1 self.assertListEqual(inputs["token_type_ids"] , [2] + [0] * sentence_len) __lowerCAmelCase : List[str] = tokenizer("UNwant\u00E9d,running" , "UNwant\u00E9d,running") self.assertListEqual(inputs["token_type_ids"] , [2] + [0] * sentence_len + [1] * sentence_len)
269
1
"""simple docstring""" import os import socket from contextlib import contextmanager import torch from ..commands.config.default import write_basic_config # noqa: F401 from ..state import PartialState from .dataclasses import DistributedType from .imports import is_deepspeed_available, is_tpu_available from .transformer_engine import convert_model from .versions import is_torch_version if is_deepspeed_available(): from deepspeed import DeepSpeedEngine if is_tpu_available(check_device=False): import torch_xla.core.xla_model as xm def _lowercase ( __snake_case ) -> Union[str, Any]: if is_torch_version("<" ,"2.0.0" ) or not hasattr(__snake_case ,"_dynamo" ): return False return isinstance(__snake_case ,torch._dynamo.eval_frame.OptimizedModule ) def _lowercase ( __snake_case ,__snake_case = True ) -> Union[str, Any]: __lowerCAmelCase : Dict = (torch.nn.parallel.DistributedDataParallel, torch.nn.DataParallel) __lowerCAmelCase : Optional[int] = is_compiled_module(__snake_case ) if is_compiled: __lowerCAmelCase : str = model __lowerCAmelCase : Any = model._orig_mod if is_deepspeed_available(): options += (DeepSpeedEngine,) while isinstance(__snake_case ,__snake_case ): __lowerCAmelCase : Any = model.module if not keep_fpaa_wrapper: __lowerCAmelCase : Dict = getattr(__snake_case ,"forward" ) __lowerCAmelCase : Dict = model.__dict__.pop("_original_forward" ,__snake_case ) if original_forward is not None: while hasattr(__snake_case ,"__wrapped__" ): __lowerCAmelCase : List[str] = forward.__wrapped__ if forward == original_forward: break __lowerCAmelCase : Optional[int] = forward if getattr(__snake_case ,"_converted_to_transformer_engine" ,__snake_case ): convert_model(__snake_case ,to_transformer_engine=__snake_case ) if is_compiled: __lowerCAmelCase : Dict = model __lowerCAmelCase : Optional[Any] = compiled_model return model def _lowercase ( ) -> List[str]: PartialState().wait_for_everyone() def _lowercase ( __snake_case ,__snake_case ) -> List[str]: if PartialState().distributed_type == DistributedType.TPU: xm.save(__snake_case ,__snake_case ) elif PartialState().local_process_index == 0: torch.save(__snake_case ,__snake_case ) @contextmanager def _lowercase ( **__snake_case ) -> Optional[int]: for key, value in kwargs.items(): __lowerCAmelCase : List[Any] = str(__snake_case ) yield for key in kwargs: if key.upper() in os.environ: del os.environ[key.upper()] def _lowercase ( __snake_case ) -> Tuple: if not hasattr(__snake_case ,"__qualname__" ) and not hasattr(__snake_case ,"__name__" ): __lowerCAmelCase : Union[str, Any] = getattr(__snake_case ,"__class__" ,__snake_case ) if hasattr(__snake_case ,"__qualname__" ): return obj.__qualname__ if hasattr(__snake_case ,"__name__" ): return obj.__name__ return str(__snake_case ) def _lowercase ( __snake_case ,__snake_case ) -> Dict: for key, value in source.items(): if isinstance(__snake_case ,__snake_case ): __lowerCAmelCase : Tuple = destination.setdefault(__snake_case ,{} ) merge_dicts(__snake_case ,__snake_case ) else: __lowerCAmelCase : List[Any] = value return destination def _lowercase ( __snake_case = None ) -> bool: if port is None: __lowerCAmelCase : str = 29_500 with socket.socket(socket.AF_INET ,socket.SOCK_STREAM ) as s: return s.connect_ex(("localhost", port) ) == 0
269
"""simple docstring""" from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def _lowercase ( __snake_case = "laptop" ) -> DataFrame: __lowerCAmelCase : str = F"""https://www.amazon.in/laptop/s?k={product}""" __lowerCAmelCase : Union[str, Any] = { "User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36\n (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36", "Accept-Language": "en-US, en;q=0.5", } __lowerCAmelCase : List[str] = BeautifulSoup(requests.get(__snake_case ,headers=__snake_case ).text ) # Initialize a Pandas dataframe with the column titles __lowerCAmelCase : Dict = DataFrame( columns=[ "Product Title", "Product Link", "Current Price of the product", "Product Rating", "MRP of the product", "Discount", ] ) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( "div" ,attrs={"class": "s-result-item", "data-component-type": "s-search-result"} ,) ,soup.find_all("div" ,attrs={"class": "a-row a-size-base a-color-base"} ) ,): try: __lowerCAmelCase : Any = item.ha.text __lowerCAmelCase : Union[str, Any] = "https://www.amazon.in/" + item.ha.a["href"] __lowerCAmelCase : Any = item.find("span" ,attrs={"class": "a-offscreen"} ).text try: __lowerCAmelCase : Union[str, Any] = item.find("span" ,attrs={"class": "a-icon-alt"} ).text except AttributeError: __lowerCAmelCase : Optional[Any] = "Not available" try: __lowerCAmelCase : Union[str, Any] = ( "₹" + item.find( "span" ,attrs={"class": "a-price a-text-price"} ).text.split("₹" )[1] ) except AttributeError: __lowerCAmelCase : Dict = "" try: __lowerCAmelCase : str = float( ( ( float(product_mrp.strip("₹" ).replace("," ,"" ) ) - float(product_price.strip("₹" ).replace("," ,"" ) ) ) / float(product_mrp.strip("₹" ).replace("," ,"" ) ) ) * 100 ) except ValueError: __lowerCAmelCase : List[str] = float("nan" ) except AttributeError: pass __lowerCAmelCase : int = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] __lowerCAmelCase : Union[str, Any] = " " __lowerCAmelCase : Union[str, Any] = " " data_frame.index += 1 return data_frame if __name__ == "__main__": __snake_case : Any = 'headphones' get_amazon_product_data(product).to_csv(F"""Amazon Product Data for {product}.csv""")
269
1
"""simple docstring""" import inspect import unittest from transformers import DPTConfig from transformers.file_utils import is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MODEL_MAPPING, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTModel from transformers.models.dpt.modeling_dpt import DPT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class A__ : '''simple docstring''' def __init__( self: Any , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[Any]=2 , _SCREAMING_SNAKE_CASE: Dict=32 , _SCREAMING_SNAKE_CASE: Union[str, Any]=16 , _SCREAMING_SNAKE_CASE: Dict=3 , _SCREAMING_SNAKE_CASE: Tuple=True , _SCREAMING_SNAKE_CASE: Dict=True , _SCREAMING_SNAKE_CASE: Optional[Any]=32 , _SCREAMING_SNAKE_CASE: Dict=4 , _SCREAMING_SNAKE_CASE: Tuple=[0, 1, 2, 3] , _SCREAMING_SNAKE_CASE: Optional[int]=4 , _SCREAMING_SNAKE_CASE: List[str]=37 , _SCREAMING_SNAKE_CASE: int="gelu" , _SCREAMING_SNAKE_CASE: Optional[int]=0.1 , _SCREAMING_SNAKE_CASE: str=0.1 , _SCREAMING_SNAKE_CASE: List[str]=0.02 , _SCREAMING_SNAKE_CASE: Dict=3 , _SCREAMING_SNAKE_CASE: Dict=[1, 384, 24, 24] , _SCREAMING_SNAKE_CASE: Any=True , _SCREAMING_SNAKE_CASE: Optional[int]=None , ) -> int: """simple docstring""" __lowerCAmelCase : Optional[int] = parent __lowerCAmelCase : List[Any] = batch_size __lowerCAmelCase : str = image_size __lowerCAmelCase : List[Any] = patch_size __lowerCAmelCase : Optional[Any] = num_channels __lowerCAmelCase : Dict = is_training __lowerCAmelCase : Optional[int] = use_labels __lowerCAmelCase : Optional[int] = hidden_size __lowerCAmelCase : List[Any] = num_hidden_layers __lowerCAmelCase : Optional[Any] = backbone_out_indices __lowerCAmelCase : Optional[Any] = num_attention_heads __lowerCAmelCase : Optional[Any] = intermediate_size __lowerCAmelCase : str = hidden_act __lowerCAmelCase : List[str] = hidden_dropout_prob __lowerCAmelCase : Optional[Any] = attention_probs_dropout_prob __lowerCAmelCase : List[Any] = initializer_range __lowerCAmelCase : Any = num_labels __lowerCAmelCase : List[Any] = backbone_featmap_shape __lowerCAmelCase : Optional[int] = scope __lowerCAmelCase : str = is_hybrid # sequence length of DPT = num_patches + 1 (we add 1 for the [CLS] token) __lowerCAmelCase : List[str] = (image_size // patch_size) ** 2 __lowerCAmelCase : List[str] = num_patches + 1 def _SCREAMING_SNAKE_CASE ( self: List[str]) -> Tuple: """simple docstring""" __lowerCAmelCase : Dict = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) __lowerCAmelCase : Any = None if self.use_labels: __lowerCAmelCase : Optional[Any] = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels) __lowerCAmelCase : str = self.get_config() return config, pixel_values, labels def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> str: """simple docstring""" __lowerCAmelCase : Dict = { "global_padding": "same", "layer_type": "bottleneck", "depths": [3, 4, 9], "out_features": ["stage1", "stage2", "stage3"], "embedding_dynamic_padding": True, "hidden_sizes": [96, 192, 384, 768], "num_groups": 2, } return DPTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , backbone_out_indices=self.backbone_out_indices , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_SCREAMING_SNAKE_CASE , initializer_range=self.initializer_range , is_hybrid=self.is_hybrid , backbone_config=_SCREAMING_SNAKE_CASE , backbone_featmap_shape=self.backbone_featmap_shape , ) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Tuple) -> Any: """simple docstring""" __lowerCAmelCase : Optional[Any] = DPTModel(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : Union[str, Any] = model(_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Union[str, Any]) -> Any: """simple docstring""" __lowerCAmelCase : str = self.num_labels __lowerCAmelCase : int = DPTForDepthEstimation(_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : List[Any] = model(_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.predicted_depth.shape , (self.batch_size, self.image_size, self.image_size)) def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Union[str, Any]) -> Optional[int]: """simple docstring""" __lowerCAmelCase : Any = self.num_labels __lowerCAmelCase : List[str] = DPTForSemanticSegmentation(_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : Dict = model(_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size)) def _SCREAMING_SNAKE_CASE ( self: Dict) -> Dict: """simple docstring""" __lowerCAmelCase : str = self.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Dict = config_and_inputs __lowerCAmelCase : List[str] = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = (DPTModel, DPTForDepthEstimation, DPTForSemanticSegmentation) if is_torch_available() else () SCREAMING_SNAKE_CASE = ( { 'depth-estimation': DPTForDepthEstimation, 'feature-extraction': DPTModel, 'image-segmentation': DPTForSemanticSegmentation, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False def _SCREAMING_SNAKE_CASE ( self: Dict) -> Tuple: """simple docstring""" __lowerCAmelCase : List[str] = DPTModelTester(self) __lowerCAmelCase : Union[str, Any] = ConfigTester(self , config_class=_SCREAMING_SNAKE_CASE , has_text_modality=_SCREAMING_SNAKE_CASE , hidden_size=37) def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> Optional[Any]: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="DPT does not use inputs_embeds") def _SCREAMING_SNAKE_CASE ( self: Dict) -> Union[str, Any]: """simple docstring""" pass def _SCREAMING_SNAKE_CASE ( self: str) -> Any: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCAmelCase : List[Any] = model_class(_SCREAMING_SNAKE_CASE) self.assertIsInstance(model.get_input_embeddings() , (nn.Module)) __lowerCAmelCase : Optional[Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_SCREAMING_SNAKE_CASE , nn.Linear)) def _SCREAMING_SNAKE_CASE ( self: Any) -> List[Any]: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCAmelCase : Any = model_class(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCAmelCase : List[str] = [*signature.parameters.keys()] __lowerCAmelCase : Optional[Any] = ["pixel_values"] self.assertListEqual(arg_names[:1] , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Dict: """simple docstring""" __lowerCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_depth_estimation(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> Any: """simple docstring""" __lowerCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: str) -> Tuple: """simple docstring""" for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue __lowerCAmelCase , __lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() __lowerCAmelCase : Tuple = True if model_class in get_values(_SCREAMING_SNAKE_CASE): continue __lowerCAmelCase : Tuple = model_class(_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.train() __lowerCAmelCase : List[Any] = self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , return_labels=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = model(**_SCREAMING_SNAKE_CASE).loss loss.backward() def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> Optional[int]: """simple docstring""" for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue __lowerCAmelCase , __lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() __lowerCAmelCase : List[Any] = False __lowerCAmelCase : Tuple = True if model_class in get_values(_SCREAMING_SNAKE_CASE) or not model_class.supports_gradient_checkpointing: continue __lowerCAmelCase : List[str] = model_class(_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.gradient_checkpointing_enable() model.train() __lowerCAmelCase : int = self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , return_labels=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = model(**_SCREAMING_SNAKE_CASE).loss loss.backward() def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> List[str]: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() __lowerCAmelCase : str = _config_zero_init(_SCREAMING_SNAKE_CASE) for model_class in self.all_model_classes: __lowerCAmelCase : Union[str, Any] = model_class(config=_SCREAMING_SNAKE_CASE) # Skip the check for the backbone __lowerCAmelCase : List[Any] = [] for name, module in model.named_modules(): if module.__class__.__name__ == "DPTViTHybridEmbeddings": __lowerCAmelCase : Dict = [F"""{name}.{key}""" for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests.") def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> str: """simple docstring""" pass @slow def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> List[str]: """simple docstring""" for model_name in DPT_PRETRAINED_MODEL_ARCHIVE_LIST[1:]: __lowerCAmelCase : Any = DPTModel.from_pretrained(_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[str]) -> str: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() __lowerCAmelCase : Any = "add" with self.assertRaises(_SCREAMING_SNAKE_CASE): __lowerCAmelCase : Tuple = DPTForDepthEstimation(_SCREAMING_SNAKE_CASE) def _lowercase ( ) -> Union[str, Any]: __lowerCAmelCase : Dict = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision @slow class A__ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : List[Any] = DPTImageProcessor.from_pretrained("Intel/dpt-hybrid-midas") __lowerCAmelCase : Tuple = DPTForDepthEstimation.from_pretrained("Intel/dpt-hybrid-midas").to(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = prepare_img() __lowerCAmelCase : int = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors="pt").to(_SCREAMING_SNAKE_CASE) # forward pass with torch.no_grad(): __lowerCAmelCase : Union[str, Any] = model(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = outputs.predicted_depth # verify the predicted depth __lowerCAmelCase : List[Any] = torch.Size((1, 384, 384)) self.assertEqual(predicted_depth.shape , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = torch.tensor( [[[5.6437, 5.6146, 5.6511], [5.4371, 5.5649, 5.5958], [5.5215, 5.5184, 5.5293]]]).to(_SCREAMING_SNAKE_CASE) self.assertTrue(torch.allclose(outputs.predicted_depth[:3, :3, :3] / 100 , _SCREAMING_SNAKE_CASE , atol=1e-4))
269
"""simple docstring""" import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('TEST_SAGEMAKER' , 'False' ) ) is not True , reason='Skipping test because should only be run when releasing minor transformers version' , ) @pytest.mark.usefixtures('sm_env' ) @parameterized_class( [ { 'framework': 'pytorch', 'script': 'run_glue.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_5_0, 'eval_accuracy': 0.6, 'eval_loss': 0.9}, }, { 'framework': 'tensorflow', 'script': 'run_tf.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_0_0, 'eval_accuracy': 0.3, 'eval_loss': 0.9}, }, ] ) class A__ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: int) -> Tuple: """simple docstring""" if self.framework == "pytorch": subprocess.run( F"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() , encoding="utf-8" , check=_SCREAMING_SNAKE_CASE , ) assert hasattr(self , "env") def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Any=1) -> Dict: """simple docstring""" return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=F"""{self.env.base_job_name}-single""" , instance_count=_SCREAMING_SNAKE_CASE , instance_type=self.instance_type , debugger_hook_config=_SCREAMING_SNAKE_CASE , hyperparameters={**self.env.hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version="py36" , ) def _SCREAMING_SNAKE_CASE ( self: str , _SCREAMING_SNAKE_CASE: List[Any]) -> Optional[Any]: """simple docstring""" TrainingJobAnalytics(_SCREAMING_SNAKE_CASE).export_csv(F"""{self.env.test_path}/{job_name}_metrics.csv""") def _SCREAMING_SNAKE_CASE ( self: str) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Tuple = self.create_estimator() # run training estimator.fit() # result dataframe __lowerCAmelCase : Tuple = TrainingJobAnalytics(estimator.latest_training_job.name).dataframe() # extract kpis __lowerCAmelCase : Optional[Any] = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"]) __lowerCAmelCase : Union[str, Any] = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"]) # get train time from SageMaker job, this includes starting, preprocessing, stopping __lowerCAmelCase : Tuple = ( Session().describe_training_job(estimator.latest_training_job.name).get("TrainingTimeInSeconds" , 99_9999) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy) assert all(t <= self.results["eval_loss"] for t in eval_loss) # dump tests result into json file to share in PR with open(F"""{estimator.latest_training_job.name}.json""" , "w") as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , _SCREAMING_SNAKE_CASE)
269
1
"""simple docstring""" from datasets.utils.patching import _PatchedModuleObj, patch_submodule from . import _test_patching def _lowercase ( ) -> Optional[Any]: import os as original_os from os import path as original_path from os import rename as original_rename from os.path import dirname as original_dirname from os.path import join as original_join assert _test_patching.os is original_os assert _test_patching.path is original_path assert _test_patching.join is original_join assert _test_patching.renamed_os is original_os assert _test_patching.renamed_path is original_path assert _test_patching.renamed_join is original_join __lowerCAmelCase : int = "__test_patch_submodule_mock__" with patch_submodule(_test_patching ,"os.path.join" ,__snake_case ): # Every way to access os.path.join must be patched, and the rest must stay untouched # check os.path.join assert isinstance(_test_patching.os ,_PatchedModuleObj ) assert isinstance(_test_patching.os.path ,_PatchedModuleObj ) assert _test_patching.os.path.join is mock # check path.join assert isinstance(_test_patching.path ,_PatchedModuleObj ) assert _test_patching.path.join is mock # check join assert _test_patching.join is mock # check that the other attributes are untouched assert _test_patching.os.rename is original_rename assert _test_patching.path.dirname is original_dirname assert _test_patching.os.path.dirname is original_dirname # Even renamed modules or objects must be patched # check renamed_os.path.join assert isinstance(_test_patching.renamed_os ,_PatchedModuleObj ) assert isinstance(_test_patching.renamed_os.path ,_PatchedModuleObj ) assert _test_patching.renamed_os.path.join is mock # check renamed_path.join assert isinstance(_test_patching.renamed_path ,_PatchedModuleObj ) assert _test_patching.renamed_path.join is mock # check renamed_join assert _test_patching.renamed_join is mock # check that the other attributes are untouched assert _test_patching.renamed_os.rename is original_rename assert _test_patching.renamed_path.dirname is original_dirname assert _test_patching.renamed_os.path.dirname is original_dirname # check that everthing is back to normal when the patch is over assert _test_patching.os is original_os assert _test_patching.path is original_path assert _test_patching.join is original_join assert _test_patching.renamed_os is original_os assert _test_patching.renamed_path is original_path assert _test_patching.renamed_join is original_join def _lowercase ( ) -> str: assert _test_patching.open is open __lowerCAmelCase : Union[str, Any] = "__test_patch_submodule_builtin_mock__" # _test_patching has "open" in its globals assert _test_patching.open is open with patch_submodule(_test_patching ,"open" ,__snake_case ): assert _test_patching.open is mock # check that everthing is back to normal when the patch is over assert _test_patching.open is open def _lowercase ( ) -> Tuple: # pandas.read_csv is not present in _test_patching __lowerCAmelCase : List[Any] = "__test_patch_submodule_missing_mock__" with patch_submodule(_test_patching ,"pandas.read_csv" ,__snake_case ): pass def _lowercase ( ) -> Tuple: # builtin should always be mocked even if they're not in the globals # in case they're loaded at one point __lowerCAmelCase : Optional[int] = "__test_patch_submodule_missing_builtin_mock__" # _test_patching doesn't have "len" in its globals assert getattr(_test_patching ,"len" ,__snake_case ) is None with patch_submodule(_test_patching ,"len" ,__snake_case ): assert _test_patching.len is mock assert _test_patching.len is len def _lowercase ( ) -> Optional[Any]: __lowerCAmelCase : int = "__test_patch_submodule_start_and_stop_mock__" __lowerCAmelCase : Tuple = patch_submodule(_test_patching ,"open" ,__snake_case ) assert _test_patching.open is open patch.start() assert _test_patching.open is mock patch.stop() assert _test_patching.open is open def _lowercase ( ) -> Optional[Any]: from os import rename as original_rename from os.path import dirname as original_dirname from os.path import join as original_join __lowerCAmelCase : Dict = "__test_patch_submodule_successive_join__" __lowerCAmelCase : Optional[int] = "__test_patch_submodule_successive_dirname__" __lowerCAmelCase : str = "__test_patch_submodule_successive_rename__" assert _test_patching.os.path.join is original_join assert _test_patching.os.path.dirname is original_dirname assert _test_patching.os.rename is original_rename with patch_submodule(_test_patching ,"os.path.join" ,__snake_case ): with patch_submodule(_test_patching ,"os.rename" ,__snake_case ): with patch_submodule(_test_patching ,"os.path.dirname" ,__snake_case ): assert _test_patching.os.path.join is mock_join assert _test_patching.os.path.dirname is mock_dirname assert _test_patching.os.rename is mock_rename # try another order with patch_submodule(_test_patching ,"os.rename" ,__snake_case ): with patch_submodule(_test_patching ,"os.path.join" ,__snake_case ): with patch_submodule(_test_patching ,"os.path.dirname" ,__snake_case ): assert _test_patching.os.path.join is mock_join assert _test_patching.os.path.dirname is mock_dirname assert _test_patching.os.rename is mock_rename assert _test_patching.os.path.join is original_join assert _test_patching.os.path.dirname is original_dirname assert _test_patching.os.rename is original_rename def _lowercase ( ) -> Tuple: __lowerCAmelCase : Optional[int] = "__test_patch_submodule_doesnt_exist_mock__" with patch_submodule(_test_patching ,"__module_that_doesn_exist__.__attribute_that_doesn_exist__" ,__snake_case ): pass with patch_submodule(_test_patching ,"os.__attribute_that_doesn_exist__" ,__snake_case ): pass
269
"""simple docstring""" from typing import TYPE_CHECKING from ..utils import _LazyModule __snake_case : Optional[int] = { 'config': [ 'EXTERNAL_DATA_FORMAT_SIZE_LIMIT', 'OnnxConfig', 'OnnxConfigWithPast', 'OnnxSeq2SeqConfigWithPast', 'PatchingSpec', ], 'convert': ['export', 'validate_model_outputs'], 'features': ['FeaturesManager'], 'utils': ['ParameterFormat', 'compute_serialized_parameters_size'], } if TYPE_CHECKING: from .config import ( EXTERNAL_DATA_FORMAT_SIZE_LIMIT, OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast, PatchingSpec, ) from .convert import export, validate_model_outputs from .features import FeaturesManager from .utils import ParameterFormat, compute_serialized_parameters_size else: import sys __snake_case : Any = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
269
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __snake_case : Tuple = {'configuration_wavlm': ['WAVLM_PRETRAINED_CONFIG_ARCHIVE_MAP', 'WavLMConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : Optional[Any] = [ 'WAVLM_PRETRAINED_MODEL_ARCHIVE_LIST', 'WavLMForAudioFrameClassification', 'WavLMForCTC', 'WavLMForSequenceClassification', 'WavLMForXVector', 'WavLMModel', 'WavLMPreTrainedModel', ] if TYPE_CHECKING: from .configuration_wavlm import WAVLM_PRETRAINED_CONFIG_ARCHIVE_MAP, WavLMConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_wavlm import ( WAVLM_PRETRAINED_MODEL_ARCHIVE_LIST, WavLMForAudioFrameClassification, WavLMForCTC, WavLMForSequenceClassification, WavLMForXVector, WavLMModel, WavLMPreTrainedModel, ) else: import sys __snake_case : Union[str, Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
269
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices __snake_case : Optional[int] = logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'maskformer-swin' SCREAMING_SNAKE_CASE = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self: Optional[int] , _SCREAMING_SNAKE_CASE: int=224 , _SCREAMING_SNAKE_CASE: Tuple=4 , _SCREAMING_SNAKE_CASE: int=3 , _SCREAMING_SNAKE_CASE: List[Any]=96 , _SCREAMING_SNAKE_CASE: Union[str, Any]=[2, 2, 6, 2] , _SCREAMING_SNAKE_CASE: Any=[3, 6, 12, 24] , _SCREAMING_SNAKE_CASE: List[str]=7 , _SCREAMING_SNAKE_CASE: List[str]=4.0 , _SCREAMING_SNAKE_CASE: Optional[int]=True , _SCREAMING_SNAKE_CASE: Tuple=0.0 , _SCREAMING_SNAKE_CASE: Any=0.0 , _SCREAMING_SNAKE_CASE: Any=0.1 , _SCREAMING_SNAKE_CASE: str="gelu" , _SCREAMING_SNAKE_CASE: Tuple=False , _SCREAMING_SNAKE_CASE: Union[str, Any]=0.02 , _SCREAMING_SNAKE_CASE: str=1e-5 , _SCREAMING_SNAKE_CASE: Optional[int]=None , _SCREAMING_SNAKE_CASE: str=None , **_SCREAMING_SNAKE_CASE: Union[str, Any] , ) -> List[str]: """simple docstring""" super().__init__(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = image_size __lowerCAmelCase : Any = patch_size __lowerCAmelCase : Tuple = num_channels __lowerCAmelCase : Any = embed_dim __lowerCAmelCase : Any = depths __lowerCAmelCase : Dict = len(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = num_heads __lowerCAmelCase : Tuple = window_size __lowerCAmelCase : Dict = mlp_ratio __lowerCAmelCase : Any = qkv_bias __lowerCAmelCase : Union[str, Any] = hidden_dropout_prob __lowerCAmelCase : int = attention_probs_dropout_prob __lowerCAmelCase : Tuple = drop_path_rate __lowerCAmelCase : int = hidden_act __lowerCAmelCase : Optional[int] = use_absolute_embeddings __lowerCAmelCase : List[str] = layer_norm_eps __lowerCAmelCase : Any = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model __lowerCAmelCase : Optional[Any] = int(embed_dim * 2 ** (len(_SCREAMING_SNAKE_CASE) - 1)) __lowerCAmelCase : Any = ["stem"] + [F"""stage{idx}""" for idx in range(1 , len(_SCREAMING_SNAKE_CASE) + 1)] __lowerCAmelCase , __lowerCAmelCase : List[str] = get_aligned_output_features_output_indices( out_features=_SCREAMING_SNAKE_CASE , out_indices=_SCREAMING_SNAKE_CASE , stage_names=self.stage_names)
269
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case : List[Any] = logging.get_logger(__name__) __snake_case : List[str] = { 'sayakpaul/vit-msn-base': 'https://huggingface.co/sayakpaul/vit-msn-base/resolve/main/config.json', # See all ViT MSN models at https://huggingface.co/models?filter=vit_msn } class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'vit_msn' def __init__( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: int=768 , _SCREAMING_SNAKE_CASE: List[str]=12 , _SCREAMING_SNAKE_CASE: int=12 , _SCREAMING_SNAKE_CASE: int=3072 , _SCREAMING_SNAKE_CASE: Union[str, Any]="gelu" , _SCREAMING_SNAKE_CASE: Dict=0.0 , _SCREAMING_SNAKE_CASE: Any=0.0 , _SCREAMING_SNAKE_CASE: List[str]=0.02 , _SCREAMING_SNAKE_CASE: str=1e-06 , _SCREAMING_SNAKE_CASE: Optional[Any]=224 , _SCREAMING_SNAKE_CASE: Tuple=16 , _SCREAMING_SNAKE_CASE: Optional[Any]=3 , _SCREAMING_SNAKE_CASE: Union[str, Any]=True , **_SCREAMING_SNAKE_CASE: List[str] , ) -> List[Any]: """simple docstring""" super().__init__(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = hidden_size __lowerCAmelCase : Union[str, Any] = num_hidden_layers __lowerCAmelCase : Dict = num_attention_heads __lowerCAmelCase : Union[str, Any] = intermediate_size __lowerCAmelCase : Optional[int] = hidden_act __lowerCAmelCase : List[str] = hidden_dropout_prob __lowerCAmelCase : List[Any] = attention_probs_dropout_prob __lowerCAmelCase : List[Any] = initializer_range __lowerCAmelCase : Any = layer_norm_eps __lowerCAmelCase : Dict = image_size __lowerCAmelCase : Optional[int] = patch_size __lowerCAmelCase : Any = num_channels __lowerCAmelCase : Optional[int] = qkv_bias
269
"""simple docstring""" import gc import threading import time import psutil import torch class A__ : '''simple docstring''' def __init__( self: str) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Optional[Any] = psutil.Process() __lowerCAmelCase : str = False def _SCREAMING_SNAKE_CASE ( self: int) -> Optional[int]: """simple docstring""" __lowerCAmelCase : Optional[Any] = -1 while True: __lowerCAmelCase : str = max(self.process.memory_info().rss , self.cpu_memory_peak) # can't sleep or will not catch the peak right (this comment is here on purpose) if not self.peak_monitoring: break def _SCREAMING_SNAKE_CASE ( self: Tuple) -> int: """simple docstring""" __lowerCAmelCase : List[str] = True __lowerCAmelCase : str = threading.Thread(target=self.peak_monitor) __lowerCAmelCase : Tuple = True self.thread.start() def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> List[str]: """simple docstring""" __lowerCAmelCase : Tuple = False self.thread.join() return self.cpu_memory_peak __snake_case : Tuple = PeakCPUMemory() def _lowercase ( ) -> str: # Time __lowerCAmelCase : str = {"time": time.time()} gc.collect() torch.cuda.empty_cache() # CPU mem __lowerCAmelCase : Optional[Any] = psutil.Process().memory_info().rss cpu_peak_tracker.start() # GPU mem for i in range(torch.cuda.device_count() ): __lowerCAmelCase : Union[str, Any] = torch.cuda.memory_allocated(__snake_case ) torch.cuda.reset_peak_memory_stats() return measures def _lowercase ( __snake_case ) -> Optional[Any]: # Time __lowerCAmelCase : str = {"time": time.time() - start_measures["time"]} gc.collect() torch.cuda.empty_cache() # CPU mem __lowerCAmelCase : str = (psutil.Process().memory_info().rss - start_measures["cpu"]) / 2**20 __lowerCAmelCase : List[str] = (cpu_peak_tracker.stop() - start_measures["cpu"]) / 2**20 # GPU mem for i in range(torch.cuda.device_count() ): __lowerCAmelCase : Union[str, Any] = (torch.cuda.memory_allocated(__snake_case ) - start_measures[str(__snake_case )]) / 2**20 __lowerCAmelCase : Any = (torch.cuda.max_memory_allocated(__snake_case ) - start_measures[str(__snake_case )]) / 2**20 return measures def _lowercase ( __snake_case ,__snake_case ) -> Dict: print(F"""{description}:""" ) print(F"""- Time: {measures['time']:.2f}s""" ) for i in range(torch.cuda.device_count() ): print(F"""- GPU {i} allocated: {measures[str(__snake_case )]:.2f}MiB""" ) __lowerCAmelCase : Optional[Any] = measures[F"""{i}-peak"""] print(F"""- GPU {i} peak: {peak:.2f}MiB""" ) print(F"""- CPU RAM allocated: {measures['cpu']:.2f}MiB""" ) print(F"""- CPU RAM peak: {measures['cpu-peak']:.2f}MiB""" )
269
1
"""simple docstring""" import argparse import torch from transformers import ( SpeechTaConfig, SpeechTaFeatureExtractor, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaProcessor, SpeechTaTokenizer, logging, ) from transformers.tokenization_utils import AddedToken logging.set_verbosity_info() __snake_case : Tuple = logging.get_logger('transformers.models.speecht5') __snake_case : Any = { 'speech_encoder_prenet.layer_norm': 'speecht5.encoder.prenet.feature_projection.layer_norm', 'speech_encoder_prenet.post_extract_proj': 'speecht5.encoder.prenet.feature_projection.projection', 'speech_encoder_prenet.pos_conv.0': 'speecht5.encoder.prenet.pos_conv_embed.conv', 'speech_encoder_prenet.mask_emb': 'speecht5.encoder.prenet.masked_spec_embed', } __snake_case : Union[str, Any] = { 'text_encoder_prenet.encoder_prenet.0': 'speecht5.encoder.prenet.embed_tokens', 'text_encoder_prenet.encoder_prenet.1.alpha': 'speecht5.encoder.prenet.encode_positions.alpha', } __snake_case : Union[str, Any] = { 'speech_decoder_prenet.decoder_prenet.0.0.prenet.0.0': 'speecht5.decoder.prenet.layers.0', 'speech_decoder_prenet.decoder_prenet.0.0.prenet.1.0': 'speecht5.decoder.prenet.layers.1', 'speech_decoder_prenet.decoder_prenet.0.1': 'speecht5.decoder.prenet.final_layer', 'speech_decoder_prenet.decoder_prenet.1.alpha': 'speecht5.decoder.prenet.encode_positions.alpha', 'speech_decoder_prenet.spkembs_layer.0': 'speecht5.decoder.prenet.speaker_embeds_layer', } __snake_case : Dict = { 'speech_decoder_postnet.feat_out': 'speech_decoder_postnet.feat_out', 'speech_decoder_postnet.prob_out': 'speech_decoder_postnet.prob_out', 'speech_decoder_postnet.postnet.postnet.0.0': 'speech_decoder_postnet.layers.0.conv', 'speech_decoder_postnet.postnet.postnet.0.1': 'speech_decoder_postnet.layers.0.batch_norm', 'speech_decoder_postnet.postnet.postnet.1.0': 'speech_decoder_postnet.layers.1.conv', 'speech_decoder_postnet.postnet.postnet.1.1': 'speech_decoder_postnet.layers.1.batch_norm', 'speech_decoder_postnet.postnet.postnet.2.0': 'speech_decoder_postnet.layers.2.conv', 'speech_decoder_postnet.postnet.postnet.2.1': 'speech_decoder_postnet.layers.2.batch_norm', 'speech_decoder_postnet.postnet.postnet.3.0': 'speech_decoder_postnet.layers.3.conv', 'speech_decoder_postnet.postnet.postnet.3.1': 'speech_decoder_postnet.layers.3.batch_norm', 'speech_decoder_postnet.postnet.postnet.4.0': 'speech_decoder_postnet.layers.4.conv', 'speech_decoder_postnet.postnet.postnet.4.1': 'speech_decoder_postnet.layers.4.batch_norm', } __snake_case : Optional[int] = { 'text_decoder_prenet.embed_tokens': 'speecht5.decoder.prenet.embed_tokens', } __snake_case : Union[str, Any] = { 'text_decoder_postnet.output_projection': 'text_decoder_postnet.lm_head', } __snake_case : Tuple = { 'encoder.layers.*.self_attn.k_proj': 'speecht5.encoder.wrapped_encoder.layers.*.attention.k_proj', 'encoder.layers.*.self_attn.v_proj': 'speecht5.encoder.wrapped_encoder.layers.*.attention.v_proj', 'encoder.layers.*.self_attn.q_proj': 'speecht5.encoder.wrapped_encoder.layers.*.attention.q_proj', 'encoder.layers.*.self_attn.out_proj': 'speecht5.encoder.wrapped_encoder.layers.*.attention.out_proj', 'encoder.layers.*.self_attn_layer_norm': 'speecht5.encoder.wrapped_encoder.layers.*.layer_norm', 'encoder.layers.*.fc1': 'speecht5.encoder.wrapped_encoder.layers.*.feed_forward.intermediate_dense', 'encoder.layers.*.fc2': 'speecht5.encoder.wrapped_encoder.layers.*.feed_forward.output_dense', 'encoder.layers.*.final_layer_norm': 'speecht5.encoder.wrapped_encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'speecht5.encoder.wrapped_encoder.layer_norm', 'encoder.pos_emb.pe_k': 'speecht5.encoder.wrapped_encoder.embed_positions.pe_k', } __snake_case : str = { 'decoder.layers.*.self_attn.k_proj': 'speecht5.decoder.wrapped_decoder.layers.*.self_attn.k_proj', 'decoder.layers.*.self_attn.v_proj': 'speecht5.decoder.wrapped_decoder.layers.*.self_attn.v_proj', 'decoder.layers.*.self_attn.q_proj': 'speecht5.decoder.wrapped_decoder.layers.*.self_attn.q_proj', 'decoder.layers.*.self_attn.out_proj': 'speecht5.decoder.wrapped_decoder.layers.*.self_attn.out_proj', 'decoder.layers.*.self_attn_layer_norm': 'speecht5.decoder.wrapped_decoder.layers.*.self_attn_layer_norm', 'decoder.layers.*.encoder_attn.k_proj': 'speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.k_proj', 'decoder.layers.*.encoder_attn.v_proj': 'speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.v_proj', 'decoder.layers.*.encoder_attn.q_proj': 'speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.q_proj', 'decoder.layers.*.encoder_attn.out_proj': 'speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.out_proj', 'decoder.layers.*.encoder_attn_layer_norm': 'speecht5.decoder.wrapped_decoder.layers.*.encoder_attn_layer_norm', 'decoder.layers.*.fc1': 'speecht5.decoder.wrapped_decoder.layers.*.feed_forward.intermediate_dense', 'decoder.layers.*.fc2': 'speecht5.decoder.wrapped_decoder.layers.*.feed_forward.output_dense', 'decoder.layers.*.final_layer_norm': 'speecht5.decoder.wrapped_decoder.layers.*.final_layer_norm', } __snake_case : Union[str, Any] = { **MAPPING_SPEECH_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_TEXT_DECODER_PRENET, **MAPPING_TEXT_DECODER_POSTNET, } __snake_case : str = { **MAPPING_TEXT_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_SPEECH_DECODER_PRENET, **MAPPING_SPEECH_DECODER_POSTNET, } __snake_case : str = { **MAPPING_SPEECH_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_SPEECH_DECODER_PRENET, **MAPPING_SPEECH_DECODER_POSTNET, } __snake_case : Optional[int] = [] __snake_case : str = [ 'encoder.version', 'encoder.layers.*.norm_k.weight', 'encoder.layers.*.norm_k.bias', 'decoder.version', 'decoder.layers.*.norm_k.weight', 'decoder.layers.*.norm_k.bias', 'decoder.pos_emb.pe_k', 'speech_encoder_prenet.embed_positions._float_tensor', 'text_decoder_prenet.embed_positions._float_tensor', ] __snake_case : Any = IGNORE_KEYS + [ 'encoder.proj', 'text_encoder_prenet.*', 'speech_decoder_prenet.*', 'speech_decoder_postnet.*', ] __snake_case : Optional[int] = IGNORE_KEYS + [ 'encoder.proj', 'speech_encoder_prenet.*', 'text_decoder_prenet.*', 'text_decoder_postnet.*', ] __snake_case : Dict = IGNORE_KEYS + [ 'encoder.proj', 'text_encoder_prenet.*', 'text_decoder_prenet.*', 'text_decoder_postnet.*', ] def _lowercase ( __snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ) -> Dict: for attribute in key.split("." ): __lowerCAmelCase : Tuple = getattr(__snake_case ,__snake_case ) if weight_type is not None: __lowerCAmelCase : Tuple = getattr(__snake_case ,__snake_case ).shape else: __lowerCAmelCase : Optional[Any] = hf_pointer.shape if hf_shape != value.shape: raise ValueError( F"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" F""" {value.shape} for {full_name}""" ) if weight_type == "weight": __lowerCAmelCase : Optional[int] = value elif weight_type == "weight_g": __lowerCAmelCase : Tuple = value elif weight_type == "weight_v": __lowerCAmelCase : str = value elif weight_type == "bias": __lowerCAmelCase : int = value elif weight_type == "running_mean": __lowerCAmelCase : List[Any] = value elif weight_type == "running_var": __lowerCAmelCase : Union[str, Any] = value elif weight_type == "num_batches_tracked": __lowerCAmelCase : Union[str, Any] = value else: __lowerCAmelCase : List[str] = value logger.info(F"""{key + ('.' + weight_type if weight_type is not None else '')} was initialized from {full_name}.""" ) def _lowercase ( __snake_case ,__snake_case ) -> Optional[int]: for key in ignore_keys: if key.endswith(".*" ): if name.startswith(key[:-1] ): return True elif ".*." in key: __lowerCAmelCase , __lowerCAmelCase : Any = key.split(".*." ) if prefix in name and suffix in name: return True elif key in name: return True return False def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> int: __lowerCAmelCase : Optional[Any] = [] if task == "s2t": __lowerCAmelCase : List[str] = hf_model.speechta.encoder.prenet.feature_encoder __lowerCAmelCase : Tuple = MAPPING_S2T __lowerCAmelCase : List[Any] = IGNORE_KEYS_S2T elif task == "t2s": __lowerCAmelCase : Any = None __lowerCAmelCase : List[Any] = MAPPING_T2S __lowerCAmelCase : Optional[int] = IGNORE_KEYS_T2S elif task == "s2s": __lowerCAmelCase : str = hf_model.speechta.encoder.prenet.feature_encoder __lowerCAmelCase : Any = MAPPING_S2S __lowerCAmelCase : Any = IGNORE_KEYS_S2S else: raise ValueError(F"""Unsupported task: {task}""" ) for name, value in fairseq_dict.items(): if should_ignore(__snake_case ,__snake_case ): logger.info(F"""{name} was ignored""" ) continue __lowerCAmelCase : Union[str, Any] = False if "conv_layers" in name: load_conv_layer( __snake_case ,__snake_case ,__snake_case ,__snake_case ,hf_model.config.feat_extract_norm == "group" ,) __lowerCAmelCase : List[Any] = True else: for key, mapped_key in MAPPING.items(): # mapped_key = "speecht5." + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if "*" in key: __lowerCAmelCase , __lowerCAmelCase : List[str] = key.split(".*." ) if prefix in name and suffix in name: __lowerCAmelCase : Optional[Any] = suffix # if key in name or key.split("w2v_model.")[-1] == name.split(".")[0]: if key in name: __lowerCAmelCase : Union[str, Any] = True if "*" in mapped_key: __lowerCAmelCase : Tuple = name.split(__snake_case )[0].split("." )[-2] __lowerCAmelCase : int = mapped_key.replace("*" ,__snake_case ) if "weight_g" in name: __lowerCAmelCase : List[str] = "weight_g" elif "weight_v" in name: __lowerCAmelCase : List[Any] = "weight_v" elif "bias" in name: __lowerCAmelCase : List[str] = "bias" elif "weight" in name: __lowerCAmelCase : List[str] = "weight" elif "running_mean" in name: __lowerCAmelCase : Optional[Any] = "running_mean" elif "running_var" in name: __lowerCAmelCase : Optional[int] = "running_var" elif "num_batches_tracked" in name: __lowerCAmelCase : Optional[int] = "num_batches_tracked" else: __lowerCAmelCase : Dict = None set_recursively(__snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ) continue if not is_used: unused_weights.append(__snake_case ) logger.warning(F"""Unused weights: {unused_weights}""" ) def _lowercase ( __snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ) -> Tuple: __lowerCAmelCase : Any = full_name.split("conv_layers." )[-1] __lowerCAmelCase : List[str] = name.split("." ) __lowerCAmelCase : int = int(items[0] ) __lowerCAmelCase : Optional[Any] = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) __lowerCAmelCase : Any = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) __lowerCAmelCase : Optional[Any] = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.""" ) __lowerCAmelCase : Dict = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.""" ) __lowerCAmelCase : Union[str, Any] = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(__snake_case ) @torch.no_grad() def _lowercase ( __snake_case ,__snake_case ,__snake_case ,__snake_case=None ,__snake_case=None ,__snake_case=None ,) -> Tuple: if config_path is not None: __lowerCAmelCase : int = SpeechTaConfig.from_pretrained(__snake_case ) else: __lowerCAmelCase : Union[str, Any] = SpeechTaConfig() if task == "s2t": __lowerCAmelCase : int = config.max_text_positions __lowerCAmelCase : Tuple = SpeechTaForSpeechToText(__snake_case ) elif task == "t2s": __lowerCAmelCase : Optional[Any] = 1_876 __lowerCAmelCase : Optional[int] = 600 __lowerCAmelCase : Any = config.max_speech_positions __lowerCAmelCase : str = SpeechTaForTextToSpeech(__snake_case ) elif task == "s2s": __lowerCAmelCase : str = 1_876 __lowerCAmelCase : Optional[int] = config.max_speech_positions __lowerCAmelCase : Union[str, Any] = SpeechTaForSpeechToSpeech(__snake_case ) else: raise ValueError(F"""Unknown task name: {task}""" ) if vocab_path: __lowerCAmelCase : Any = SpeechTaTokenizer(__snake_case ,model_max_length=config.max_text_positions ) # Mask token behaves like a normal word, i.e. include the space before it __lowerCAmelCase : Tuple = AddedToken("<mask>" ,lstrip=__snake_case ,rstrip=__snake_case ) __lowerCAmelCase : Union[str, Any] = mask_token tokenizer.add_special_tokens({"mask_token": mask_token} ) tokenizer.add_tokens(["<ctc_blank>"] ) __lowerCAmelCase : List[str] = SpeechTaFeatureExtractor() __lowerCAmelCase : List[Any] = SpeechTaProcessor(tokenizer=__snake_case ,feature_extractor=__snake_case ) processor.save_pretrained(__snake_case ) __lowerCAmelCase : List[Any] = torch.load(__snake_case ) recursively_load_weights(fairseq_checkpoint["model"] ,__snake_case ,__snake_case ) model.save_pretrained(__snake_case ) if repo_id: print("Pushing to the hub..." ) processor.push_to_hub(__snake_case ) model.push_to_hub(__snake_case ) if __name__ == "__main__": __snake_case : Any = argparse.ArgumentParser() parser.add_argument( '--task', default='s2t', type=str, help='Type of the SpeechT5 model you\'d like to convert. Should be one of \'s2t\', \'t2s\', \'s2s\'.', ) parser.add_argument('--checkpoint_path', required=True, default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--vocab_path', default=None, type=str, help='Path to SentencePiece model') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--pytorch_dump_folder_path', required=True, default=None, type=str, help='Path to the output PyTorch model.' ) parser.add_argument( '--push_to_hub', default=None, type=str, help='Where to upload the converted model on the 🤗 hub.' ) __snake_case : Any = parser.parse_args() convert_speechta_checkpoint( args.task, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.vocab_path, args.push_to_hub, )
269
"""simple docstring""" from ....configuration_utils import PretrainedConfig from ....utils import logging __snake_case : List[Any] = logging.get_logger(__name__) __snake_case : Any = { 'speechbrain/m-ctc-t-large': 'https://huggingface.co/speechbrain/m-ctc-t-large/resolve/main/config.json', # See all M-CTC-T models at https://huggingface.co/models?filter=mctct } class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'mctct' def __init__( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: str=8065 , _SCREAMING_SNAKE_CASE: str=1536 , _SCREAMING_SNAKE_CASE: str=36 , _SCREAMING_SNAKE_CASE: Optional[Any]=6144 , _SCREAMING_SNAKE_CASE: Optional[Any]=4 , _SCREAMING_SNAKE_CASE: Union[str, Any]=384 , _SCREAMING_SNAKE_CASE: Optional[Any]=920 , _SCREAMING_SNAKE_CASE: Union[str, Any]=1e-5 , _SCREAMING_SNAKE_CASE: List[Any]=0.3 , _SCREAMING_SNAKE_CASE: Optional[Any]="relu" , _SCREAMING_SNAKE_CASE: Optional[int]=0.02 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.3 , _SCREAMING_SNAKE_CASE: Dict=0.3 , _SCREAMING_SNAKE_CASE: List[Any]=1 , _SCREAMING_SNAKE_CASE: Optional[Any]=0 , _SCREAMING_SNAKE_CASE: List[str]=2 , _SCREAMING_SNAKE_CASE: Union[str, Any]=1 , _SCREAMING_SNAKE_CASE: Tuple=0.3 , _SCREAMING_SNAKE_CASE: Dict=1 , _SCREAMING_SNAKE_CASE: int=(7,) , _SCREAMING_SNAKE_CASE: str=(3,) , _SCREAMING_SNAKE_CASE: Union[str, Any]=80 , _SCREAMING_SNAKE_CASE: Tuple=1 , _SCREAMING_SNAKE_CASE: Dict=None , _SCREAMING_SNAKE_CASE: Tuple="sum" , _SCREAMING_SNAKE_CASE: List[str]=False , **_SCREAMING_SNAKE_CASE: Tuple , ) -> Tuple: """simple docstring""" super().__init__(**_SCREAMING_SNAKE_CASE , pad_token_id=_SCREAMING_SNAKE_CASE , bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = vocab_size __lowerCAmelCase : str = hidden_size __lowerCAmelCase : str = num_hidden_layers __lowerCAmelCase : str = intermediate_size __lowerCAmelCase : List[Any] = num_attention_heads __lowerCAmelCase : Dict = attention_head_dim __lowerCAmelCase : Optional[int] = max_position_embeddings __lowerCAmelCase : str = layer_norm_eps __lowerCAmelCase : Tuple = layerdrop __lowerCAmelCase : str = hidden_act __lowerCAmelCase : List[Any] = initializer_range __lowerCAmelCase : int = hidden_dropout_prob __lowerCAmelCase : Optional[Any] = attention_probs_dropout_prob __lowerCAmelCase : str = pad_token_id __lowerCAmelCase : Optional[int] = bos_token_id __lowerCAmelCase : Union[str, Any] = eos_token_id __lowerCAmelCase : Any = conv_glu_dim __lowerCAmelCase : Optional[int] = conv_dropout __lowerCAmelCase : Union[str, Any] = num_conv_layers __lowerCAmelCase : Optional[int] = input_feat_per_channel __lowerCAmelCase : Union[str, Any] = input_channels __lowerCAmelCase : Optional[Any] = conv_channels __lowerCAmelCase : Dict = ctc_loss_reduction __lowerCAmelCase : int = ctc_zero_infinity # prevents config testing fail with exporting to json __lowerCAmelCase : List[str] = list(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = list(_SCREAMING_SNAKE_CASE) if len(self.conv_kernel) != self.num_conv_layers: raise ValueError( "Configuration for convolutional module is incorrect. " "It is required that `len(config.conv_kernel)` == `config.num_conv_layers` " F"""but is `len(config.conv_kernel) = {len(self.conv_kernel)}`, """ F"""`config.num_conv_layers = {self.num_conv_layers}`.""")
269
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tensorflow_text_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __snake_case : Optional[int] = { 'configuration_bert': ['BERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BertConfig', 'BertOnnxConfig'], 'tokenization_bert': ['BasicTokenizer', 'BertTokenizer', 'WordpieceTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : Any = ['BertTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : Any = [ 'BERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'BertForMaskedLM', 'BertForMultipleChoice', 'BertForNextSentencePrediction', 'BertForPreTraining', 'BertForQuestionAnswering', 'BertForSequenceClassification', 'BertForTokenClassification', 'BertLayer', 'BertLMHeadModel', 'BertModel', 'BertPreTrainedModel', 'load_tf_weights_in_bert', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : Optional[int] = [ 'TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFBertEmbeddings', 'TFBertForMaskedLM', 'TFBertForMultipleChoice', 'TFBertForNextSentencePrediction', 'TFBertForPreTraining', 'TFBertForQuestionAnswering', 'TFBertForSequenceClassification', 'TFBertForTokenClassification', 'TFBertLMHeadModel', 'TFBertMainLayer', 'TFBertModel', 'TFBertPreTrainedModel', ] try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : Optional[int] = ['TFBertTokenizer'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : Dict = [ 'FlaxBertForCausalLM', 'FlaxBertForMaskedLM', 'FlaxBertForMultipleChoice', 'FlaxBertForNextSentencePrediction', 'FlaxBertForPreTraining', 'FlaxBertForQuestionAnswering', 'FlaxBertForSequenceClassification', 'FlaxBertForTokenClassification', 'FlaxBertModel', 'FlaxBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_bert import BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BertConfig, BertOnnxConfig from .tokenization_bert import BasicTokenizer, BertTokenizer, WordpieceTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_fast import BertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bert import ( BERT_PRETRAINED_MODEL_ARCHIVE_LIST, BertForMaskedLM, BertForMultipleChoice, BertForNextSentencePrediction, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertForTokenClassification, BertLayer, BertLMHeadModel, BertModel, BertPreTrainedModel, load_tf_weights_in_bert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_bert import ( TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFBertEmbeddings, TFBertForMaskedLM, TFBertForMultipleChoice, TFBertForNextSentencePrediction, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertForTokenClassification, TFBertLMHeadModel, TFBertMainLayer, TFBertModel, TFBertPreTrainedModel, ) try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_tf import TFBertTokenizer try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_bert import ( FlaxBertForCausalLM, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, FlaxBertPreTrainedModel, ) else: import sys __snake_case : Optional[int] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
269
"""simple docstring""" from __future__ import annotations import time import numpy as np __snake_case : Optional[Any] = [8, 5, 9, 7] __snake_case : List[Any] = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] __snake_case : Optional[int] = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class A__ : '''simple docstring''' def __init__( self: Any , _SCREAMING_SNAKE_CASE: list[int] , _SCREAMING_SNAKE_CASE: list[list[int]] , _SCREAMING_SNAKE_CASE: list[list[int]] , ) -> None: """simple docstring""" __lowerCAmelCase : Any = claim_vector __lowerCAmelCase : Tuple = allocated_resources_table __lowerCAmelCase : Tuple = maximum_claim_table def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> list[int]: """simple docstring""" return [ sum(p_item[i] for p_item in self.__allocated_resources_table) for i in range(len(self.__allocated_resources_table[0])) ] def _SCREAMING_SNAKE_CASE ( self: int) -> list[int]: """simple docstring""" return np.array(self.__claim_vector) - np.array( self.__processes_resource_summation()) def _SCREAMING_SNAKE_CASE ( self: int) -> list[list[int]]: """simple docstring""" return [ list(np.array(self.__maximum_claim_table[i]) - np.array(_SCREAMING_SNAKE_CASE)) for i, allocated_resource in enumerate(self.__allocated_resources_table) ] def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> dict[int, list[int]]: """simple docstring""" return {self.__need().index(_SCREAMING_SNAKE_CASE): i for i in self.__need()} def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , **_SCREAMING_SNAKE_CASE: List[Any]) -> None: """simple docstring""" __lowerCAmelCase : Optional[int] = self.__need() __lowerCAmelCase : int = self.__allocated_resources_table __lowerCAmelCase : Dict = self.__available_resources() __lowerCAmelCase : str = self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print("_" * 50 + "\n") while need_list: __lowerCAmelCase : int = False for each_need in need_list: __lowerCAmelCase : Dict = True for index, need in enumerate(_SCREAMING_SNAKE_CASE): if need > available_resources[index]: __lowerCAmelCase : Dict = False break if execution: __lowerCAmelCase : Any = True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: __lowerCAmelCase : Union[str, Any] = original_need_index print(F"""Process {process_number + 1} is executing.""") # remove the process run from stack need_list.remove(_SCREAMING_SNAKE_CASE) # update available/freed resources stack __lowerCAmelCase : Dict = np.array(_SCREAMING_SNAKE_CASE) + np.array( alloc_resources_table[process_number]) print( "Updated available resource stack for processes: " + " ".join([str(_SCREAMING_SNAKE_CASE) for x in available_resources])) break if safe: print("The process is in a safe state.\n") else: print("System in unsafe state. Aborting...\n") break def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> List[Any]: """simple docstring""" print(" " * 9 + "Allocated Resource Table") for item in self.__allocated_resources_table: print( F"""P{self.__allocated_resources_table.index(_SCREAMING_SNAKE_CASE) + 1}""" + " ".join(F"""{it:>8}""" for it in item) + "\n") print(" " * 9 + "System Resource Table") for item in self.__maximum_claim_table: print( F"""P{self.__maximum_claim_table.index(_SCREAMING_SNAKE_CASE) + 1}""" + " ".join(F"""{it:>8}""" for it in item) + "\n") print( "Current Usage by Active Processes: " + " ".join(str(_SCREAMING_SNAKE_CASE) for x in self.__claim_vector)) print( "Initial Available Resources: " + " ".join(str(_SCREAMING_SNAKE_CASE) for x in self.__available_resources())) time.sleep(1) if __name__ == "__main__": import doctest doctest.testmod()
269
1
"""simple docstring""" import argparse import os import re import packaging.version __snake_case : List[str] = 'examples/' __snake_case : int = { 'examples': (re.compile(R'^check_min_version\("[^"]+"\)\s*$', re.MULTILINE), 'check_min_version("VERSION")\n'), 'init': (re.compile(R'^__version__\s+=\s+"([^"]+)"\s*$', re.MULTILINE), '__version__ = "VERSION"\n'), 'setup': (re.compile(R'^(\s*)version\s*=\s*"[^"]+",', re.MULTILINE), R'\1version="VERSION",'), 'doc': (re.compile(R'^(\s*)release\s*=\s*"[^"]+"$', re.MULTILINE), 'release = "VERSION"\n'), } __snake_case : List[Any] = { 'init': 'src/transformers/__init__.py', 'setup': 'setup.py', } __snake_case : Any = 'README.md' def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> Optional[int]: with open(__snake_case ,"r" ,encoding="utf-8" ,newline="\n" ) as f: __lowerCAmelCase : List[str] = f.read() __lowerCAmelCase , __lowerCAmelCase : Optional[int] = REPLACE_PATTERNS[pattern] __lowerCAmelCase : Optional[int] = replace.replace("VERSION" ,__snake_case ) __lowerCAmelCase : Tuple = re_pattern.sub(__snake_case ,__snake_case ) with open(__snake_case ,"w" ,encoding="utf-8" ,newline="\n" ) as f: f.write(__snake_case ) def _lowercase ( __snake_case ) -> Optional[int]: for folder, directories, fnames in os.walk(__snake_case ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove("research_projects" ) if "legacy" in directories: directories.remove("legacy" ) for fname in fnames: if fname.endswith(".py" ): update_version_in_file(os.path.join(__snake_case ,__snake_case ) ,__snake_case ,pattern="examples" ) def _lowercase ( __snake_case ,__snake_case=False ) -> List[Any]: for pattern, fname in REPLACE_FILES.items(): update_version_in_file(__snake_case ,__snake_case ,__snake_case ) if not patch: update_version_in_examples(__snake_case ) def _lowercase ( ) -> str: __lowerCAmelCase : Tuple = "🤗 Transformers currently provides the following architectures" __lowerCAmelCase : Tuple = "1. Want to contribute a new model?" with open(__snake_case ,"r" ,encoding="utf-8" ,newline="\n" ) as f: __lowerCAmelCase : str = f.readlines() # Find the start of the list. __lowerCAmelCase : int = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 __lowerCAmelCase : Union[str, Any] = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith("1." ): __lowerCAmelCase : Any = lines[index].replace( "https://huggingface.co/docs/transformers/main/model_doc" ,"https://huggingface.co/docs/transformers/model_doc" ,) index += 1 with open(__snake_case ,"w" ,encoding="utf-8" ,newline="\n" ) as f: f.writelines(__snake_case ) def _lowercase ( ) -> List[Any]: with open(REPLACE_FILES["init"] ,"r" ) as f: __lowerCAmelCase : int = f.read() __lowerCAmelCase : Union[str, Any] = REPLACE_PATTERNS["init"][0].search(__snake_case ).groups()[0] return packaging.version.parse(__snake_case ) def _lowercase ( __snake_case=False ) -> int: __lowerCAmelCase : Union[str, Any] = get_version() if patch and default_version.is_devrelease: raise ValueError("Can't create a patch version from the dev branch, checkout a released version!" ) if default_version.is_devrelease: __lowerCAmelCase : Optional[int] = default_version.base_version elif patch: __lowerCAmelCase : Union[str, Any] = F"""{default_version.major}.{default_version.minor}.{default_version.micro + 1}""" else: __lowerCAmelCase : str = F"""{default_version.major}.{default_version.minor + 1}.0""" # Now let's ask nicely if that's the right one. __lowerCAmelCase : Optional[int] = input(F"""Which version are you releasing? [{default_version}]""" ) if len(__snake_case ) == 0: __lowerCAmelCase : int = default_version print(F"""Updating version to {version}.""" ) global_version_update(__snake_case ,patch=__snake_case ) if not patch: print("Cleaning main README, don't forget to run `make fix-copies`." ) clean_main_ref_in_model_list() def _lowercase ( ) -> List[str]: __lowerCAmelCase : Optional[Any] = get_version() __lowerCAmelCase : Tuple = F"""{current_version.major}.{current_version.minor + 1}.0.dev0""" __lowerCAmelCase : Any = current_version.base_version # Check with the user we got that right. __lowerCAmelCase : Optional[Any] = input(F"""Which version are we developing now? [{dev_version}]""" ) if len(__snake_case ) == 0: __lowerCAmelCase : str = dev_version print(F"""Updating version to {version}.""" ) global_version_update(__snake_case ) print("Cleaning main README, don't forget to run `make fix-copies`." ) clean_main_ref_in_model_list() if __name__ == "__main__": __snake_case : Tuple = argparse.ArgumentParser() parser.add_argument('--post_release', action='store_true', help='Whether this is pre or post release.') parser.add_argument('--patch', action='store_true', help='Whether or not this is a patch release.') __snake_case : List[str] = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print('Nothing to do after a patch :-)') else: post_release_work()
269
"""simple docstring""" # # This a `torch.distributed` diagnostics script that checks that all GPUs in the cluster (one or # many nodes) can talk to each other via nccl and allocate gpu memory. # # To run first adjust the number of processes and nodes: # # python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # You may need to add --master_addr $MASTER_ADDR --master_port $MASTER_PORT if using a custom addr:port # # You can also use the rdzv API: --rdzv_endpoint $MASTER_ADDR:$MASTER_PORT --rdzv_backend c10d # # use torch.distributed.launch instead of torch.distributed.run for torch < 1.9 # # If you get a hanging in `barrier` calls you have some network issues, you may try to debug this with: # # NCCL_DEBUG=INFO python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # which should tell you what's going on behind the scenes. # # # This script can be run via `srun` in the SLURM environment as well. Here is a SLURM script that # runs on 2 nodes of 4 gpus per node: # # #SBATCH --job-name=test-nodes # name # #SBATCH --nodes=2 # nodes # #SBATCH --ntasks-per-node=1 # crucial - only 1 task per dist per node! # #SBATCH --cpus-per-task=10 # number of cores per tasks # #SBATCH --gres=gpu:4 # number of gpus # #SBATCH --time 0:05:00 # maximum execution time (HH:MM:SS) # #SBATCH --output=%x-%j.out # output file name # # GPUS_PER_NODE=4 # MASTER_ADDR=$(scontrol show hostnames $SLURM_JOB_NODELIST | head -n 1) # MASTER_PORT=6000 # # srun --jobid $SLURM_JOBID bash -c 'python -m torch.distributed.run \ # --nproc_per_node $GPUS_PER_NODE --nnodes $SLURM_NNODES --node_rank $SLURM_PROCID \ # --master_addr $MASTER_ADDR --master_port $MASTER_PORT \ # torch-distributed-gpu-test.py' # import fcntl import os import socket import torch import torch.distributed as dist def _lowercase ( *__snake_case ) -> Optional[Any]: with open(__snake_case ,"r" ) as fh: fcntl.flock(__snake_case ,fcntl.LOCK_EX ) try: print(*__snake_case ) finally: fcntl.flock(__snake_case ,fcntl.LOCK_UN ) __snake_case : List[Any] = int(os.environ['LOCAL_RANK']) torch.cuda.set_device(local_rank) __snake_case : List[str] = torch.device('cuda', local_rank) __snake_case : Optional[Any] = socket.gethostname() __snake_case : str = F"""[{hostname}-{local_rank}]""" try: # test distributed dist.init_process_group('nccl') dist.all_reduce(torch.ones(1).to(device), op=dist.ReduceOp.SUM) dist.barrier() # test cuda is available and can allocate memory torch.cuda.is_available() torch.ones(1).cuda(local_rank) # global rank __snake_case : Tuple = dist.get_rank() __snake_case : Any = dist.get_world_size() printflock(F"""{gpu} is OK (global rank: {rank}/{world_size})""") dist.barrier() if rank == 0: printflock(F"""pt={torch.__version__}, cuda={torch.version.cuda}, nccl={torch.cuda.nccl.version()}""") except Exception: printflock(F"""{gpu} is broken""") raise
269
1
"""simple docstring""" from ....configuration_utils import PretrainedConfig from ....utils import logging __snake_case : List[Any] = logging.get_logger(__name__) __snake_case : Any = { 'speechbrain/m-ctc-t-large': 'https://huggingface.co/speechbrain/m-ctc-t-large/resolve/main/config.json', # See all M-CTC-T models at https://huggingface.co/models?filter=mctct } class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'mctct' def __init__( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: str=8065 , _SCREAMING_SNAKE_CASE: str=1536 , _SCREAMING_SNAKE_CASE: str=36 , _SCREAMING_SNAKE_CASE: Optional[Any]=6144 , _SCREAMING_SNAKE_CASE: Optional[Any]=4 , _SCREAMING_SNAKE_CASE: Union[str, Any]=384 , _SCREAMING_SNAKE_CASE: Optional[Any]=920 , _SCREAMING_SNAKE_CASE: Union[str, Any]=1e-5 , _SCREAMING_SNAKE_CASE: List[Any]=0.3 , _SCREAMING_SNAKE_CASE: Optional[Any]="relu" , _SCREAMING_SNAKE_CASE: Optional[int]=0.02 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.3 , _SCREAMING_SNAKE_CASE: Dict=0.3 , _SCREAMING_SNAKE_CASE: List[Any]=1 , _SCREAMING_SNAKE_CASE: Optional[Any]=0 , _SCREAMING_SNAKE_CASE: List[str]=2 , _SCREAMING_SNAKE_CASE: Union[str, Any]=1 , _SCREAMING_SNAKE_CASE: Tuple=0.3 , _SCREAMING_SNAKE_CASE: Dict=1 , _SCREAMING_SNAKE_CASE: int=(7,) , _SCREAMING_SNAKE_CASE: str=(3,) , _SCREAMING_SNAKE_CASE: Union[str, Any]=80 , _SCREAMING_SNAKE_CASE: Tuple=1 , _SCREAMING_SNAKE_CASE: Dict=None , _SCREAMING_SNAKE_CASE: Tuple="sum" , _SCREAMING_SNAKE_CASE: List[str]=False , **_SCREAMING_SNAKE_CASE: Tuple , ) -> Tuple: """simple docstring""" super().__init__(**_SCREAMING_SNAKE_CASE , pad_token_id=_SCREAMING_SNAKE_CASE , bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = vocab_size __lowerCAmelCase : str = hidden_size __lowerCAmelCase : str = num_hidden_layers __lowerCAmelCase : str = intermediate_size __lowerCAmelCase : List[Any] = num_attention_heads __lowerCAmelCase : Dict = attention_head_dim __lowerCAmelCase : Optional[int] = max_position_embeddings __lowerCAmelCase : str = layer_norm_eps __lowerCAmelCase : Tuple = layerdrop __lowerCAmelCase : str = hidden_act __lowerCAmelCase : List[Any] = initializer_range __lowerCAmelCase : int = hidden_dropout_prob __lowerCAmelCase : Optional[Any] = attention_probs_dropout_prob __lowerCAmelCase : str = pad_token_id __lowerCAmelCase : Optional[int] = bos_token_id __lowerCAmelCase : Union[str, Any] = eos_token_id __lowerCAmelCase : Any = conv_glu_dim __lowerCAmelCase : Optional[int] = conv_dropout __lowerCAmelCase : Union[str, Any] = num_conv_layers __lowerCAmelCase : Optional[int] = input_feat_per_channel __lowerCAmelCase : Union[str, Any] = input_channels __lowerCAmelCase : Optional[Any] = conv_channels __lowerCAmelCase : Dict = ctc_loss_reduction __lowerCAmelCase : int = ctc_zero_infinity # prevents config testing fail with exporting to json __lowerCAmelCase : List[str] = list(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = list(_SCREAMING_SNAKE_CASE) if len(self.conv_kernel) != self.num_conv_layers: raise ValueError( "Configuration for convolutional module is incorrect. " "It is required that `len(config.conv_kernel)` == `config.num_conv_layers` " F"""but is `len(config.conv_kernel) = {len(self.conv_kernel)}`, """ F"""`config.num_conv_layers = {self.num_conv_layers}`.""")
269
"""simple docstring""" import numpy as np from scipy.spatial.distance import cdist from sklearn.metrics import fa_score import datasets __snake_case : Optional[int] = '\\n @inproceedings{kakwani2020indicnlpsuite,\n title={{IndicNLPSuite: Monolingual Corpora, Evaluation Benchmarks and Pre-trained Multilingual Language Models for Indian Languages}},\n author={Divyanshu Kakwani and Anoop Kunchukuttan and Satish Golla and Gokul N.C. and Avik Bhattacharyya and Mitesh M. Khapra and Pratyush Kumar},\n year={2020},\n booktitle={Findings of EMNLP},\n}\n' __snake_case : str = '\\n IndicGLUE is a natural language understanding benchmark for Indian languages. It contains a wide\n variety of tasks and covers 11 major Indian languages - as, bn, gu, hi, kn, ml, mr, or, pa, ta, te.\n' __snake_case : str = '\nCompute IndicGLUE evaluation metric associated to each IndicGLUE dataset.\nArgs:\n predictions: list of predictions to score (as int64),\n except for \'cvit-mkb-clsr\' where each prediction is a vector (of float32).\n references: list of ground truth labels corresponding to the predictions (as int64),\n except for \'cvit-mkb-clsr\' where each reference is a vector (of float32).\nReturns: depending on the IndicGLUE subset, one or several of:\n "accuracy": Accuracy\n "f1": F1 score\n "precision": Precision@10\nExamples:\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wnli\') # \'wnli\' or any of ["copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md"]\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0}\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wiki-ner\')\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0, \'f1\': 1.0}\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'cvit-mkb-clsr\')\n >>> references = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]]\n >>> predictions = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'precision@10\': 1.0}\n\n' def _lowercase ( __snake_case ,__snake_case ) -> Union[str, Any]: return float((preds == labels).mean() ) def _lowercase ( __snake_case ,__snake_case ) -> str: __lowerCAmelCase : str = simple_accuracy(__snake_case ,__snake_case ) __lowerCAmelCase : Any = float(fa_score(y_true=__snake_case ,y_pred=__snake_case ) ) return { "accuracy": acc, "f1": fa, } def _lowercase ( __snake_case ,__snake_case ) -> int: __lowerCAmelCase : Union[str, Any] = np.array(__snake_case ) __lowerCAmelCase : Tuple = np.array(__snake_case ) __lowerCAmelCase : List[Any] = en_sentvecs.shape[0] # mean centering __lowerCAmelCase : Union[str, Any] = en_sentvecs - np.mean(__snake_case ,axis=0 ) __lowerCAmelCase : int = in_sentvecs - np.mean(__snake_case ,axis=0 ) __lowerCAmelCase : Optional[Any] = cdist(__snake_case ,__snake_case ,"cosine" ) __lowerCAmelCase : int = np.array(range(__snake_case ) ) __lowerCAmelCase : int = sim.argsort(axis=1 )[:, :10] __lowerCAmelCase : Optional[Any] = np.any(preds == actual[:, None] ,axis=1 ) return float(matches.mean() ) @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A__ ( datasets.Metric ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: int) -> str: """simple docstring""" if self.config_name not in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", "wiki-ner", ]: raise KeyError( "You should supply a configuration name selected in " "[\"wnli\", \"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", " "\"cvit-mkb-clsr\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\", " "\"wiki-ner\"]") return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("int64") if self.config_name != "cvit-mkb-clsr" else datasets.Sequence(datasets.Value("float32")), "references": datasets.Value("int64") if self.config_name != "cvit-mkb-clsr" else datasets.Sequence(datasets.Value("float32")), }) , codebase_urls=[] , reference_urls=[] , format="numpy" if self.config_name != "cvit-mkb-clsr" else None , ) def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Optional[Any]) -> int: """simple docstring""" if self.config_name == "cvit-mkb-clsr": return {"precision@10": precision_at_aa(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE)} elif self.config_name in ["wiki-ner"]: return acc_and_fa(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) elif self.config_name in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md", ]: return {"accuracy": simple_accuracy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE)} else: raise KeyError( "You should supply a configuration name selected in " "[\"wnli\", \"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", " "\"cvit-mkb-clsr\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\", " "\"wiki-ner\"]")
269
1
"""simple docstring""" import argparse import re import requests import torch # git clone https://github.com/salesforce/BLIP.git from models.blip import blip_decoder from models.blip_itm import blip_itm from models.blip_vqa import blip_vqa from PIL import Image from torchvision import transforms from torchvision.transforms.functional import InterpolationMode from transformers import ( BertTokenizer, BlipConfig, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, ) def _lowercase ( __snake_case ,__snake_case ) -> List[str]: __lowerCAmelCase : int = "https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg" __lowerCAmelCase : List[str] = Image.open(requests.get(__snake_case ,stream=__snake_case ).raw ).convert("RGB" ) __lowerCAmelCase : Optional[int] = transforms.Compose( [ transforms.Resize((image_size, image_size) ,interpolation=InterpolationMode.BICUBIC ), transforms.ToTensor(), transforms.Normalize((0.48145466, 0.4578275, 0.40821073) ,(0.26862954, 0.26130258, 0.27577711) ), ] ) __lowerCAmelCase : Union[str, Any] = transform(__snake_case ).unsqueeze(0 ).to(__snake_case ) return image def _lowercase ( __snake_case ) -> int: if "visual_encoder" in key: __lowerCAmelCase : List[str] = re.sub("visual_encoder*" ,"vision_model.encoder" ,__snake_case ) if "blocks" in key: __lowerCAmelCase : Union[str, Any] = re.sub(r"blocks" ,"layers" ,__snake_case ) if "attn" in key: __lowerCAmelCase : Tuple = re.sub(r"attn" ,"self_attn" ,__snake_case ) if "norm1" in key: __lowerCAmelCase : Optional[int] = re.sub(r"norm1" ,"layer_norm1" ,__snake_case ) if "norm2" in key: __lowerCAmelCase : Dict = re.sub(r"norm2" ,"layer_norm2" ,__snake_case ) if "encoder.norm" in key: __lowerCAmelCase : List[Any] = re.sub(r"encoder.norm" ,"post_layernorm" ,__snake_case ) if "encoder.patch_embed.proj" in key: __lowerCAmelCase : Any = re.sub(r"encoder.patch_embed.proj" ,"embeddings.patch_embedding" ,__snake_case ) if "encoder.pos_embed" in key: __lowerCAmelCase : Dict = re.sub(r"encoder.pos_embed" ,"embeddings.position_embedding" ,__snake_case ) if "encoder.cls_token" in key: __lowerCAmelCase : int = re.sub(r"encoder.cls_token" ,"embeddings.class_embedding" ,__snake_case ) if "self_attn" in key: __lowerCAmelCase : List[Any] = re.sub(r"self_attn.proj" ,"self_attn.projection" ,__snake_case ) return key @torch.no_grad() def _lowercase ( __snake_case ,__snake_case=None ) -> str: if config_path is not None: __lowerCAmelCase : Union[str, Any] = BlipConfig.from_pretrained(__snake_case ) else: __lowerCAmelCase : Optional[Any] = BlipConfig(projection_dim=512 ,text_config={} ,vision_config={} ) __lowerCAmelCase : Tuple = BlipForConditionalGeneration(__snake_case ).eval() __lowerCAmelCase : str = "https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_capfilt_large.pth" __lowerCAmelCase : Tuple = blip_decoder(pretrained=__snake_case ,image_size=384 ,vit="base" ) __lowerCAmelCase : str = pt_model.eval() __lowerCAmelCase : List[Any] = pt_model.state_dict() for key in modified_state_dict.copy(): __lowerCAmelCase : List[Any] = modified_state_dict.pop(__snake_case ) __lowerCAmelCase : int = rename_key(__snake_case ) __lowerCAmelCase : int = value hf_model.load_state_dict(__snake_case ) __lowerCAmelCase : List[Any] = 384 __lowerCAmelCase : Union[str, Any] = load_demo_image(image_size=__snake_case ,device="cpu" ) __lowerCAmelCase : Union[str, Any] = BertTokenizer.from_pretrained("bert-base-uncased" ) __lowerCAmelCase : Tuple = tokenizer(["a picture of"] ).input_ids __lowerCAmelCase : List[str] = hf_model.generate(__snake_case ,__snake_case ) assert out[0].tolist() == [30_522, 1_037, 3_861, 1_997, 1_037, 2_450, 3_564, 2_006, 1_996, 3_509, 2_007, 2_014, 3_899, 102] __lowerCAmelCase : str = hf_model.generate(__snake_case ) assert out[0].tolist() == [30_522, 1_037, 2_450, 3_564, 2_006, 1_996, 3_509, 2_007, 2_014, 3_899, 102] if pytorch_dump_folder_path is not None: hf_model.save_pretrained(__snake_case ) # model_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_vqa.pth' __lowerCAmelCase : Optional[int] = ( "https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_vqa_capfilt_large.pth" ) __lowerCAmelCase : List[Any] = blip_vqa(pretrained=__snake_case ,image_size=__snake_case ,vit="base" ) vqa_model.eval() __lowerCAmelCase : Union[str, Any] = vqa_model.state_dict() for key in modified_state_dict.copy(): __lowerCAmelCase : Union[str, Any] = modified_state_dict.pop(__snake_case ) __lowerCAmelCase : Optional[Any] = rename_key(__snake_case ) __lowerCAmelCase : str = value __lowerCAmelCase : List[str] = BlipForQuestionAnswering(__snake_case ) hf_vqa_model.load_state_dict(__snake_case ) __lowerCAmelCase : Optional[int] = ["How many dogs are in this image?"] __lowerCAmelCase : Any = tokenizer(__snake_case ,return_tensors="pt" ).input_ids __lowerCAmelCase : int = hf_vqa_model.generate(__snake_case ,__snake_case ) print(tokenizer.decode(answer[0] ) ) assert tokenizer.decode(answer[0] ) == "[UNK] 1 [SEP]" if pytorch_dump_folder_path is not None: hf_vqa_model.save_pretrained(pytorch_dump_folder_path + "_vqa" ) __lowerCAmelCase : str = "https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_retrieval_coco.pth" __lowerCAmelCase : Any = blip_itm(pretrained=__snake_case ,image_size=__snake_case ,vit="base" ) itm_model.eval() __lowerCAmelCase : Dict = itm_model.state_dict() for key in modified_state_dict.copy(): __lowerCAmelCase : Optional[Any] = modified_state_dict.pop(__snake_case ) __lowerCAmelCase : List[str] = rename_key(__snake_case ) __lowerCAmelCase : List[Any] = value __lowerCAmelCase : Dict = BlipForImageTextRetrieval(__snake_case ) __lowerCAmelCase : List[str] = ["A picture of a woman with a dog sitting in a beach"] __lowerCAmelCase : List[str] = tokenizer( __snake_case ,return_tensors="pt" ,padding="max_length" ,truncation=__snake_case ,max_length=35 ,).input_ids hf_itm_model.load_state_dict(__snake_case ) hf_itm_model.eval() __lowerCAmelCase : Tuple = hf_itm_model(__snake_case ,__snake_case ,use_itm_head=__snake_case ) __lowerCAmelCase : int = hf_itm_model(__snake_case ,__snake_case ,use_itm_head=__snake_case ) assert out[0].item() == 0.2110687494277954 assert torch.nn.functional.softmax(out_itm[0] ,dim=1 )[:, 1].item() == 0.45698845386505127 if pytorch_dump_folder_path is not None: hf_itm_model.save_pretrained(pytorch_dump_folder_path + "_itm" ) if __name__ == "__main__": __snake_case : Optional[Any] = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') __snake_case : List[str] = parser.parse_args() convert_blip_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
269
"""simple docstring""" def _lowercase ( __snake_case ,__snake_case ) -> float: if digit_amount > 0: return round(number - int(__snake_case ) ,__snake_case ) return number - int(__snake_case ) if __name__ == "__main__": print(decimal_isolate(1.53, 0)) print(decimal_isolate(35.345, 1)) print(decimal_isolate(35.345, 2)) print(decimal_isolate(35.345, 3)) print(decimal_isolate(-14.789, 3)) print(decimal_isolate(0, 2)) print(decimal_isolate(-14.123, 1)) print(decimal_isolate(-14.123, 2)) print(decimal_isolate(-14.123, 3))
269
1
"""simple docstring""" import numpy as np import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModelWithProjection, PreTrainedModel from ...utils import logging __snake_case : str = logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = CLIPConfig SCREAMING_SNAKE_CASE = ['CLIPEncoderLayer'] def __init__( self: Optional[Any] , _SCREAMING_SNAKE_CASE: CLIPConfig) -> Any: """simple docstring""" super().__init__(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = CLIPVisionModelWithProjection(config.vision_config) __lowerCAmelCase : Union[str, Any] = nn.Linear(config.vision_config.projection_dim , 1) __lowerCAmelCase : str = nn.Linear(config.vision_config.projection_dim , 1) @torch.no_grad() def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Dict=0.5 , _SCREAMING_SNAKE_CASE: str=0.5) -> Dict: """simple docstring""" __lowerCAmelCase : Optional[int] = self.vision_model(_SCREAMING_SNAKE_CASE)[0] __lowerCAmelCase : Dict = self.p_head(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = nsfw_detected.flatten() __lowerCAmelCase : int = nsfw_detected > p_threshold __lowerCAmelCase : Optional[Any] = nsfw_detected.tolist() if any(_SCREAMING_SNAKE_CASE): logger.warning( "Potential NSFW content was detected in one or more images. A black image will be returned instead." " Try again with a different prompt and/or seed.") for idx, nsfw_detected_ in enumerate(_SCREAMING_SNAKE_CASE): if nsfw_detected_: __lowerCAmelCase : int = np.zeros(images[idx].shape) __lowerCAmelCase : Tuple = self.w_head(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = watermark_detected.flatten() __lowerCAmelCase : Any = watermark_detected > w_threshold __lowerCAmelCase : Tuple = watermark_detected.tolist() if any(_SCREAMING_SNAKE_CASE): logger.warning( "Potential watermarked content was detected in one or more images. A black image will be returned instead." " Try again with a different prompt and/or seed.") for idx, watermark_detected_ in enumerate(_SCREAMING_SNAKE_CASE): if watermark_detected_: __lowerCAmelCase : List[str] = np.zeros(images[idx].shape) return images, nsfw_detected, watermark_detected
269
"""simple docstring""" from ..utils import DummyObject, requires_backends class A__ ( metaclass=__SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = ['torch', 'torchsde'] def __init__( self: int , *_SCREAMING_SNAKE_CASE: Optional[Any] , **_SCREAMING_SNAKE_CASE: str) -> Dict: """simple docstring""" requires_backends(self , ["torch", "torchsde"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls: Optional[Any] , *_SCREAMING_SNAKE_CASE: Optional[int] , **_SCREAMING_SNAKE_CASE: Optional[int]) -> Optional[Any]: """simple docstring""" requires_backends(cls , ["torch", "torchsde"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls: Dict , *_SCREAMING_SNAKE_CASE: List[Any] , **_SCREAMING_SNAKE_CASE: Any) -> Optional[int]: """simple docstring""" requires_backends(cls , ["torch", "torchsde"])
269
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available __snake_case : Any = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : List[str] = ['MLukeTokenizer'] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mluke import MLukeTokenizer else: import sys __snake_case : Any = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
269
"""simple docstring""" def _lowercase ( ) -> int: return 1 def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else two_pence(x - 2 ) + one_pence() def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else five_pence(x - 5 ) + two_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else ten_pence(x - 10 ) + five_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else twenty_pence(x - 20 ) + ten_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else fifty_pence(x - 50 ) + twenty_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else one_pound(x - 100 ) + fifty_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else two_pound(x - 200 ) + one_pound(__snake_case ) def _lowercase ( __snake_case = 200 ) -> int: return two_pound(__snake_case ) if __name__ == "__main__": print(solution(int(input().strip())))
269
1
"""simple docstring""" from __future__ import annotations import os from typing import Any import requests __snake_case : Dict = 'https://api.github.com' # https://docs.github.com/en/free-pro-team@latest/rest/reference/users#get-the-authenticated-user __snake_case : Optional[int] = BASE_URL + '/user' # https://github.com/settings/tokens __snake_case : Optional[Any] = os.environ.get('USER_TOKEN', '') def _lowercase ( __snake_case ) -> dict[Any, Any]: __lowerCAmelCase : Dict = { "Authorization": F"""token {auth_token}""", "Accept": "application/vnd.github.v3+json", } return requests.get(__snake_case ,headers=__snake_case ).json() if __name__ == "__main__": # pragma: no cover if USER_TOKEN: for key, value in fetch_github_info(USER_TOKEN).items(): print(F"""{key}: {value}""") else: raise ValueError('\'USER_TOKEN\' field cannot be empty.')
269
"""simple docstring""" import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Optional[int] , _SCREAMING_SNAKE_CASE: str) -> Dict: """simple docstring""" with open(_SCREAMING_SNAKE_CASE , encoding="utf-8") as input_file: __lowerCAmelCase : List[str] = re.compile(r"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)") __lowerCAmelCase : List[Any] = input_file.read() __lowerCAmelCase : Any = regexp.search(_SCREAMING_SNAKE_CASE) return match def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: str) -> Optional[Any]: """simple docstring""" with open(_SCREAMING_SNAKE_CASE , encoding="utf-8") as input_file: __lowerCAmelCase : Any = re.compile(r"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL) __lowerCAmelCase : Optional[int] = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` __lowerCAmelCase : int = regexp.finditer(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = [match for match in matches if match is not None and match.group(1) is not None] return matches[0] if matches else None def _SCREAMING_SNAKE_CASE ( self: str) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[Any] = Path("./datasets") __lowerCAmelCase : Optional[int] = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_encoding_on_file_open(str(_SCREAMING_SNAKE_CASE)): raise AssertionError(F"""open(...) must use utf-8 encoding in {dataset}""") def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> Optional[int]: """simple docstring""" __lowerCAmelCase : Dict = Path("./datasets") __lowerCAmelCase : Union[str, Any] = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_print_statements(str(_SCREAMING_SNAKE_CASE)): raise AssertionError(F"""print statement found in {dataset}. Use datasets.logger/logging instead.""")
269
1
"""simple docstring""" import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, LDMTextToImagePipeline, UNetaDConditionModel from diffusers.utils.testing_utils import ( enable_full_determinism, load_numpy, nightly, require_torch_gpu, slow, torch_device, ) from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = LDMTextToImagePipeline SCREAMING_SNAKE_CASE = TEXT_TO_IMAGE_PARAMS - { 'negative_prompt', 'negative_prompt_embeds', 'cross_attention_kwargs', 'prompt_embeds', } SCREAMING_SNAKE_CASE = PipelineTesterMixin.required_optional_params - { 'num_images_per_prompt', 'callback', 'callback_steps', } SCREAMING_SNAKE_CASE = TEXT_TO_IMAGE_BATCH_PARAMS SCREAMING_SNAKE_CASE = False def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> str: """simple docstring""" torch.manual_seed(0) __lowerCAmelCase : Tuple = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) __lowerCAmelCase : Optional[int] = DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=_SCREAMING_SNAKE_CASE , set_alpha_to_one=_SCREAMING_SNAKE_CASE , ) torch.manual_seed(0) __lowerCAmelCase : Optional[int] = AutoencoderKL( block_out_channels=(32, 64) , in_channels=3 , out_channels=3 , down_block_types=("DownEncoderBlock2D", "DownEncoderBlock2D") , up_block_types=("UpDecoderBlock2D", "UpDecoderBlock2D") , latent_channels=4 , ) torch.manual_seed(0) __lowerCAmelCase : List[str] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) __lowerCAmelCase : Optional[int] = CLIPTextModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip") __lowerCAmelCase : Optional[Any] = { "unet": unet, "scheduler": scheduler, "vqvae": vae, "bert": text_encoder, "tokenizer": tokenizer, } return components def _SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: List[Any]=0) -> Any: """simple docstring""" if str(_SCREAMING_SNAKE_CASE).startswith("mps"): __lowerCAmelCase : Union[str, Any] = torch.manual_seed(_SCREAMING_SNAKE_CASE) else: __lowerCAmelCase : Optional[Any] = torch.Generator(device=_SCREAMING_SNAKE_CASE).manual_seed(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def _SCREAMING_SNAKE_CASE ( self: Tuple) -> List[str]: """simple docstring""" __lowerCAmelCase : Any = "cpu" # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : Optional[int] = self.get_dummy_components() __lowerCAmelCase : List[Any] = LDMTextToImagePipeline(**_SCREAMING_SNAKE_CASE) pipe.to(_SCREAMING_SNAKE_CASE) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = self.get_dummy_inputs(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = pipe(**_SCREAMING_SNAKE_CASE).images __lowerCAmelCase : List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 16, 16, 3) __lowerCAmelCase : List[str] = np.array([0.6101, 0.6156, 0.5622, 0.4895, 0.6661, 0.3804, 0.5748, 0.6136, 0.5014]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-3 @slow @require_torch_gpu class A__ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Dict) -> List[Any]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _SCREAMING_SNAKE_CASE ( self: str , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: Tuple=torch.floataa , _SCREAMING_SNAKE_CASE: Optional[int]=0) -> str: """simple docstring""" __lowerCAmelCase : List[Any] = torch.manual_seed(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = np.random.RandomState(_SCREAMING_SNAKE_CASE).standard_normal((1, 4, 32, 32)) __lowerCAmelCase : Union[str, Any] = torch.from_numpy(_SCREAMING_SNAKE_CASE).to(device=_SCREAMING_SNAKE_CASE , dtype=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = { "prompt": "A painting of a squirrel eating a burger", "latents": latents, "generator": generator, "num_inference_steps": 3, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def _SCREAMING_SNAKE_CASE ( self: int) -> Optional[int]: """simple docstring""" __lowerCAmelCase : Optional[int] = LDMTextToImagePipeline.from_pretrained("CompVis/ldm-text2im-large-256").to(_SCREAMING_SNAKE_CASE) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = self.get_inputs(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = pipe(**_SCREAMING_SNAKE_CASE).images __lowerCAmelCase : Tuple = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 256, 256, 3) __lowerCAmelCase : Any = np.array([0.5_1825, 0.5_2850, 0.5_2543, 0.5_4258, 0.5_2304, 0.5_2569, 0.5_4363, 0.5_5276, 0.5_6878]) __lowerCAmelCase : Union[str, Any] = np.abs(expected_slice - image_slice).max() assert max_diff < 1e-3 @nightly @require_torch_gpu class A__ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> Optional[int]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: List[Any]=torch.floataa , _SCREAMING_SNAKE_CASE: Optional[Any]=0) -> Optional[int]: """simple docstring""" __lowerCAmelCase : int = torch.manual_seed(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = np.random.RandomState(_SCREAMING_SNAKE_CASE).standard_normal((1, 4, 32, 32)) __lowerCAmelCase : List[Any] = torch.from_numpy(_SCREAMING_SNAKE_CASE).to(device=_SCREAMING_SNAKE_CASE , dtype=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = { "prompt": "A painting of a squirrel eating a burger", "latents": latents, "generator": generator, "num_inference_steps": 50, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def _SCREAMING_SNAKE_CASE ( self: List[str]) -> str: """simple docstring""" __lowerCAmelCase : Optional[int] = LDMTextToImagePipeline.from_pretrained("CompVis/ldm-text2im-large-256").to(_SCREAMING_SNAKE_CASE) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = self.get_inputs(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = pipe(**_SCREAMING_SNAKE_CASE).images[0] __lowerCAmelCase : int = load_numpy( "https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/ldm_text2img/ldm_large_256_ddim.npy") __lowerCAmelCase : Optional[int] = np.abs(expected_image - image).max() assert max_diff < 1e-3
269
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxSeqaSeqConfigWithPast from ...utils import logging if TYPE_CHECKING: from ...feature_extraction_utils import FeatureExtractionMixin from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType __snake_case : Optional[Any] = logging.get_logger(__name__) __snake_case : Optional[Any] = { 'openai/whisper-base': 'https://huggingface.co/openai/whisper-base/resolve/main/config.json', } # fmt: off __snake_case : Any = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 357, 366, 438, 532, 685, 705, 796, 930, 1_058, 1_220, 1_267, 1_279, 1_303, 1_343, 1_377, 1_391, 1_635, 1_782, 1_875, 2_162, 2_361, 2_488, 3_467, 4_008, 4_211, 4_600, 4_808, 5_299, 5_855, 6_329, 7_203, 9_609, 9_959, 10_563, 10_786, 11_420, 11_709, 11_907, 13_163, 13_697, 13_700, 14_808, 15_306, 16_410, 16_791, 17_992, 19_203, 19_510, 20_724, 22_305, 22_935, 27_007, 30_109, 30_420, 33_409, 34_949, 40_283, 40_493, 40_549, 47_282, 49_146, 50_257, 50_359, 50_360, 50_361 ] __snake_case : str = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 359, 503, 522, 542, 873, 893, 902, 918, 922, 931, 1_350, 1_853, 1_982, 2_460, 2_627, 3_246, 3_253, 3_268, 3_536, 3_846, 3_961, 4_183, 4_667, 6_585, 6_647, 7_273, 9_061, 9_383, 10_428, 10_929, 11_938, 12_033, 12_331, 12_562, 13_793, 14_157, 14_635, 15_265, 15_618, 16_553, 16_604, 18_362, 18_956, 20_075, 21_675, 22_520, 26_130, 26_161, 26_435, 28_279, 29_464, 31_650, 32_302, 32_470, 36_865, 42_863, 47_425, 49_870, 50_254, 50_258, 50_360, 50_361, 50_362 ] class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'whisper' SCREAMING_SNAKE_CASE = ['past_key_values'] SCREAMING_SNAKE_CASE = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self: Optional[int] , _SCREAMING_SNAKE_CASE: Any=5_1865 , _SCREAMING_SNAKE_CASE: Optional[Any]=80 , _SCREAMING_SNAKE_CASE: Optional[int]=6 , _SCREAMING_SNAKE_CASE: Any=4 , _SCREAMING_SNAKE_CASE: Dict=6 , _SCREAMING_SNAKE_CASE: Dict=4 , _SCREAMING_SNAKE_CASE: Optional[Any]=1536 , _SCREAMING_SNAKE_CASE: List[str]=1536 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.0 , _SCREAMING_SNAKE_CASE: str=0.0 , _SCREAMING_SNAKE_CASE: List[str]=5_0257 , _SCREAMING_SNAKE_CASE: Optional[Any]=True , _SCREAMING_SNAKE_CASE: List[str]=True , _SCREAMING_SNAKE_CASE: Optional[int]="gelu" , _SCREAMING_SNAKE_CASE: Tuple=256 , _SCREAMING_SNAKE_CASE: str=0.0 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.0 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.0 , _SCREAMING_SNAKE_CASE: List[Any]=0.02 , _SCREAMING_SNAKE_CASE: Tuple=False , _SCREAMING_SNAKE_CASE: Union[str, Any]=1500 , _SCREAMING_SNAKE_CASE: str=448 , _SCREAMING_SNAKE_CASE: Any=5_0256 , _SCREAMING_SNAKE_CASE: Any=5_0256 , _SCREAMING_SNAKE_CASE: List[str]=5_0256 , _SCREAMING_SNAKE_CASE: Dict=None , _SCREAMING_SNAKE_CASE: List[Any]=[220, 5_0256] , _SCREAMING_SNAKE_CASE: Dict=False , _SCREAMING_SNAKE_CASE: str=256 , _SCREAMING_SNAKE_CASE: List[str]=False , _SCREAMING_SNAKE_CASE: Tuple=0.05 , _SCREAMING_SNAKE_CASE: List[str]=10 , _SCREAMING_SNAKE_CASE: str=2 , _SCREAMING_SNAKE_CASE: Any=0.0 , _SCREAMING_SNAKE_CASE: Optional[int]=10 , _SCREAMING_SNAKE_CASE: int=0 , _SCREAMING_SNAKE_CASE: Any=7 , **_SCREAMING_SNAKE_CASE: List[str] , ) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[int] = vocab_size __lowerCAmelCase : Optional[Any] = num_mel_bins __lowerCAmelCase : int = d_model __lowerCAmelCase : List[Any] = encoder_layers __lowerCAmelCase : List[Any] = encoder_attention_heads __lowerCAmelCase : List[str] = decoder_layers __lowerCAmelCase : Tuple = decoder_attention_heads __lowerCAmelCase : Any = decoder_ffn_dim __lowerCAmelCase : Tuple = encoder_ffn_dim __lowerCAmelCase : List[str] = dropout __lowerCAmelCase : Union[str, Any] = attention_dropout __lowerCAmelCase : Union[str, Any] = activation_dropout __lowerCAmelCase : Dict = activation_function __lowerCAmelCase : Tuple = init_std __lowerCAmelCase : str = encoder_layerdrop __lowerCAmelCase : int = decoder_layerdrop __lowerCAmelCase : Optional[int] = use_cache __lowerCAmelCase : Union[str, Any] = encoder_layers __lowerCAmelCase : List[str] = scale_embedding # scale factor will be sqrt(d_model) if True __lowerCAmelCase : int = max_source_positions __lowerCAmelCase : Any = max_target_positions # Audio Classification-specific parameters. Feel free to ignore for other classes. __lowerCAmelCase : Dict = classifier_proj_size __lowerCAmelCase : Dict = use_weighted_layer_sum # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __lowerCAmelCase : int = apply_spec_augment __lowerCAmelCase : Union[str, Any] = mask_time_prob __lowerCAmelCase : str = mask_time_length __lowerCAmelCase : int = mask_time_min_masks __lowerCAmelCase : List[Any] = mask_feature_prob __lowerCAmelCase : Tuple = mask_feature_length __lowerCAmelCase : Any = mask_feature_min_masks __lowerCAmelCase : Union[str, Any] = median_filter_width super().__init__( pad_token_id=_SCREAMING_SNAKE_CASE , bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE , is_encoder_decoder=_SCREAMING_SNAKE_CASE , decoder_start_token_id=_SCREAMING_SNAKE_CASE , suppress_tokens=_SCREAMING_SNAKE_CASE , begin_suppress_tokens=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' @property def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> Mapping[str, Mapping[int, str]]: """simple docstring""" __lowerCAmelCase : List[str] = OrderedDict( [ ("input_features", {0: "batch", 1: "feature_size", 2: "encoder_sequence"}), ]) if self.use_past: __lowerCAmelCase : Tuple = {0: "batch"} else: __lowerCAmelCase : Union[str, Any] = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(_SCREAMING_SNAKE_CASE , direction="inputs") return common_inputs def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: Union["PreTrainedTokenizerBase", "FeatureExtractionMixin"] , _SCREAMING_SNAKE_CASE: int = -1 , _SCREAMING_SNAKE_CASE: int = -1 , _SCREAMING_SNAKE_CASE: bool = False , _SCREAMING_SNAKE_CASE: Optional["TensorType"] = None , _SCREAMING_SNAKE_CASE: int = 2_2050 , _SCREAMING_SNAKE_CASE: float = 5.0 , _SCREAMING_SNAKE_CASE: int = 220 , ) -> Mapping[str, Any]: """simple docstring""" __lowerCAmelCase : int = OrderedDict() __lowerCAmelCase : Optional[Any] = OnnxConfig.generate_dummy_inputs( self , preprocessor=preprocessor.feature_extractor , batch_size=_SCREAMING_SNAKE_CASE , framework=_SCREAMING_SNAKE_CASE , sampling_rate=_SCREAMING_SNAKE_CASE , time_duration=_SCREAMING_SNAKE_CASE , frequency=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : List[str] = encoder_inputs["input_features"].shape[2] __lowerCAmelCase : List[str] = encoder_sequence_length // 2 if self.use_past else seq_length __lowerCAmelCase : List[Any] = super().generate_dummy_inputs( preprocessor.tokenizer , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = encoder_inputs.pop("input_features") __lowerCAmelCase : List[Any] = decoder_inputs.pop("decoder_input_ids") if "past_key_values" in decoder_inputs: __lowerCAmelCase : int = decoder_inputs.pop("past_key_values") return dummy_inputs @property def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> float: """simple docstring""" return 1e-3
269
1
"""simple docstring""" from binascii import hexlify from hashlib import shaaaa from os import urandom # RFC 3526 - More Modular Exponential (MODP) Diffie-Hellman groups for # Internet Key Exchange (IKE) https://tools.ietf.org/html/rfc3526 __snake_case : List[str] = { # 1536-bit 5: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA237327FFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 2048-bit 14: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AACAA68FFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 3072-bit 15: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64' + 'ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7' + 'ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B' + 'F12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31' + '43DB5BFCE0FD108E4B82D120A93AD2CAFFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 4096-bit 16: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64' + 'ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7' + 'ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B' + 'F12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31' + '43DB5BFCE0FD108E4B82D120A92108011A723C12A787E6D7' + '88719A10BDBA5B2699C327186AF4E23C1A946834B6150BDA' + '2583E9CA2AD44CE8DBBBC2DB04DE8EF92E8EFC141FBECAA6' + '287C59474E6BC05D99B2964FA090C3A2233BA186515BE7ED' + '1F612970CEE2D7AFB81BDD762170481CD0069127D5B05AA9' + '93B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934063199' + 'FFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 6144-bit 17: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD129024E08' + '8A67CC74020BBEA63B139B22514A08798E3404DDEF9519B3CD3A431B' + '302B0A6DF25F14374FE1356D6D51C245E485B576625E7EC6F44C42E9' + 'A637ED6B0BFF5CB6F406B7EDEE386BFB5A899FA5AE9F24117C4B1FE6' + '49286651ECE45B3DC2007CB8A163BF0598DA48361C55D39A69163FA8' + 'FD24CF5F83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3BE39E772C' + '180E86039B2783A2EC07A28FB5C55DF06F4C52C9DE2BCBF695581718' + '3995497CEA956AE515D2261898FA051015728E5A8AAAC42DAD33170D' + '04507A33A85521ABDF1CBA64ECFB850458DBEF0A8AEA71575D060C7D' + 'B3970F85A6E1E4C7ABF5AE8CDB0933D71E8C94E04A25619DCEE3D226' + '1AD2EE6BF12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB3143DB5BFC' + 'E0FD108E4B82D120A92108011A723C12A787E6D788719A10BDBA5B26' + '99C327186AF4E23C1A946834B6150BDA2583E9CA2AD44CE8DBBBC2DB' + '04DE8EF92E8EFC141FBECAA6287C59474E6BC05D99B2964FA090C3A2' + '233BA186515BE7ED1F612970CEE2D7AFB81BDD762170481CD0069127' + 'D5B05AA993B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934028492' + '36C3FAB4D27C7026C1D4DCB2602646DEC9751E763DBA37BDF8FF9406' + 'AD9E530EE5DB382F413001AEB06A53ED9027D831179727B0865A8918' + 'DA3EDBEBCF9B14ED44CE6CBACED4BB1BDB7F1447E6CC254B33205151' + '2BD7AF426FB8F401378CD2BF5983CA01C64B92ECF032EA15D1721D03' + 'F482D7CE6E74FEF6D55E702F46980C82B5A84031900B1C9E59E7C97F' + 'BEC7E8F323A97A7E36CC88BE0F1D45B7FF585AC54BD407B22B4154AA' + 'CC8F6D7EBF48E1D814CC5ED20F8037E0A79715EEF29BE32806A1D58B' + 'B7C5DA76F550AA3D8A1FBFF0EB19CCB1A313D55CDA56C9EC2EF29632' + '387FE8D76E3C0468043E8F663F4860EE12BF2D5B0B7474D6E694F91E' + '6DCC4024FFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 8192-bit 18: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64' + 'ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7' + 'ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B' + 'F12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31' + '43DB5BFCE0FD108E4B82D120A92108011A723C12A787E6D7' + '88719A10BDBA5B2699C327186AF4E23C1A946834B6150BDA' + '2583E9CA2AD44CE8DBBBC2DB04DE8EF92E8EFC141FBECAA6' + '287C59474E6BC05D99B2964FA090C3A2233BA186515BE7ED' + '1F612970CEE2D7AFB81BDD762170481CD0069127D5B05AA9' + '93B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934028492' + '36C3FAB4D27C7026C1D4DCB2602646DEC9751E763DBA37BD' + 'F8FF9406AD9E530EE5DB382F413001AEB06A53ED9027D831' + '179727B0865A8918DA3EDBEBCF9B14ED44CE6CBACED4BB1B' + 'DB7F1447E6CC254B332051512BD7AF426FB8F401378CD2BF' + '5983CA01C64B92ECF032EA15D1721D03F482D7CE6E74FEF6' + 'D55E702F46980C82B5A84031900B1C9E59E7C97FBEC7E8F3' + '23A97A7E36CC88BE0F1D45B7FF585AC54BD407B22B4154AA' + 'CC8F6D7EBF48E1D814CC5ED20F8037E0A79715EEF29BE328' + '06A1D58BB7C5DA76F550AA3D8A1FBFF0EB19CCB1A313D55C' + 'DA56C9EC2EF29632387FE8D76E3C0468043E8F663F4860EE' + '12BF2D5B0B7474D6E694F91E6DBE115974A3926F12FEE5E4' + '38777CB6A932DF8CD8BEC4D073B931BA3BC832B68D9DD300' + '741FA7BF8AFC47ED2576F6936BA424663AAB639C5AE4F568' + '3423B4742BF1C978238F16CBE39D652DE3FDB8BEFC848AD9' + '22222E04A4037C0713EB57A81A23F0C73473FC646CEA306B' + '4BCBC8862F8385DDFA9D4B7FA2C087E879683303ED5BDD3A' + '062B3CF5B3A278A66D2A13F83F44F82DDF310EE074AB6A36' + '4597E899A0255DC164F31CC50846851DF9AB48195DED7EA1' + 'B1D510BD7EE74D73FAF36BC31ECFA268359046F4EB879F92' + '4009438B481C6CD7889A002ED5EE382BC9190DA6FC026E47' + '9558E4475677E9AA9E3050E2765694DFC81F56E880B96E71' + '60C980DD98EDD3DFFFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, } class A__ : '''simple docstring''' def __init__( self: Tuple , _SCREAMING_SNAKE_CASE: int = 14) -> None: """simple docstring""" if group not in primes: raise ValueError("Unsupported Group") __lowerCAmelCase : Optional[int] = primes[group]["prime"] __lowerCAmelCase : Union[str, Any] = primes[group]["generator"] __lowerCAmelCase : Any = int(hexlify(urandom(32)) , base=16) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> str: """simple docstring""" return hex(self.__private_key)[2:] def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> str: """simple docstring""" __lowerCAmelCase : Dict = pow(self.generator , self.__private_key , self.prime) return hex(_SCREAMING_SNAKE_CASE)[2:] def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: int) -> bool: """simple docstring""" return ( 2 <= key <= self.prime - 2 and pow(_SCREAMING_SNAKE_CASE , (self.prime - 1) // 2 , self.prime) == 1 ) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: str) -> str: """simple docstring""" __lowerCAmelCase : List[Any] = int(_SCREAMING_SNAKE_CASE , base=16) if not self.is_valid_public_key(_SCREAMING_SNAKE_CASE): raise ValueError("Invalid public key") __lowerCAmelCase : Dict = pow(_SCREAMING_SNAKE_CASE , self.__private_key , self.prime) return shaaaa(str(_SCREAMING_SNAKE_CASE).encode()).hexdigest() @staticmethod def _SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: int) -> bool: """simple docstring""" return ( 2 <= remote_public_key_str <= prime - 2 and pow(_SCREAMING_SNAKE_CASE , (prime - 1) // 2 , _SCREAMING_SNAKE_CASE) == 1 ) @staticmethod def _SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: int = 14) -> str: """simple docstring""" __lowerCAmelCase : str = int(_SCREAMING_SNAKE_CASE , base=16) __lowerCAmelCase : List[Any] = int(_SCREAMING_SNAKE_CASE , base=16) __lowerCAmelCase : Dict = primes[group]["prime"] if not DiffieHellman.is_valid_public_key_static(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE): raise ValueError("Invalid public key") __lowerCAmelCase : Dict = pow(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) return shaaaa(str(_SCREAMING_SNAKE_CASE).encode()).hexdigest() if __name__ == "__main__": import doctest doctest.testmod()
269
"""simple docstring""" import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration __snake_case : Optional[int] = 50_000 __snake_case : Dict = 5_000 __snake_case , __snake_case : Union[str, Any] = os.path.split(__file__) __snake_case : Any = os.path.join(RESULTS_BASEPATH, 'results', RESULTS_FILENAME.replace('.py', '.json')) @get_duration def _lowercase ( __snake_case ,__snake_case ) -> Dict: for i in range(__snake_case ): __lowerCAmelCase : Union[str, Any] = dataset[i] @get_duration def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> Dict: for i in range(0 ,len(__snake_case ) ,__snake_case ): __lowerCAmelCase : List[str] = dataset[i : i + batch_size] @get_duration def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> Dict: with dataset.formatted_as(type=__snake_case ): for i in range(__snake_case ): __lowerCAmelCase : Union[str, Any] = dataset[i] @get_duration def _lowercase ( __snake_case ,__snake_case ,__snake_case ,__snake_case ) -> str: with dataset.formatted_as(type=__snake_case ): for i in range(0 ,__snake_case ,__snake_case ): __lowerCAmelCase : Optional[int] = dataset[i : i + batch_size] def _lowercase ( ) -> Union[str, Any]: __lowerCAmelCase : Optional[int] = {"num examples": SPEED_TEST_N_EXAMPLES} __lowerCAmelCase : Optional[int] = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted, {"type": "pandas", "length": SMALL_TEST}), (read_formatted, {"type": "torch", "length": SMALL_TEST}), (read_formatted, {"type": "tensorflow", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] __lowerCAmelCase : Any = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print("generating dataset" ) __lowerCAmelCase : int = datasets.Features( {"list": datasets.Sequence(datasets.Value("float32" ) ), "numbers": datasets.Value("float32" )} ) __lowerCAmelCase : str = generate_example_dataset( os.path.join(__snake_case ,"dataset.arrow" ) ,__snake_case ,num_examples=__snake_case ,seq_shapes={"list": (100,)} ,) print("first set of iterations" ) for func, kwargs in functions: print(func.__name__ ,str(__snake_case ) ) __lowerCAmelCase : str = func(__snake_case ,**__snake_case ) print("shuffling dataset" ) __lowerCAmelCase : Optional[int] = dataset.shuffle() print("Second set of iterations (after shuffling" ) for func, kwargs in functions_shuffled: print("shuffled " ,func.__name__ ,str(__snake_case ) ) __lowerCAmelCase : List[str] = func( __snake_case ,**__snake_case ) with open(__snake_case ,"wb" ) as f: f.write(json.dumps(__snake_case ).encode("utf-8" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
269
1
"""simple docstring""" from __future__ import annotations import os from collections.abc import Mapping __snake_case : Dict = tuple[int, int] class A__ : '''simple docstring''' def __init__( self: Optional[Any] , _SCREAMING_SNAKE_CASE: set[int] , _SCREAMING_SNAKE_CASE: Mapping[EdgeT, int]) -> None: """simple docstring""" __lowerCAmelCase : set[int] = vertices __lowerCAmelCase : dict[EdgeT, int] = { (min(_SCREAMING_SNAKE_CASE), max(_SCREAMING_SNAKE_CASE)): weight for edge, weight in edges.items() } def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: EdgeT , _SCREAMING_SNAKE_CASE: int) -> None: """simple docstring""" self.vertices.add(edge[0]) self.vertices.add(edge[1]) __lowerCAmelCase : str = weight def _SCREAMING_SNAKE_CASE ( self: str) -> Graph: """simple docstring""" __lowerCAmelCase : Graph = Graph({min(self.vertices)} , {}) __lowerCAmelCase : EdgeT __lowerCAmelCase : int __lowerCAmelCase : EdgeT __lowerCAmelCase : int while len(subgraph.vertices) < len(self.vertices): __lowerCAmelCase : List[Any] = max(self.edges.values()) + 1 for edge, weight in self.edges.items(): if (edge[0] in subgraph.vertices) ^ (edge[1] in subgraph.vertices): if weight < min_weight: __lowerCAmelCase : str = edge __lowerCAmelCase : List[Any] = weight subgraph.add_edge(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) return subgraph def _lowercase ( __snake_case = "p107_network.txt" ) -> int: __lowerCAmelCase : str = os.path.abspath(os.path.dirname(__snake_case ) ) __lowerCAmelCase : str = os.path.join(__snake_case ,__snake_case ) __lowerCAmelCase : dict[EdgeT, int] = {} __lowerCAmelCase : list[str] __lowerCAmelCase : int __lowerCAmelCase : int with open(__snake_case ) as f: __lowerCAmelCase : List[str] = f.read().strip().split("\n" ) __lowerCAmelCase : Optional[int] = [line.split("," ) for line in data] for edgea in range(1 ,len(__snake_case ) ): for edgea in range(__snake_case ): if adjaceny_matrix[edgea][edgea] != "-": __lowerCAmelCase : Dict = int(adjaceny_matrix[edgea][edgea] ) __lowerCAmelCase : Graph = Graph(set(range(len(__snake_case ) ) ) ,__snake_case ) __lowerCAmelCase : Graph = graph.prims_algorithm() __lowerCAmelCase : int = sum(graph.edges.values() ) __lowerCAmelCase : int = sum(subgraph.edges.values() ) return initial_total - optimal_total if __name__ == "__main__": print(F"""{solution() = }""")
269
"""simple docstring""" import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self: Tuple , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: List[str]=13 , _SCREAMING_SNAKE_CASE: Tuple=7 , _SCREAMING_SNAKE_CASE: int=True , _SCREAMING_SNAKE_CASE: Optional[Any]=True , _SCREAMING_SNAKE_CASE: str=False , _SCREAMING_SNAKE_CASE: Optional[Any]=True , _SCREAMING_SNAKE_CASE: int=99 , _SCREAMING_SNAKE_CASE: int=32 , _SCREAMING_SNAKE_CASE: List[str]=5 , _SCREAMING_SNAKE_CASE: Union[str, Any]=4 , _SCREAMING_SNAKE_CASE: int=64 , _SCREAMING_SNAKE_CASE: List[str]="gelu" , _SCREAMING_SNAKE_CASE: str=0.1 , _SCREAMING_SNAKE_CASE: Any=0.1 , _SCREAMING_SNAKE_CASE: Optional[int]=512 , _SCREAMING_SNAKE_CASE: Tuple=16 , _SCREAMING_SNAKE_CASE: Any=2 , _SCREAMING_SNAKE_CASE: List[str]=0.02 , _SCREAMING_SNAKE_CASE: Tuple=3 , _SCREAMING_SNAKE_CASE: Optional[Any]=4 , _SCREAMING_SNAKE_CASE: int=None , _SCREAMING_SNAKE_CASE: int=2 , _SCREAMING_SNAKE_CASE: str=2 , _SCREAMING_SNAKE_CASE: Union[str, Any]=2 , _SCREAMING_SNAKE_CASE: List[Any]=2 , _SCREAMING_SNAKE_CASE: int=4 , _SCREAMING_SNAKE_CASE: List[str]=1 , ) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : List[str] = parent __lowerCAmelCase : Optional[Any] = batch_size __lowerCAmelCase : Union[str, Any] = seq_length __lowerCAmelCase : Optional[Any] = is_training __lowerCAmelCase : Optional[int] = use_input_mask __lowerCAmelCase : Dict = use_token_type_ids __lowerCAmelCase : Dict = use_labels __lowerCAmelCase : Dict = vocab_size __lowerCAmelCase : Tuple = hidden_size __lowerCAmelCase : List[Any] = num_hidden_layers __lowerCAmelCase : Union[str, Any] = num_attention_heads __lowerCAmelCase : Tuple = intermediate_size __lowerCAmelCase : List[Any] = hidden_act __lowerCAmelCase : Optional[Any] = hidden_dropout_prob __lowerCAmelCase : Optional[Any] = attention_probs_dropout_prob __lowerCAmelCase : Optional[int] = max_position_embeddings __lowerCAmelCase : Union[str, Any] = type_vocab_size __lowerCAmelCase : Optional[int] = type_sequence_label_size __lowerCAmelCase : Dict = initializer_range __lowerCAmelCase : Tuple = num_labels __lowerCAmelCase : Optional[Any] = num_choices __lowerCAmelCase : Union[str, Any] = scope __lowerCAmelCase : Optional[Any] = q_groups __lowerCAmelCase : Optional[int] = k_groups __lowerCAmelCase : Any = v_groups __lowerCAmelCase : int = post_attention_groups __lowerCAmelCase : List[str] = intermediate_groups __lowerCAmelCase : Optional[Any] = output_groups def _SCREAMING_SNAKE_CASE ( self: Dict) -> List[str]: """simple docstring""" __lowerCAmelCase : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) __lowerCAmelCase : Union[str, Any] = None if self.use_input_mask: __lowerCAmelCase : Tuple = random_attention_mask([self.batch_size, self.seq_length]) __lowerCAmelCase : Optional[int] = None __lowerCAmelCase : List[Any] = None __lowerCAmelCase : str = None if self.use_labels: __lowerCAmelCase : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size) __lowerCAmelCase : str = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) __lowerCAmelCase : str = ids_tensor([self.batch_size] , self.num_choices) __lowerCAmelCase : Any = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _SCREAMING_SNAKE_CASE ( self: Tuple) -> int: """simple docstring""" return SqueezeBertConfig( embedding_size=self.hidden_size , vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , attention_probs_dropout_prob=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , q_groups=self.q_groups , k_groups=self.k_groups , v_groups=self.v_groups , post_attention_groups=self.post_attention_groups , intermediate_groups=self.intermediate_groups , output_groups=self.output_groups , ) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Tuple , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Tuple) -> Tuple: """simple docstring""" __lowerCAmelCase : List[Any] = SqueezeBertModel(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : List[Any] = model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = model(_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Tuple) -> Dict: """simple docstring""" __lowerCAmelCase : int = SqueezeBertForMaskedLM(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : Dict = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def _SCREAMING_SNAKE_CASE ( self: Optional[int] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Union[str, Any]) -> int: """simple docstring""" __lowerCAmelCase : str = SqueezeBertForQuestionAnswering(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : Union[str, Any] = model( _SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , start_positions=_SCREAMING_SNAKE_CASE , end_positions=_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: Tuple , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: Tuple) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : List[Any] = self.num_labels __lowerCAmelCase : Union[str, Any] = SqueezeBertForSequenceClassification(_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : int = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[int]) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : Dict = self.num_labels __lowerCAmelCase : Optional[int] = SqueezeBertForTokenClassification(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : List[str] = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: int) -> Tuple: """simple docstring""" __lowerCAmelCase : List[str] = self.num_choices __lowerCAmelCase : str = SqueezeBertForMultipleChoice(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : int = input_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() __lowerCAmelCase : Union[str, Any] = input_mask.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() __lowerCAmelCase : str = model( _SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices)) def _SCREAMING_SNAKE_CASE ( self: str) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[int] = self.prepare_config_and_inputs() ((__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase)) : Union[str, Any] = config_and_inputs __lowerCAmelCase : int = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) SCREAMING_SNAKE_CASE = ( { 'feature-extraction': SqueezeBertModel, 'fill-mask': SqueezeBertForMaskedLM, 'question-answering': SqueezeBertForQuestionAnswering, 'text-classification': SqueezeBertForSequenceClassification, 'token-classification': SqueezeBertForTokenClassification, 'zero-shot': SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = False def _SCREAMING_SNAKE_CASE ( self: Dict) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Any = SqueezeBertModelTester(self) __lowerCAmelCase : Optional[int] = ConfigTester(self , config_class=_SCREAMING_SNAKE_CASE , dim=37) def _SCREAMING_SNAKE_CASE ( self: Any) -> Tuple: """simple docstring""" self.config_tester.run_common_tests() def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> Any: """simple docstring""" __lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> List[Any]: """simple docstring""" __lowerCAmelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> int: """simple docstring""" __lowerCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Any) -> int: """simple docstring""" __lowerCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> str: """simple docstring""" __lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*_SCREAMING_SNAKE_CASE) @slow def _SCREAMING_SNAKE_CASE ( self: Any) -> Dict: """simple docstring""" for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase : Optional[Any] = SqueezeBertModel.from_pretrained(_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) @require_sentencepiece @require_tokenizers @require_torch class A__ ( unittest.TestCase ): '''simple docstring''' @slow def _SCREAMING_SNAKE_CASE ( self: int) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[Any] = SqueezeBertForSequenceClassification.from_pretrained("squeezebert/squeezebert-mnli") __lowerCAmelCase : List[Any] = torch.tensor([[1, 2_9414, 232, 328, 740, 1140, 1_2695, 69, 13, 1588, 2]]) __lowerCAmelCase : List[Any] = model(_SCREAMING_SNAKE_CASE)[0] __lowerCAmelCase : Any = torch.Size((1, 3)) self.assertEqual(output.shape , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = torch.tensor([[0.6401, -0.0349, -0.6041]]) self.assertTrue(torch.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=1e-4))
269
1
"""simple docstring""" import argparse import json import re from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileNetVaConfig, MobileNetVaForImageClassification, MobileNetVaImageProcessor, load_tf_weights_in_mobilenet_va, ) from transformers.utils import logging logging.set_verbosity_info() __snake_case : Dict = logging.get_logger(__name__) def _lowercase ( __snake_case ) -> List[Any]: __lowerCAmelCase : str = MobileNetVaConfig(layer_norm_eps=0.001 ) if "_quant" in model_name: raise ValueError("Quantized models are not supported." ) __lowerCAmelCase : str = re.match(r"^mobilenet_v1_([^_]*)_([^_]*)$" ,__snake_case ) if matches: __lowerCAmelCase : List[Any] = float(matches[1] ) __lowerCAmelCase : Tuple = int(matches[2] ) # The TensorFlow version of MobileNetV1 predicts 1001 classes instead of # the usual 1000. The first class (index 0) is "background". __lowerCAmelCase : Optional[int] = 1_001 __lowerCAmelCase : str = "imagenet-1k-id2label.json" __lowerCAmelCase : Optional[Any] = "huggingface/label-files" __lowerCAmelCase : Dict = json.load(open(hf_hub_download(__snake_case ,__snake_case ,repo_type="dataset" ) ,"r" ) ) __lowerCAmelCase : List[str] = {int(__snake_case ) + 1: v for k, v in idalabel.items()} __lowerCAmelCase : str = "background" __lowerCAmelCase : int = idalabel __lowerCAmelCase : int = {v: k for k, v in idalabel.items()} return config def _lowercase ( ) -> List[Any]: __lowerCAmelCase : int = "http://images.cocodataset.org/val2017/000000039769.jpg" __lowerCAmelCase : Optional[Any] = Image.open(requests.get(__snake_case ,stream=__snake_case ).raw ) return im @torch.no_grad() def _lowercase ( __snake_case ,__snake_case ,__snake_case ,__snake_case=False ) -> List[str]: __lowerCAmelCase : List[Any] = get_mobilenet_va_config(__snake_case ) # Load 🤗 model __lowerCAmelCase : Tuple = MobileNetVaForImageClassification(__snake_case ).eval() # Load weights from TensorFlow checkpoint load_tf_weights_in_mobilenet_va(__snake_case ,__snake_case ,__snake_case ) # Check outputs on an image, prepared by MobileNetV1ImageProcessor __lowerCAmelCase : Any = MobileNetVaImageProcessor( crop_size={"width": config.image_size, "height": config.image_size} ,size={"shortest_edge": config.image_size + 32} ,) __lowerCAmelCase : Tuple = image_processor(images=prepare_img() ,return_tensors="pt" ) __lowerCAmelCase : Any = model(**__snake_case ) __lowerCAmelCase : List[Any] = outputs.logits assert logits.shape == (1, 1_001) if model_name == "mobilenet_v1_1.0_224": __lowerCAmelCase : Tuple = torch.tensor([-4.1739, -1.1233, 3.1205] ) elif model_name == "mobilenet_v1_0.75_192": __lowerCAmelCase : Any = torch.tensor([-3.9440, -2.3141, -0.3333] ) else: __lowerCAmelCase : Union[str, Any] = None if expected_logits is not None: assert torch.allclose(logits[0, :3] ,__snake_case ,atol=1e-4 ) Path(__snake_case ).mkdir(exist_ok=__snake_case ) print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__snake_case ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__snake_case ) if push_to_hub: print("Pushing to the hub..." ) __lowerCAmelCase : str = "google/" + model_name image_processor.push_to_hub(__snake_case ) model.push_to_hub(__snake_case ) if __name__ == "__main__": __snake_case : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='mobilenet_v1_1.0_224', type=str, help='Name of the MobileNetV1 model you\'d like to convert. Should in the form \'mobilenet_v1_<depth>_<size>\'.', ) parser.add_argument( '--checkpoint_path', required=True, type=str, help='Path to the original TensorFlow checkpoint (.ckpt file).' ) parser.add_argument( '--pytorch_dump_folder_path', required=True, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model to the 🤗 hub.' ) __snake_case : Tuple = parser.parse_args() convert_movilevit_checkpoint( args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
269
"""simple docstring""" import os from math import logaa def _lowercase ( __snake_case = "base_exp.txt" ) -> int: __lowerCAmelCase : float = 0 __lowerCAmelCase : Any = 0 for i, line in enumerate(open(os.path.join(os.path.dirname(__snake_case ) ,__snake_case ) ) ): __lowerCAmelCase , __lowerCAmelCase : List[str] = list(map(__snake_case ,line.split("," ) ) ) if x * logaa(__snake_case ) > largest: __lowerCAmelCase : Tuple = x * logaa(__snake_case ) __lowerCAmelCase : Optional[Any] = i + 1 return result if __name__ == "__main__": print(solution())
269
1
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: __snake_case : Optional[Any] = None __snake_case : Tuple = logging.get_logger(__name__) __snake_case : str = {'vocab_file': 'sentencepiece.bpe.model', 'tokenizer_file': 'tokenizer.json'} __snake_case : int = { 'vocab_file': { 'facebook/nllb-200-distilled-600M': ( 'https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model' ), }, 'tokenizer_file': { 'facebook/nllb-200-distilled-600M': ( 'https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json' ), }, } __snake_case : Any = { 'facebook/nllb-large-en-ro': 1_024, 'facebook/nllb-200-distilled-600M': 1_024, } # fmt: off __snake_case : List[Any] = ['ace_Arab', 'ace_Latn', 'acm_Arab', 'acq_Arab', 'aeb_Arab', 'afr_Latn', 'ajp_Arab', 'aka_Latn', 'amh_Ethi', 'apc_Arab', 'arb_Arab', 'ars_Arab', 'ary_Arab', 'arz_Arab', 'asm_Beng', 'ast_Latn', 'awa_Deva', 'ayr_Latn', 'azb_Arab', 'azj_Latn', 'bak_Cyrl', 'bam_Latn', 'ban_Latn', 'bel_Cyrl', 'bem_Latn', 'ben_Beng', 'bho_Deva', 'bjn_Arab', 'bjn_Latn', 'bod_Tibt', 'bos_Latn', 'bug_Latn', 'bul_Cyrl', 'cat_Latn', 'ceb_Latn', 'ces_Latn', 'cjk_Latn', 'ckb_Arab', 'crh_Latn', 'cym_Latn', 'dan_Latn', 'deu_Latn', 'dik_Latn', 'dyu_Latn', 'dzo_Tibt', 'ell_Grek', 'eng_Latn', 'epo_Latn', 'est_Latn', 'eus_Latn', 'ewe_Latn', 'fao_Latn', 'pes_Arab', 'fij_Latn', 'fin_Latn', 'fon_Latn', 'fra_Latn', 'fur_Latn', 'fuv_Latn', 'gla_Latn', 'gle_Latn', 'glg_Latn', 'grn_Latn', 'guj_Gujr', 'hat_Latn', 'hau_Latn', 'heb_Hebr', 'hin_Deva', 'hne_Deva', 'hrv_Latn', 'hun_Latn', 'hye_Armn', 'ibo_Latn', 'ilo_Latn', 'ind_Latn', 'isl_Latn', 'ita_Latn', 'jav_Latn', 'jpn_Jpan', 'kab_Latn', 'kac_Latn', 'kam_Latn', 'kan_Knda', 'kas_Arab', 'kas_Deva', 'kat_Geor', 'knc_Arab', 'knc_Latn', 'kaz_Cyrl', 'kbp_Latn', 'kea_Latn', 'khm_Khmr', 'kik_Latn', 'kin_Latn', 'kir_Cyrl', 'kmb_Latn', 'kon_Latn', 'kor_Hang', 'kmr_Latn', 'lao_Laoo', 'lvs_Latn', 'lij_Latn', 'lim_Latn', 'lin_Latn', 'lit_Latn', 'lmo_Latn', 'ltg_Latn', 'ltz_Latn', 'lua_Latn', 'lug_Latn', 'luo_Latn', 'lus_Latn', 'mag_Deva', 'mai_Deva', 'mal_Mlym', 'mar_Deva', 'min_Latn', 'mkd_Cyrl', 'plt_Latn', 'mlt_Latn', 'mni_Beng', 'khk_Cyrl', 'mos_Latn', 'mri_Latn', 'zsm_Latn', 'mya_Mymr', 'nld_Latn', 'nno_Latn', 'nob_Latn', 'npi_Deva', 'nso_Latn', 'nus_Latn', 'nya_Latn', 'oci_Latn', 'gaz_Latn', 'ory_Orya', 'pag_Latn', 'pan_Guru', 'pap_Latn', 'pol_Latn', 'por_Latn', 'prs_Arab', 'pbt_Arab', 'quy_Latn', 'ron_Latn', 'run_Latn', 'rus_Cyrl', 'sag_Latn', 'san_Deva', 'sat_Beng', 'scn_Latn', 'shn_Mymr', 'sin_Sinh', 'slk_Latn', 'slv_Latn', 'smo_Latn', 'sna_Latn', 'snd_Arab', 'som_Latn', 'sot_Latn', 'spa_Latn', 'als_Latn', 'srd_Latn', 'srp_Cyrl', 'ssw_Latn', 'sun_Latn', 'swe_Latn', 'swh_Latn', 'szl_Latn', 'tam_Taml', 'tat_Cyrl', 'tel_Telu', 'tgk_Cyrl', 'tgl_Latn', 'tha_Thai', 'tir_Ethi', 'taq_Latn', 'taq_Tfng', 'tpi_Latn', 'tsn_Latn', 'tso_Latn', 'tuk_Latn', 'tum_Latn', 'tur_Latn', 'twi_Latn', 'tzm_Tfng', 'uig_Arab', 'ukr_Cyrl', 'umb_Latn', 'urd_Arab', 'uzn_Latn', 'vec_Latn', 'vie_Latn', 'war_Latn', 'wol_Latn', 'xho_Latn', 'ydd_Hebr', 'yor_Latn', 'yue_Hant', 'zho_Hans', 'zho_Hant', 'zul_Latn'] class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE = ['input_ids', 'attention_mask'] SCREAMING_SNAKE_CASE = NllbTokenizer SCREAMING_SNAKE_CASE = [] SCREAMING_SNAKE_CASE = [] def __init__( self: List[str] , _SCREAMING_SNAKE_CASE: str=None , _SCREAMING_SNAKE_CASE: Tuple=None , _SCREAMING_SNAKE_CASE: str="<s>" , _SCREAMING_SNAKE_CASE: Union[str, Any]="</s>" , _SCREAMING_SNAKE_CASE: List[str]="</s>" , _SCREAMING_SNAKE_CASE: Any="<s>" , _SCREAMING_SNAKE_CASE: List[str]="<unk>" , _SCREAMING_SNAKE_CASE: int="<pad>" , _SCREAMING_SNAKE_CASE: int="<mask>" , _SCREAMING_SNAKE_CASE: Dict=None , _SCREAMING_SNAKE_CASE: Any=None , _SCREAMING_SNAKE_CASE: Any=None , _SCREAMING_SNAKE_CASE: Optional[int]=False , **_SCREAMING_SNAKE_CASE: Union[str, Any] , ) -> Tuple: """simple docstring""" __lowerCAmelCase : Optional[int] = AddedToken(_SCREAMING_SNAKE_CASE , lstrip=_SCREAMING_SNAKE_CASE , rstrip=_SCREAMING_SNAKE_CASE) if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) else mask_token __lowerCAmelCase : Union[str, Any] = legacy_behaviour super().__init__( vocab_file=_SCREAMING_SNAKE_CASE , tokenizer_file=_SCREAMING_SNAKE_CASE , bos_token=_SCREAMING_SNAKE_CASE , eos_token=_SCREAMING_SNAKE_CASE , sep_token=_SCREAMING_SNAKE_CASE , cls_token=_SCREAMING_SNAKE_CASE , unk_token=_SCREAMING_SNAKE_CASE , pad_token=_SCREAMING_SNAKE_CASE , mask_token=_SCREAMING_SNAKE_CASE , src_lang=_SCREAMING_SNAKE_CASE , tgt_lang=_SCREAMING_SNAKE_CASE , additional_special_tokens=_SCREAMING_SNAKE_CASE , legacy_behaviour=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Tuple = vocab_file __lowerCAmelCase : int = False if not self.vocab_file else True __lowerCAmelCase : Optional[Any] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens]) self.add_special_tokens({"additional_special_tokens": _additional_special_tokens}) __lowerCAmelCase : Dict = { lang_code: self.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE) for lang_code in FAIRSEQ_LANGUAGE_CODES } __lowerCAmelCase : Dict = src_lang if src_lang is not None else "eng_Latn" __lowerCAmelCase : Dict = self.convert_tokens_to_ids(self._src_lang) __lowerCAmelCase : Any = tgt_lang self.set_src_lang_special_tokens(self._src_lang) @property def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> str: """simple docstring""" return self._src_lang @src_lang.setter def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: str) -> None: """simple docstring""" __lowerCAmelCase : Tuple = new_src_lang self.set_src_lang_special_tokens(self._src_lang) def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: List[int] , _SCREAMING_SNAKE_CASE: Optional[List[int]] = None) -> List[int]: """simple docstring""" if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: List[int] , _SCREAMING_SNAKE_CASE: Optional[List[int]] = None) -> List[int]: """simple docstring""" __lowerCAmelCase : List[Any] = [self.sep_token_id] __lowerCAmelCase : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0] def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: Optional[str] , _SCREAMING_SNAKE_CASE: Optional[str] , **_SCREAMING_SNAKE_CASE: List[Any]) -> int: """simple docstring""" if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model") __lowerCAmelCase : Dict = src_lang __lowerCAmelCase : List[str] = self(_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE , return_tensors=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = self.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = tgt_lang_id return inputs def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: str = "eng_Latn" , _SCREAMING_SNAKE_CASE: Optional[List[str]] = None , _SCREAMING_SNAKE_CASE: str = "fra_Latn" , **_SCREAMING_SNAKE_CASE: List[Any] , ) -> BatchEncoding: """simple docstring""" __lowerCAmelCase : Any = src_lang __lowerCAmelCase : Union[str, Any] = tgt_lang return super().prepare_seqaseq_batch(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> Dict: """simple docstring""" return self.set_src_lang_special_tokens(self.src_lang) def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> List[Any]: """simple docstring""" return self.set_tgt_lang_special_tokens(self.tgt_lang) def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: Optional[int]) -> None: """simple docstring""" __lowerCAmelCase : List[str] = self.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE) if self.legacy_behaviour: __lowerCAmelCase : str = [] __lowerCAmelCase : Union[str, Any] = [self.eos_token_id, self.cur_lang_code] else: __lowerCAmelCase : str = [self.cur_lang_code] __lowerCAmelCase : List[str] = [self.eos_token_id] __lowerCAmelCase : Union[str, Any] = self.convert_ids_to_tokens(self.prefix_tokens) __lowerCAmelCase : List[str] = self.convert_ids_to_tokens(self.suffix_tokens) __lowerCAmelCase : Optional[int] = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens)) , ) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: str) -> None: """simple docstring""" __lowerCAmelCase : Any = self.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE) if self.legacy_behaviour: __lowerCAmelCase : str = [] __lowerCAmelCase : str = [self.eos_token_id, self.cur_lang_code] else: __lowerCAmelCase : Tuple = [self.cur_lang_code] __lowerCAmelCase : List[Any] = [self.eos_token_id] __lowerCAmelCase : str = self.convert_ids_to_tokens(self.prefix_tokens) __lowerCAmelCase : Union[str, Any] = self.convert_ids_to_tokens(self.suffix_tokens) __lowerCAmelCase : Optional[int] = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens)) , ) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: Optional[str] = None) -> Tuple[str]: """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer.") if not os.path.isdir(_SCREAMING_SNAKE_CASE): logger.error(F"""Vocabulary path ({save_directory}) should be a directory.""") return __lowerCAmelCase : str = os.path.join( _SCREAMING_SNAKE_CASE , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(_SCREAMING_SNAKE_CASE): copyfile(self.vocab_file , _SCREAMING_SNAKE_CASE) return (out_vocab_file,)
269
"""simple docstring""" import collections import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_flax_cross_test, require_flax, require_torch, require_vision, slow, torch_device, ) from transformers.utils import is_flax_available, is_torch_available, is_vision_available from ...test_modeling_flax_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_flax_bert import FlaxBertModelTester from ..clip.test_modeling_flax_clip import FlaxCLIPVisionModelTester from ..vit.test_modeling_flax_vit import FlaxViTModelTester if is_flax_available(): from transformers import ( FlaxBertModel, FlaxCLIPVisionModel, FlaxVisionTextDualEncoderModel, FlaxViTModel, VisionTextDualEncoderConfig, VisionTextDualEncoderProcessor, ) from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) if is_torch_available(): import torch from transformers import VisionTextDualEncoderModel if is_vision_available(): from PIL import Image def _lowercase ( __snake_case ) -> List[str]: if isinstance(__snake_case ,collections.abc.Iterable ): return x return (x, x) @require_flax class A__ : '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Optional[Any]) -> str: """simple docstring""" pass def _SCREAMING_SNAKE_CASE ( self: str) -> int: """simple docstring""" pass def _SCREAMING_SNAKE_CASE ( self: List[str]) -> Tuple: """simple docstring""" pass def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: np.ndarray , _SCREAMING_SNAKE_CASE: np.ndarray , _SCREAMING_SNAKE_CASE: float) -> List[Any]: """simple docstring""" __lowerCAmelCase : str = np.abs((a - b)).max() self.assertLessEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , F"""Difference between torch and flax is {diff} (>= {tol}).""") def _SCREAMING_SNAKE_CASE ( self: str , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Any=None , **_SCREAMING_SNAKE_CASE: Tuple) -> List[str]: """simple docstring""" __lowerCAmelCase : Tuple = VisionTextDualEncoderConfig.from_vision_text_configs(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = FlaxVisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = model(input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], config.projection_dim)) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], config.projection_dim)) def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Optional[Any]=None , **_SCREAMING_SNAKE_CASE: List[str]) -> Dict: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : List[Any] = self.get_vision_text_model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = {"vision_model": vision_model, "text_model": text_model} __lowerCAmelCase : Optional[Any] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = model(input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], model.config.projection_dim)) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], model.config.projection_dim)) def _SCREAMING_SNAKE_CASE ( self: Dict , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: List[Any]=None , **_SCREAMING_SNAKE_CASE: Tuple) -> str: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : Optional[Any] = self.get_vision_text_model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = {"vision_model": vision_model, "text_model": text_model} __lowerCAmelCase : Any = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = model(input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = output[0] with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = FlaxVisionTextDualEncoderModel.from_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = model(input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = after_output[0] __lowerCAmelCase : Any = np.amax(np.abs(out_a - out_a)) self.assertLessEqual(_SCREAMING_SNAKE_CASE , 1e-3) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Any=None , **_SCREAMING_SNAKE_CASE: List[str]) -> Optional[int]: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : Any = self.get_vision_text_model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = {"vision_model": vision_model, "text_model": text_model} __lowerCAmelCase : Optional[Any] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = model( input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , output_attentions=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = output.vision_model_output.attentions self.assertEqual(len(_SCREAMING_SNAKE_CASE) , vision_config.num_hidden_layers) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) __lowerCAmelCase : List[str] = to_atuple(vision_model.config.image_size) __lowerCAmelCase : Any = to_atuple(vision_model.config.patch_size) __lowerCAmelCase : Union[str, Any] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) __lowerCAmelCase : Tuple = num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len)) __lowerCAmelCase : Union[str, Any] = output.text_model_output.attentions self.assertEqual(len(_SCREAMING_SNAKE_CASE) , text_config.num_hidden_layers) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: int) -> str: """simple docstring""" pt_model.to(_SCREAMING_SNAKE_CASE) pt_model.eval() # prepare inputs __lowerCAmelCase : Union[str, Any] = inputs_dict __lowerCAmelCase : Union[str, Any] = {k: torch.tensor(v.tolist()) for k, v in flax_inputs.items()} with torch.no_grad(): __lowerCAmelCase : Any = pt_model(**_SCREAMING_SNAKE_CASE).to_tuple() __lowerCAmelCase : List[Any] = fx_model(**_SCREAMING_SNAKE_CASE).to_tuple() self.assertEqual(len(_SCREAMING_SNAKE_CASE) , len(_SCREAMING_SNAKE_CASE) , "Output lengths differ between Flax and PyTorch") for fx_output, pt_output in zip(fx_outputs[:4] , pt_outputs[:4]): self.assert_almost_equals(_SCREAMING_SNAKE_CASE , pt_output.numpy() , 4e-2) # PT -> Flax with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = FlaxVisionTextDualEncoderModel.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = fx_model_loaded(**_SCREAMING_SNAKE_CASE).to_tuple() self.assertEqual(len(_SCREAMING_SNAKE_CASE) , len(_SCREAMING_SNAKE_CASE) , "Output lengths differ between Flax and PyTorch") for fx_output_loaded, pt_output in zip(fx_outputs_loaded[:4] , pt_outputs[:4]): self.assert_almost_equals(_SCREAMING_SNAKE_CASE , pt_output.numpy() , 4e-2) # Flax -> PT with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = VisionTextDualEncoderModel.from_pretrained(_SCREAMING_SNAKE_CASE , from_flax=_SCREAMING_SNAKE_CASE) pt_model_loaded.to(_SCREAMING_SNAKE_CASE) pt_model_loaded.eval() with torch.no_grad(): __lowerCAmelCase : Optional[Any] = pt_model_loaded(**_SCREAMING_SNAKE_CASE).to_tuple() self.assertEqual(len(_SCREAMING_SNAKE_CASE) , len(_SCREAMING_SNAKE_CASE) , "Output lengths differ between Flax and PyTorch") for fx_output, pt_output_loaded in zip(fx_outputs[:4] , pt_outputs_loaded[:4]): self.assert_almost_equals(_SCREAMING_SNAKE_CASE , pt_output_loaded.numpy() , 4e-2) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Dict) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Union[str, Any] = VisionTextDualEncoderConfig.from_vision_text_configs(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = VisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = FlaxVisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = fx_state self.check_pt_flax_equivalence(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Tuple , _SCREAMING_SNAKE_CASE: Dict) -> str: """simple docstring""" __lowerCAmelCase : Optional[int] = VisionTextDualEncoderConfig.from_vision_text_configs(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = VisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = FlaxVisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = load_flax_weights_in_pytorch_model(_SCREAMING_SNAKE_CASE , fx_model.params) self.check_pt_flax_equivalence(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> str: """simple docstring""" __lowerCAmelCase : List[str] = self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Dict) -> int: """simple docstring""" __lowerCAmelCase : List[Any] = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> List[Any]: """simple docstring""" __lowerCAmelCase : Dict = self.prepare_config_and_inputs() self.check_save_load(**_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: int) -> Dict: """simple docstring""" __lowerCAmelCase : List[str] = self.prepare_config_and_inputs() self.check_vision_text_output_attention(**_SCREAMING_SNAKE_CASE) @is_pt_flax_cross_test def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> Any: """simple docstring""" __lowerCAmelCase : Dict = self.prepare_config_and_inputs() __lowerCAmelCase : List[Any] = config_inputs_dict.pop("vision_config") __lowerCAmelCase : str = config_inputs_dict.pop("text_config") __lowerCAmelCase : Union[str, Any] = config_inputs_dict self.check_equivalence_pt_to_flax(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) self.check_equivalence_flax_to_pt(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) @slow def _SCREAMING_SNAKE_CASE ( self: str) -> Dict: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : Dict = self.get_pretrained_model_and_inputs() __lowerCAmelCase : Union[str, Any] = model_a(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = outputs[0] with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = FlaxVisionTextDualEncoderModel.from_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = model_a(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = after_outputs[0] __lowerCAmelCase : List[Any] = np.amax(np.abs(out_a - out_a)) self.assertLessEqual(_SCREAMING_SNAKE_CASE , 1e-5) @require_flax class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: List[str]) -> List[Any]: """simple docstring""" __lowerCAmelCase : Any = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( "hf-internal-testing/tiny-random-vit" , "hf-internal-testing/tiny-bert" , vision_from_pt=_SCREAMING_SNAKE_CASE , text_from_pt=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Union[str, Any] = 13 __lowerCAmelCase : Optional[int] = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ]) __lowerCAmelCase : List[Any] = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size) __lowerCAmelCase : List[Any] = random_attention_mask([batch_size, 4]) __lowerCAmelCase : str = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def _SCREAMING_SNAKE_CASE ( self: int , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Union[str, Any]) -> Optional[int]: """simple docstring""" __lowerCAmelCase : List[str] = FlaxViTModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = FlaxBertModel(_SCREAMING_SNAKE_CASE) return vision_model, text_model def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> int: """simple docstring""" __lowerCAmelCase : List[Any] = FlaxViTModelTester(self) __lowerCAmelCase : Optional[Any] = FlaxBertModelTester(self) __lowerCAmelCase : int = vit_model_tester.prepare_config_and_inputs() __lowerCAmelCase : List[str] = bert_model_tester.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase : Tuple = vision_config_and_inputs __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Union[str, Any] = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_torch class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Dict) -> List[str]: """simple docstring""" __lowerCAmelCase : Optional[int] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( "hf-internal-testing/tiny-random-clip" , "hf-internal-testing/tiny-bert" , vision_from_pt=_SCREAMING_SNAKE_CASE , text_from_pt=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Optional[int] = 13 __lowerCAmelCase : List[str] = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ]) __lowerCAmelCase : Any = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size) __lowerCAmelCase : str = random_attention_mask([batch_size, 4]) __lowerCAmelCase : Optional[int] = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Union[str, Any]) -> int: """simple docstring""" __lowerCAmelCase : int = FlaxCLIPVisionModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = FlaxBertModel(_SCREAMING_SNAKE_CASE) return vision_model, text_model def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : List[Any] = FlaxCLIPVisionModelTester(self) __lowerCAmelCase : str = FlaxBertModelTester(self) __lowerCAmelCase : Optional[Any] = clip_model_tester.prepare_config_and_inputs() __lowerCAmelCase : Dict = bert_model_tester.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase : Any = vision_config_and_inputs __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : List[Any] = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_flax @require_vision class A__ ( unittest.TestCase ): '''simple docstring''' @slow def _SCREAMING_SNAKE_CASE ( self: List[str]) -> List[str]: """simple docstring""" __lowerCAmelCase : Dict = FlaxVisionTextDualEncoderModel.from_pretrained("clip-italian/clip-italian" , logit_scale_init_value=1.0) __lowerCAmelCase : str = VisionTextDualEncoderProcessor.from_pretrained("clip-italian/clip-italian") __lowerCAmelCase : Tuple = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png") __lowerCAmelCase : Optional[int] = processor( text=["una foto di un gatto", "una foto di un cane"] , images=_SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE , return_tensors="np") __lowerCAmelCase : List[str] = model(**_SCREAMING_SNAKE_CASE) # verify the logits self.assertEqual(outputs.logits_per_image.shape , (inputs.pixel_values.shape[0], inputs.input_ids.shape[0])) self.assertEqual( outputs.logits_per_text.shape , (inputs.input_ids.shape[0], inputs.pixel_values.shape[0]) , ) __lowerCAmelCase : List[str] = np.array([[1.228_4727, 0.310_4122]]) self.assertTrue(np.allclose(outputs.logits_per_image , _SCREAMING_SNAKE_CASE , atol=1e-3))
269
1
"""simple docstring""" import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> int: # Initialise PyTorch model __lowerCAmelCase : List[Any] = MobileBertConfig.from_json_file(__snake_case ) print(F"""Building PyTorch model from configuration: {config}""" ) __lowerCAmelCase : str = MobileBertForPreTraining(__snake_case ) # Load weights from tf checkpoint __lowerCAmelCase : Any = load_tf_weights_in_mobilebert(__snake_case ,__snake_case ,__snake_case ) # Save pytorch-model print(F"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() ,__snake_case ) if __name__ == "__main__": __snake_case : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--mobilebert_config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained MobileBERT model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) __snake_case : Tuple = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.mobilebert_config_file, args.pytorch_dump_path)
269
"""simple docstring""" from __future__ import annotations def _lowercase ( __snake_case ,__snake_case ,__snake_case ,__snake_case ) -> list: __lowerCAmelCase : Dict = [] __lowerCAmelCase , __lowerCAmelCase : Any = input_list[low:mid], input_list[mid : high + 1] while left and right: result.append((left if left[0] <= right[0] else right).pop(0 ) ) __lowerCAmelCase : int = result + left + right return input_list def _lowercase ( __snake_case ) -> list: if len(__snake_case ) <= 1: return input_list __lowerCAmelCase : int = list(__snake_case ) # iteration for two-way merging __lowerCAmelCase : Optional[int] = 2 while p <= len(__snake_case ): # getting low, high and middle value for merge-sort of single list for i in range(0 ,len(__snake_case ) ,__snake_case ): __lowerCAmelCase : Union[str, Any] = i __lowerCAmelCase : Tuple = i + p - 1 __lowerCAmelCase : Optional[Any] = (low + high + 1) // 2 __lowerCAmelCase : Any = merge(__snake_case ,__snake_case ,__snake_case ,__snake_case ) # final merge of last two parts if p * 2 >= len(__snake_case ): __lowerCAmelCase : Optional[Any] = i __lowerCAmelCase : Union[str, Any] = merge(__snake_case ,0 ,__snake_case ,len(__snake_case ) - 1 ) break p *= 2 return input_list if __name__ == "__main__": __snake_case : Union[str, Any] = input('Enter numbers separated by a comma:\n').strip() if user_input == "": __snake_case : Optional[int] = [] else: __snake_case : int = [int(item.strip()) for item in user_input.split(',')] print(iter_merge_sort(unsorted))
269
1
"""simple docstring""" import argparse import json from typing import List from ltp import LTP from transformers import BertTokenizer def _lowercase ( __snake_case ) -> List[Any]: # This defines a "chinese character" as anything in the CJK Unicode block: # https://en.wikipedia.org/wiki/CJK_Unified_Ideographs_(Unicode_block) # # Note that the CJK Unicode block is NOT all Japanese and Korean characters, # despite its name. The modern Korean Hangul alphabet is a different block, # as is Japanese Hiragana and Katakana. Those alphabets are used to write # space-separated words, so they are not treated specially and handled # like the all of the other languages. if ( (cp >= 0X4e00 and cp <= 0X9fff) or (cp >= 0X3400 and cp <= 0X4dbf) # or (cp >= 0X20000 and cp <= 0X2a6df) # or (cp >= 0X2a700 and cp <= 0X2b73f) # or (cp >= 0X2b740 and cp <= 0X2b81f) # or (cp >= 0X2b820 and cp <= 0X2ceaf) # or (cp >= 0Xf900 and cp <= 0Xfaff) or (cp >= 0X2f800 and cp <= 0X2fa1f) # ): # return True return False def _lowercase ( __snake_case ) -> Optional[int]: # word like '180' or '身高' or '神' for char in word: __lowerCAmelCase : Tuple = ord(__snake_case ) if not _is_chinese_char(__snake_case ): return 0 return 1 def _lowercase ( __snake_case ) -> int: __lowerCAmelCase : Dict = set() for token in tokens: __lowerCAmelCase : Any = len(__snake_case ) > 1 and is_chinese(__snake_case ) if chinese_word: word_set.add(__snake_case ) __lowerCAmelCase : Optional[int] = list(__snake_case ) return word_list def _lowercase ( __snake_case ,__snake_case ) -> List[Any]: if not chinese_word_set: return bert_tokens __lowerCAmelCase : List[Any] = max([len(__snake_case ) for w in chinese_word_set] ) __lowerCAmelCase : Optional[Any] = bert_tokens __lowerCAmelCase , __lowerCAmelCase : Tuple = 0, len(__snake_case ) while start < end: __lowerCAmelCase : Optional[int] = True if is_chinese(bert_word[start] ): __lowerCAmelCase : Dict = min(end - start ,__snake_case ) for i in range(__snake_case ,1 ,-1 ): __lowerCAmelCase : int = "".join(bert_word[start : start + i] ) if whole_word in chinese_word_set: for j in range(start + 1 ,start + i ): __lowerCAmelCase : List[Any] = "##" + bert_word[j] __lowerCAmelCase : List[str] = start + i __lowerCAmelCase : Dict = False break if single_word: start += 1 return bert_word def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> Optional[int]: __lowerCAmelCase : Dict = [] for i in range(0 ,len(__snake_case ) ,100 ): __lowerCAmelCase : Optional[int] = ltp_tokenizer.seg(lines[i : i + 100] )[0] __lowerCAmelCase : int = [get_chinese_word(__snake_case ) for r in res] ltp_res.extend(__snake_case ) assert len(__snake_case ) == len(__snake_case ) __lowerCAmelCase : Union[str, Any] = [] for i in range(0 ,len(__snake_case ) ,100 ): __lowerCAmelCase : Tuple = bert_tokenizer(lines[i : i + 100] ,add_special_tokens=__snake_case ,truncation=__snake_case ,max_length=512 ) bert_res.extend(res["input_ids"] ) assert len(__snake_case ) == len(__snake_case ) __lowerCAmelCase : List[str] = [] for input_ids, chinese_word in zip(__snake_case ,__snake_case ): __lowerCAmelCase : Union[str, Any] = [] for id in input_ids: __lowerCAmelCase : Any = bert_tokenizer._convert_id_to_token(__snake_case ) input_tokens.append(__snake_case ) __lowerCAmelCase : Any = add_sub_symbol(__snake_case ,__snake_case ) __lowerCAmelCase : Any = [] # We only save pos of chinese subwords start with ##, which mean is part of a whole word. for i, token in enumerate(__snake_case ): if token[:2] == "##": __lowerCAmelCase : Union[str, Any] = token[2:] # save chinese tokens' pos if len(__snake_case ) == 1 and _is_chinese_char(ord(__snake_case ) ): ref_id.append(__snake_case ) ref_ids.append(__snake_case ) assert len(__snake_case ) == len(__snake_case ) return ref_ids def _lowercase ( __snake_case ) -> Tuple: # For Chinese (Ro)Bert, the best result is from : RoBERTa-wwm-ext (https://github.com/ymcui/Chinese-BERT-wwm) # If we want to fine-tune these model, we have to use same tokenizer : LTP (https://github.com/HIT-SCIR/ltp) with open(args.file_name ,"r" ,encoding="utf-8" ) as f: __lowerCAmelCase : Tuple = f.readlines() __lowerCAmelCase : Dict = [line.strip() for line in data if len(__snake_case ) > 0 and not line.isspace()] # avoid delimiter like '\u2029' __lowerCAmelCase : Optional[Any] = LTP(args.ltp ) # faster in GPU device __lowerCAmelCase : str = BertTokenizer.from_pretrained(args.bert ) __lowerCAmelCase : int = prepare_ref(__snake_case ,__snake_case ,__snake_case ) with open(args.save_path ,"w" ,encoding="utf-8" ) as f: __lowerCAmelCase : Tuple = [json.dumps(__snake_case ) + "\n" for ref in ref_ids] f.writelines(__snake_case ) if __name__ == "__main__": __snake_case : int = argparse.ArgumentParser(description='prepare_chinese_ref') parser.add_argument( '--file_name', type=str, default='./resources/chinese-demo.txt', help='file need process, same as training data in lm', ) parser.add_argument( '--ltp', type=str, default='./resources/ltp', help='resources for LTP tokenizer, usually a path' ) parser.add_argument('--bert', type=str, default='./resources/robert', help='resources for Bert tokenizer') parser.add_argument('--save_path', type=str, default='./resources/ref.txt', help='path to save res') __snake_case : List[Any] = parser.parse_args() main(args)
269
"""simple docstring""" import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class A__ ( unittest.TestCase ): '''simple docstring''' @slow def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> str: """simple docstring""" __lowerCAmelCase : Optional[Any] = FlaxMTaForConditionalGeneration.from_pretrained("google/mt5-small") __lowerCAmelCase : Optional[Any] = AutoTokenizer.from_pretrained("google/mt5-small") __lowerCAmelCase : Tuple = tokenizer("Hello there" , return_tensors="np").input_ids __lowerCAmelCase : Dict = tokenizer("Hi I am" , return_tensors="np").input_ids __lowerCAmelCase : str = shift_tokens_right(_SCREAMING_SNAKE_CASE , model.config.pad_token_id , model.config.decoder_start_token_id) __lowerCAmelCase : Optional[int] = model(_SCREAMING_SNAKE_CASE , decoder_input_ids=_SCREAMING_SNAKE_CASE).logits __lowerCAmelCase : int = optax.softmax_cross_entropy(_SCREAMING_SNAKE_CASE , onehot(_SCREAMING_SNAKE_CASE , logits.shape[-1])).mean() __lowerCAmelCase : List[str] = -(labels.shape[-1] * loss.item()) __lowerCAmelCase : str = -84.9127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE) < 1e-4)
269
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_segformer import SegformerImageProcessor __snake_case : Optional[Any] = logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self: str , *_SCREAMING_SNAKE_CASE: int , **_SCREAMING_SNAKE_CASE: Optional[Any]) -> None: """simple docstring""" warnings.warn( "The class SegformerFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use SegformerImageProcessor instead." , _SCREAMING_SNAKE_CASE , ) super().__init__(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE)
269
"""simple docstring""" import re def _lowercase ( __snake_case ) -> str: if len(re.findall("[ATCG]" ,__snake_case ) ) != len(__snake_case ): raise ValueError("Invalid Strand" ) return dna.translate(dna.maketrans("ATCG" ,"TAGC" ) ) if __name__ == "__main__": import doctest doctest.testmod()
269
1
"""simple docstring""" from heapq import heappop, heappush import numpy as np def _lowercase ( __snake_case ,__snake_case ,__snake_case ,__snake_case ,) -> tuple[float | int, list[tuple[int, int]]]: __lowerCAmelCase , __lowerCAmelCase : Optional[int] = grid.shape __lowerCAmelCase : List[str] = [-1, 1, 0, 0] __lowerCAmelCase : Tuple = [0, 0, -1, 1] if allow_diagonal: dx += [-1, -1, 1, 1] dy += [-1, 1, -1, 1] __lowerCAmelCase , __lowerCAmelCase : List[Any] = [(0, source)], set() __lowerCAmelCase : int = np.full((rows, cols) ,np.inf ) __lowerCAmelCase : Optional[int] = 0 __lowerCAmelCase : Optional[Any] = np.empty((rows, cols) ,dtype=__snake_case ) __lowerCAmelCase : Union[str, Any] = None while queue: ((__lowerCAmelCase) , (__lowerCAmelCase)) : Optional[Any] = heappop(__snake_case ) if (x, y) in visited: continue visited.add((x, y) ) if (x, y) == destination: __lowerCAmelCase : Tuple = [] while (x, y) != source: path.append((x, y) ) __lowerCAmelCase , __lowerCAmelCase : Tuple = predecessors[x, y] path.append(__snake_case ) # add the source manually path.reverse() return matrix[destination], path for i in range(len(__snake_case ) ): __lowerCAmelCase , __lowerCAmelCase : Optional[Any] = x + dx[i], y + dy[i] if 0 <= nx < rows and 0 <= ny < cols: __lowerCAmelCase : Optional[int] = grid[nx][ny] if next_node == 1 and matrix[nx, ny] > dist + 1: heappush(__snake_case ,(dist + 1, (nx, ny)) ) __lowerCAmelCase : str = dist + 1 __lowerCAmelCase : Union[str, Any] = (x, y) return np.inf, [] if __name__ == "__main__": import doctest doctest.testmod()
269
"""simple docstring""" import os import unittest from transformers import FunnelTokenizer, FunnelTokenizerFast from transformers.models.funnel.tokenization_funnel import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = FunnelTokenizer SCREAMING_SNAKE_CASE = FunnelTokenizerFast SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = True def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Optional[int]: """simple docstring""" super().setUp() __lowerCAmelCase : str = [ "<unk>", "<cls>", "<sep>", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] __lowerCAmelCase : int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"]) with open(self.vocab_file , "w" , encoding="utf-8") as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens])) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , **_SCREAMING_SNAKE_CASE: Union[str, Any]) -> int: """simple docstring""" return FunnelTokenizer.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Any , **_SCREAMING_SNAKE_CASE: Any) -> str: """simple docstring""" return FunnelTokenizerFast.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: str) -> Any: """simple docstring""" __lowerCAmelCase : Union[str, Any] = "UNwant\u00E9d,running" __lowerCAmelCase : str = "unwanted, running" return input_text, output_text def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> List[str]: """simple docstring""" __lowerCAmelCase : Any = self.tokenizer_class(self.vocab_file) __lowerCAmelCase : Any = tokenizer.tokenize("UNwant\u00E9d,running") self.assertListEqual(_SCREAMING_SNAKE_CASE , ["un", "##want", "##ed", ",", "runn", "##ing"]) self.assertListEqual(tokenizer.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE) , [7, 4, 5, 10, 8, 9]) def _SCREAMING_SNAKE_CASE ( self: List[str]) -> List[str]: """simple docstring""" __lowerCAmelCase : Optional[Any] = self.get_tokenizers(do_lower_case=_SCREAMING_SNAKE_CASE) for tokenizer in tokenizers: __lowerCAmelCase : List[str] = tokenizer("UNwant\u00E9d,running") __lowerCAmelCase : Optional[int] = len(inputs["input_ids"]) - 1 self.assertListEqual(inputs["token_type_ids"] , [2] + [0] * sentence_len) __lowerCAmelCase : List[str] = tokenizer("UNwant\u00E9d,running" , "UNwant\u00E9d,running") self.assertListEqual(inputs["token_type_ids"] , [2] + [0] * sentence_len + [1] * sentence_len)
269
1
"""simple docstring""" class A__ : '''simple docstring''' def __init__( self: List[Any] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: List[Any]) -> List[Any]: """simple docstring""" __lowerCAmelCase : Tuple = None __lowerCAmelCase : List[str] = None __lowerCAmelCase : Union[str, Any] = graph self._normalize_graph(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = len(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = None def _SCREAMING_SNAKE_CASE ( self: str , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: List[Any]) -> Any: """simple docstring""" if sources is int: __lowerCAmelCase : List[Any] = [sources] if sinks is int: __lowerCAmelCase : List[Any] = [sinks] if len(_SCREAMING_SNAKE_CASE) == 0 or len(_SCREAMING_SNAKE_CASE) == 0: return __lowerCAmelCase : Optional[Any] = sources[0] __lowerCAmelCase : Union[str, Any] = sinks[0] # make fake vertex if there are more # than one source or sink if len(_SCREAMING_SNAKE_CASE) > 1 or len(_SCREAMING_SNAKE_CASE) > 1: __lowerCAmelCase : int = 0 for i in sources: max_input_flow += sum(self.graph[i]) __lowerCAmelCase : Dict = len(self.graph) + 1 for room in self.graph: room.insert(0 , 0) self.graph.insert(0 , [0] * size) for i in sources: __lowerCAmelCase : Optional[Any] = max_input_flow __lowerCAmelCase : Optional[int] = 0 __lowerCAmelCase : Dict = len(self.graph) + 1 for room in self.graph: room.append(0) self.graph.append([0] * size) for i in sinks: __lowerCAmelCase : List[str] = max_input_flow __lowerCAmelCase : Any = size - 1 def _SCREAMING_SNAKE_CASE ( self: int) -> Any: """simple docstring""" if self.maximum_flow_algorithm is None: raise Exception("You need to set maximum flow algorithm before.") if self.source_index is None or self.sink_index is None: return 0 self.maximum_flow_algorithm.execute() return self.maximum_flow_algorithm.getMaximumFlow() def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: Dict) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[int] = algorithm(self) class A__ : '''simple docstring''' def __init__( self: int , _SCREAMING_SNAKE_CASE: int) -> Any: """simple docstring""" __lowerCAmelCase : List[str] = flow_network __lowerCAmelCase : Optional[int] = flow_network.verticesCount __lowerCAmelCase : List[Any] = flow_network.sourceIndex __lowerCAmelCase : int = flow_network.sinkIndex # it's just a reference, so you shouldn't change # it in your algorithms, use deep copy before doing that __lowerCAmelCase : Any = flow_network.graph __lowerCAmelCase : List[str] = False def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Any: """simple docstring""" if not self.executed: self._algorithm() __lowerCAmelCase : Optional[Any] = True def _SCREAMING_SNAKE_CASE ( self: int) -> List[str]: """simple docstring""" pass class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self: str , _SCREAMING_SNAKE_CASE: str) -> Optional[int]: """simple docstring""" super().__init__(_SCREAMING_SNAKE_CASE) # use this to save your result __lowerCAmelCase : Optional[Any] = -1 def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> List[Any]: """simple docstring""" if not self.executed: raise Exception("You should execute algorithm before using its result!") return self.maximum_flow class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: str) -> List[str]: """simple docstring""" super().__init__(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = [[0] * self.verticies_count for i in range(self.verticies_count)] __lowerCAmelCase : Tuple = [0] * self.verticies_count __lowerCAmelCase : int = [0] * self.verticies_count def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Optional[int]: """simple docstring""" __lowerCAmelCase : Optional[int] = self.verticies_count # push some substance to graph for nextvertex_index, bandwidth in enumerate(self.graph[self.source_index]): self.preflow[self.source_index][nextvertex_index] += bandwidth self.preflow[nextvertex_index][self.source_index] -= bandwidth self.excesses[nextvertex_index] += bandwidth # Relabel-to-front selection rule __lowerCAmelCase : Dict = [ i for i in range(self.verticies_count) if i != self.source_index and i != self.sink_index ] # move through list __lowerCAmelCase : Tuple = 0 while i < len(_SCREAMING_SNAKE_CASE): __lowerCAmelCase : List[Any] = vertices_list[i] __lowerCAmelCase : Union[str, Any] = self.heights[vertex_index] self.process_vertex(_SCREAMING_SNAKE_CASE) if self.heights[vertex_index] > previous_height: # if it was relabeled, swap elements # and start from 0 index vertices_list.insert(0 , vertices_list.pop(_SCREAMING_SNAKE_CASE)) __lowerCAmelCase : Dict = 0 else: i += 1 __lowerCAmelCase : Tuple = sum(self.preflow[self.source_index]) def _SCREAMING_SNAKE_CASE ( self: Dict , _SCREAMING_SNAKE_CASE: Union[str, Any]) -> Union[str, Any]: """simple docstring""" while self.excesses[vertex_index] > 0: for neighbour_index in range(self.verticies_count): # if it's neighbour and current vertex is higher if ( self.graph[vertex_index][neighbour_index] - self.preflow[vertex_index][neighbour_index] > 0 and self.heights[vertex_index] > self.heights[neighbour_index] ): self.push(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) self.relabel(_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Tuple) -> Dict: """simple docstring""" __lowerCAmelCase : List[str] = min( self.excesses[from_index] , self.graph[from_index][to_index] - self.preflow[from_index][to_index] , ) self.preflow[from_index][to_index] += preflow_delta self.preflow[to_index][from_index] -= preflow_delta self.excesses[from_index] -= preflow_delta self.excesses[to_index] += preflow_delta def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: List[Any]) -> int: """simple docstring""" __lowerCAmelCase : List[Any] = None for to_index in range(self.verticies_count): if ( self.graph[vertex_index][to_index] - self.preflow[vertex_index][to_index] > 0 ) and (min_height is None or self.heights[to_index] < min_height): __lowerCAmelCase : Optional[int] = self.heights[to_index] if min_height is not None: __lowerCAmelCase : List[str] = min_height + 1 if __name__ == "__main__": __snake_case : Tuple = [0] __snake_case : Tuple = [3] # graph = [ # [0, 0, 4, 6, 0, 0], # [0, 0, 5, 2, 0, 0], # [0, 0, 0, 0, 4, 4], # [0, 0, 0, 0, 6, 6], # [0, 0, 0, 0, 0, 0], # [0, 0, 0, 0, 0, 0], # ] __snake_case : int = [[0, 7, 0, 0], [0, 0, 6, 0], [0, 0, 0, 8], [9, 0, 0, 0]] # prepare our network __snake_case : List[Any] = FlowNetwork(graph, entrances, exits) # set algorithm flow_network.set_maximum_flow_algorithm(PushRelabelExecutor) # and calculate __snake_case : Any = flow_network.find_maximum_flow() print(F"""maximum flow is {maximum_flow}""")
269
"""simple docstring""" from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def _lowercase ( __snake_case = "laptop" ) -> DataFrame: __lowerCAmelCase : str = F"""https://www.amazon.in/laptop/s?k={product}""" __lowerCAmelCase : Union[str, Any] = { "User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36\n (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36", "Accept-Language": "en-US, en;q=0.5", } __lowerCAmelCase : List[str] = BeautifulSoup(requests.get(__snake_case ,headers=__snake_case ).text ) # Initialize a Pandas dataframe with the column titles __lowerCAmelCase : Dict = DataFrame( columns=[ "Product Title", "Product Link", "Current Price of the product", "Product Rating", "MRP of the product", "Discount", ] ) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( "div" ,attrs={"class": "s-result-item", "data-component-type": "s-search-result"} ,) ,soup.find_all("div" ,attrs={"class": "a-row a-size-base a-color-base"} ) ,): try: __lowerCAmelCase : Any = item.ha.text __lowerCAmelCase : Union[str, Any] = "https://www.amazon.in/" + item.ha.a["href"] __lowerCAmelCase : Any = item.find("span" ,attrs={"class": "a-offscreen"} ).text try: __lowerCAmelCase : Union[str, Any] = item.find("span" ,attrs={"class": "a-icon-alt"} ).text except AttributeError: __lowerCAmelCase : Optional[Any] = "Not available" try: __lowerCAmelCase : Union[str, Any] = ( "₹" + item.find( "span" ,attrs={"class": "a-price a-text-price"} ).text.split("₹" )[1] ) except AttributeError: __lowerCAmelCase : Dict = "" try: __lowerCAmelCase : str = float( ( ( float(product_mrp.strip("₹" ).replace("," ,"" ) ) - float(product_price.strip("₹" ).replace("," ,"" ) ) ) / float(product_mrp.strip("₹" ).replace("," ,"" ) ) ) * 100 ) except ValueError: __lowerCAmelCase : List[str] = float("nan" ) except AttributeError: pass __lowerCAmelCase : int = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] __lowerCAmelCase : Union[str, Any] = " " __lowerCAmelCase : Union[str, Any] = " " data_frame.index += 1 return data_frame if __name__ == "__main__": __snake_case : Any = 'headphones' get_amazon_product_data(product).to_csv(F"""Amazon Product Data for {product}.csv""")
269
1
"""simple docstring""" from __future__ import annotations import time import numpy as np __snake_case : Optional[Any] = [8, 5, 9, 7] __snake_case : List[Any] = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] __snake_case : Optional[int] = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class A__ : '''simple docstring''' def __init__( self: Any , _SCREAMING_SNAKE_CASE: list[int] , _SCREAMING_SNAKE_CASE: list[list[int]] , _SCREAMING_SNAKE_CASE: list[list[int]] , ) -> None: """simple docstring""" __lowerCAmelCase : Any = claim_vector __lowerCAmelCase : Tuple = allocated_resources_table __lowerCAmelCase : Tuple = maximum_claim_table def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> list[int]: """simple docstring""" return [ sum(p_item[i] for p_item in self.__allocated_resources_table) for i in range(len(self.__allocated_resources_table[0])) ] def _SCREAMING_SNAKE_CASE ( self: int) -> list[int]: """simple docstring""" return np.array(self.__claim_vector) - np.array( self.__processes_resource_summation()) def _SCREAMING_SNAKE_CASE ( self: int) -> list[list[int]]: """simple docstring""" return [ list(np.array(self.__maximum_claim_table[i]) - np.array(_SCREAMING_SNAKE_CASE)) for i, allocated_resource in enumerate(self.__allocated_resources_table) ] def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> dict[int, list[int]]: """simple docstring""" return {self.__need().index(_SCREAMING_SNAKE_CASE): i for i in self.__need()} def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , **_SCREAMING_SNAKE_CASE: List[Any]) -> None: """simple docstring""" __lowerCAmelCase : Optional[int] = self.__need() __lowerCAmelCase : int = self.__allocated_resources_table __lowerCAmelCase : Dict = self.__available_resources() __lowerCAmelCase : str = self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print("_" * 50 + "\n") while need_list: __lowerCAmelCase : int = False for each_need in need_list: __lowerCAmelCase : Dict = True for index, need in enumerate(_SCREAMING_SNAKE_CASE): if need > available_resources[index]: __lowerCAmelCase : Dict = False break if execution: __lowerCAmelCase : Any = True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: __lowerCAmelCase : Union[str, Any] = original_need_index print(F"""Process {process_number + 1} is executing.""") # remove the process run from stack need_list.remove(_SCREAMING_SNAKE_CASE) # update available/freed resources stack __lowerCAmelCase : Dict = np.array(_SCREAMING_SNAKE_CASE) + np.array( alloc_resources_table[process_number]) print( "Updated available resource stack for processes: " + " ".join([str(_SCREAMING_SNAKE_CASE) for x in available_resources])) break if safe: print("The process is in a safe state.\n") else: print("System in unsafe state. Aborting...\n") break def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> List[Any]: """simple docstring""" print(" " * 9 + "Allocated Resource Table") for item in self.__allocated_resources_table: print( F"""P{self.__allocated_resources_table.index(_SCREAMING_SNAKE_CASE) + 1}""" + " ".join(F"""{it:>8}""" for it in item) + "\n") print(" " * 9 + "System Resource Table") for item in self.__maximum_claim_table: print( F"""P{self.__maximum_claim_table.index(_SCREAMING_SNAKE_CASE) + 1}""" + " ".join(F"""{it:>8}""" for it in item) + "\n") print( "Current Usage by Active Processes: " + " ".join(str(_SCREAMING_SNAKE_CASE) for x in self.__claim_vector)) print( "Initial Available Resources: " + " ".join(str(_SCREAMING_SNAKE_CASE) for x in self.__available_resources())) time.sleep(1) if __name__ == "__main__": import doctest doctest.testmod()
269
"""simple docstring""" import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('TEST_SAGEMAKER' , 'False' ) ) is not True , reason='Skipping test because should only be run when releasing minor transformers version' , ) @pytest.mark.usefixtures('sm_env' ) @parameterized_class( [ { 'framework': 'pytorch', 'script': 'run_glue.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_5_0, 'eval_accuracy': 0.6, 'eval_loss': 0.9}, }, { 'framework': 'tensorflow', 'script': 'run_tf.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_0_0, 'eval_accuracy': 0.3, 'eval_loss': 0.9}, }, ] ) class A__ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: int) -> Tuple: """simple docstring""" if self.framework == "pytorch": subprocess.run( F"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() , encoding="utf-8" , check=_SCREAMING_SNAKE_CASE , ) assert hasattr(self , "env") def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Any=1) -> Dict: """simple docstring""" return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=F"""{self.env.base_job_name}-single""" , instance_count=_SCREAMING_SNAKE_CASE , instance_type=self.instance_type , debugger_hook_config=_SCREAMING_SNAKE_CASE , hyperparameters={**self.env.hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version="py36" , ) def _SCREAMING_SNAKE_CASE ( self: str , _SCREAMING_SNAKE_CASE: List[Any]) -> Optional[Any]: """simple docstring""" TrainingJobAnalytics(_SCREAMING_SNAKE_CASE).export_csv(F"""{self.env.test_path}/{job_name}_metrics.csv""") def _SCREAMING_SNAKE_CASE ( self: str) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Tuple = self.create_estimator() # run training estimator.fit() # result dataframe __lowerCAmelCase : Tuple = TrainingJobAnalytics(estimator.latest_training_job.name).dataframe() # extract kpis __lowerCAmelCase : Optional[Any] = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"]) __lowerCAmelCase : Union[str, Any] = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"]) # get train time from SageMaker job, this includes starting, preprocessing, stopping __lowerCAmelCase : Tuple = ( Session().describe_training_job(estimator.latest_training_job.name).get("TrainingTimeInSeconds" , 99_9999) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy) assert all(t <= self.results["eval_loss"] for t in eval_loss) # dump tests result into json file to share in PR with open(F"""{estimator.latest_training_job.name}.json""" , "w") as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , _SCREAMING_SNAKE_CASE)
269
1
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DetrConfig, DetrForObjectDetection, DetrForSegmentation, DetrImageProcessor, ResNetConfig from transformers.utils import logging logging.set_verbosity_info() __snake_case : Union[str, Any] = logging.get_logger(__name__) def _lowercase ( __snake_case ) -> Tuple: # initialize config if "resnet-50" in model_name: __lowerCAmelCase : List[str] = ResNetConfig.from_pretrained("microsoft/resnet-50" ) elif "resnet-101" in model_name: __lowerCAmelCase : List[str] = ResNetConfig.from_pretrained("microsoft/resnet-101" ) else: raise ValueError("Model name should include either resnet50 or resnet101" ) __lowerCAmelCase : str = DetrConfig(use_timm_backbone=__snake_case ,backbone_config=__snake_case ) # set label attributes __lowerCAmelCase : Optional[Any] = "panoptic" in model_name if is_panoptic: __lowerCAmelCase : Dict = 250 else: __lowerCAmelCase : str = 91 __lowerCAmelCase : List[str] = "huggingface/label-files" __lowerCAmelCase : Dict = "coco-detection-id2label.json" __lowerCAmelCase : Tuple = json.load(open(hf_hub_download(__snake_case ,__snake_case ,repo_type="dataset" ) ,"r" ) ) __lowerCAmelCase : List[str] = {int(__snake_case ): v for k, v in idalabel.items()} __lowerCAmelCase : Dict = idalabel __lowerCAmelCase : List[Any] = {v: k for k, v in idalabel.items()} return config, is_panoptic def _lowercase ( __snake_case ) -> Optional[Any]: # here we list all keys to be renamed (original name on the left, our name on the right) __lowerCAmelCase : Dict = [] # stem # fmt: off rename_keys.append(("backbone.0.body.conv1.weight", "backbone.conv_encoder.model.embedder.embedder.convolution.weight") ) rename_keys.append(("backbone.0.body.bn1.weight", "backbone.conv_encoder.model.embedder.embedder.normalization.weight") ) rename_keys.append(("backbone.0.body.bn1.bias", "backbone.conv_encoder.model.embedder.embedder.normalization.bias") ) rename_keys.append(("backbone.0.body.bn1.running_mean", "backbone.conv_encoder.model.embedder.embedder.normalization.running_mean") ) rename_keys.append(("backbone.0.body.bn1.running_var", "backbone.conv_encoder.model.embedder.embedder.normalization.running_var") ) # stages for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): # shortcut if layer_idx == 0: rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.0.weight""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.convolution.weight""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.weight""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.weight""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.bias""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.bias""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_mean""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_mean""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_var""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_var""", ) ) # 3 convs for i in range(3 ): rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.conv{i+1}.weight""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.convolution.weight""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.weight""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.weight""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.bias""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.bias""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_mean""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_mean""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_var""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_var""", ) ) # fmt: on for i in range(config.encoder_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( ( F"""transformer.encoder.layers.{i}.self_attn.out_proj.weight""", F"""encoder.layers.{i}.self_attn.out_proj.weight""", ) ) rename_keys.append( (F"""transformer.encoder.layers.{i}.self_attn.out_proj.bias""", F"""encoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear1.weight""", F"""encoder.layers.{i}.fc1.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear1.bias""", F"""encoder.layers.{i}.fc1.bias""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear2.weight""", F"""encoder.layers.{i}.fc2.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear2.bias""", F"""encoder.layers.{i}.fc2.bias""") ) rename_keys.append( (F"""transformer.encoder.layers.{i}.norm1.weight""", F"""encoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append( (F"""transformer.encoder.layers.{i}.norm1.bias""", F"""encoder.layers.{i}.self_attn_layer_norm.bias""") ) rename_keys.append( (F"""transformer.encoder.layers.{i}.norm2.weight""", F"""encoder.layers.{i}.final_layer_norm.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.norm2.bias""", F"""encoder.layers.{i}.final_layer_norm.bias""") ) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( ( F"""transformer.decoder.layers.{i}.self_attn.out_proj.weight""", F"""decoder.layers.{i}.self_attn.out_proj.weight""", ) ) rename_keys.append( (F"""transformer.decoder.layers.{i}.self_attn.out_proj.bias""", F"""decoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append( ( F"""transformer.decoder.layers.{i}.multihead_attn.out_proj.weight""", F"""decoder.layers.{i}.encoder_attn.out_proj.weight""", ) ) rename_keys.append( ( F"""transformer.decoder.layers.{i}.multihead_attn.out_proj.bias""", F"""decoder.layers.{i}.encoder_attn.out_proj.bias""", ) ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear1.weight""", F"""decoder.layers.{i}.fc1.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear1.bias""", F"""decoder.layers.{i}.fc1.bias""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear2.weight""", F"""decoder.layers.{i}.fc2.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear2.bias""", F"""decoder.layers.{i}.fc2.bias""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm1.weight""", F"""decoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm1.bias""", F"""decoder.layers.{i}.self_attn_layer_norm.bias""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm2.weight""", F"""decoder.layers.{i}.encoder_attn_layer_norm.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm2.bias""", F"""decoder.layers.{i}.encoder_attn_layer_norm.bias""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm3.weight""", F"""decoder.layers.{i}.final_layer_norm.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm3.bias""", F"""decoder.layers.{i}.final_layer_norm.bias""") ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("input_proj.weight", "input_projection.weight"), ("input_proj.bias", "input_projection.bias"), ("query_embed.weight", "query_position_embeddings.weight"), ("transformer.decoder.norm.weight", "decoder.layernorm.weight"), ("transformer.decoder.norm.bias", "decoder.layernorm.bias"), ("class_embed.weight", "class_labels_classifier.weight"), ("class_embed.bias", "class_labels_classifier.bias"), ("bbox_embed.layers.0.weight", "bbox_predictor.layers.0.weight"), ("bbox_embed.layers.0.bias", "bbox_predictor.layers.0.bias"), ("bbox_embed.layers.1.weight", "bbox_predictor.layers.1.weight"), ("bbox_embed.layers.1.bias", "bbox_predictor.layers.1.bias"), ("bbox_embed.layers.2.weight", "bbox_predictor.layers.2.weight"), ("bbox_embed.layers.2.bias", "bbox_predictor.layers.2.bias"), ] ) return rename_keys def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> Any: __lowerCAmelCase : Union[str, Any] = state_dict.pop(__snake_case ) __lowerCAmelCase : Optional[Any] = val def _lowercase ( __snake_case ,__snake_case=False ) -> Dict: __lowerCAmelCase : Optional[int] = "" if is_panoptic: __lowerCAmelCase : Optional[Any] = "detr." # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) __lowerCAmelCase : Optional[Any] = state_dict.pop(F"""{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight""" ) __lowerCAmelCase : Dict = state_dict.pop(F"""{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict __lowerCAmelCase : int = in_proj_weight[:256, :] __lowerCAmelCase : List[Any] = in_proj_bias[:256] __lowerCAmelCase : Tuple = in_proj_weight[256:512, :] __lowerCAmelCase : Any = in_proj_bias[256:512] __lowerCAmelCase : Any = in_proj_weight[-256:, :] __lowerCAmelCase : Optional[int] = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention __lowerCAmelCase : int = state_dict.pop(F"""{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight""" ) __lowerCAmelCase : Tuple = state_dict.pop(F"""{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict __lowerCAmelCase : str = in_proj_weight[:256, :] __lowerCAmelCase : Any = in_proj_bias[:256] __lowerCAmelCase : str = in_proj_weight[256:512, :] __lowerCAmelCase : int = in_proj_bias[256:512] __lowerCAmelCase : Tuple = in_proj_weight[-256:, :] __lowerCAmelCase : Optional[Any] = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention __lowerCAmelCase : int = state_dict.pop( F"""{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight""" ) __lowerCAmelCase : Any = state_dict.pop(F"""{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) of cross-attention to the state dict __lowerCAmelCase : List[str] = in_proj_weight_cross_attn[:256, :] __lowerCAmelCase : Dict = in_proj_bias_cross_attn[:256] __lowerCAmelCase : Optional[int] = in_proj_weight_cross_attn[256:512, :] __lowerCAmelCase : Optional[int] = in_proj_bias_cross_attn[256:512] __lowerCAmelCase : Any = in_proj_weight_cross_attn[-256:, :] __lowerCAmelCase : Optional[Any] = in_proj_bias_cross_attn[-256:] def _lowercase ( ) -> List[Any]: __lowerCAmelCase : Dict = "http://images.cocodataset.org/val2017/000000039769.jpg" __lowerCAmelCase : Union[str, Any] = Image.open(requests.get(__snake_case ,stream=__snake_case ).raw ) return im @torch.no_grad() def _lowercase ( __snake_case ,__snake_case=None ,__snake_case=False ) -> List[str]: __lowerCAmelCase , __lowerCAmelCase : Optional[Any] = get_detr_config(__snake_case ) # load original model from torch hub __lowerCAmelCase : List[Any] = { "detr-resnet-50": "detr_resnet50", "detr-resnet-101": "detr_resnet101", } logger.info(F"""Converting model {model_name}...""" ) __lowerCAmelCase : Tuple = torch.hub.load("facebookresearch/detr" ,model_name_to_original_name[model_name] ,pretrained=__snake_case ).eval() __lowerCAmelCase : List[str] = detr.state_dict() # rename keys for src, dest in create_rename_keys(__snake_case ): if is_panoptic: __lowerCAmelCase : Optional[Any] = "detr." + src rename_key(__snake_case ,__snake_case ,__snake_case ) # query, key and value matrices need special treatment read_in_q_k_v(__snake_case ,is_panoptic=__snake_case ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them __lowerCAmelCase : Tuple = "detr.model." if is_panoptic else "model." for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("detr" ) and not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ) ): __lowerCAmelCase : Tuple = state_dict.pop(__snake_case ) __lowerCAmelCase : Optional[int] = val elif "class_labels_classifier" in key or "bbox_predictor" in key: __lowerCAmelCase : List[str] = state_dict.pop(__snake_case ) __lowerCAmelCase : Union[str, Any] = val elif key.startswith("bbox_attention" ) or key.startswith("mask_head" ): continue else: __lowerCAmelCase : Dict = state_dict.pop(__snake_case ) __lowerCAmelCase : Tuple = val else: if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): __lowerCAmelCase : str = state_dict.pop(__snake_case ) __lowerCAmelCase : int = val # finally, create HuggingFace model and load state dict __lowerCAmelCase : Dict = DetrForSegmentation(__snake_case ) if is_panoptic else DetrForObjectDetection(__snake_case ) model.load_state_dict(__snake_case ) model.eval() # verify our conversion on an image __lowerCAmelCase : Optional[Any] = "coco_panoptic" if is_panoptic else "coco_detection" __lowerCAmelCase : Dict = DetrImageProcessor(format=__snake_case ) __lowerCAmelCase : List[str] = processor(images=prepare_img() ,return_tensors="pt" ) __lowerCAmelCase : int = encoding["pixel_values"] __lowerCAmelCase : Union[str, Any] = detr(__snake_case ) __lowerCAmelCase : Optional[Any] = model(__snake_case ) assert torch.allclose(outputs.logits ,original_outputs["pred_logits"] ,atol=1e-3 ) assert torch.allclose(outputs.pred_boxes ,original_outputs["pred_boxes"] ,atol=1e-3 ) if is_panoptic: assert torch.allclose(outputs.pred_masks ,original_outputs["pred_masks"] ,atol=1e-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(F"""Saving PyTorch model and image processor to {pytorch_dump_folder_path}...""" ) Path(__snake_case ).mkdir(exist_ok=__snake_case ) model.save_pretrained(__snake_case ) processor.save_pretrained(__snake_case ) if push_to_hub: # Upload model and image processor to the hub logger.info("Uploading PyTorch model and image processor to the hub..." ) model.push_to_hub(F"""nielsr/{model_name}""" ) processor.push_to_hub(F"""nielsr/{model_name}""" ) if __name__ == "__main__": __snake_case : Optional[int] = argparse.ArgumentParser() parser.add_argument( '--model_name', default='detr-resnet-50', type=str, choices=['detr-resnet-50', 'detr-resnet-101'], help='Name of the DETR model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) parser.add_argument('--push_to_hub', action='store_true', help='Whether to push the model to the hub or not.') __snake_case : List[str] = parser.parse_args() convert_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
269
"""simple docstring""" from typing import TYPE_CHECKING from ..utils import _LazyModule __snake_case : Optional[int] = { 'config': [ 'EXTERNAL_DATA_FORMAT_SIZE_LIMIT', 'OnnxConfig', 'OnnxConfigWithPast', 'OnnxSeq2SeqConfigWithPast', 'PatchingSpec', ], 'convert': ['export', 'validate_model_outputs'], 'features': ['FeaturesManager'], 'utils': ['ParameterFormat', 'compute_serialized_parameters_size'], } if TYPE_CHECKING: from .config import ( EXTERNAL_DATA_FORMAT_SIZE_LIMIT, OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast, PatchingSpec, ) from .convert import export, validate_model_outputs from .features import FeaturesManager from .utils import ParameterFormat, compute_serialized_parameters_size else: import sys __snake_case : Any = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
269
1
"""simple docstring""" def _lowercase ( __snake_case ) -> list[int]: __lowerCAmelCase : Optional[Any] = [0 for i in range(len(__snake_case ) )] # initialize interval's left pointer and right pointer __lowerCAmelCase , __lowerCAmelCase : Dict = 0, 0 for i in range(1 ,len(__snake_case ) ): # case when current index is inside the interval if i <= right_pointer: __lowerCAmelCase : Optional[Any] = min(right_pointer - i + 1 ,z_result[i - left_pointer] ) __lowerCAmelCase : int = min_edge while go_next(__snake_case ,__snake_case ,__snake_case ): z_result[i] += 1 # if new index's result gives us more right interval, # we've to update left_pointer and right_pointer if i + z_result[i] - 1 > right_pointer: __lowerCAmelCase , __lowerCAmelCase : Optional[int] = i, i + z_result[i] - 1 return z_result def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> bool: return i + z_result[i] < len(__snake_case ) and s[z_result[i]] == s[i + z_result[i]] def _lowercase ( __snake_case ,__snake_case ) -> int: __lowerCAmelCase : Dict = 0 # concatenate 'pattern' and 'input_str' and call z_function # with concatenated string __lowerCAmelCase : List[str] = z_function(pattern + input_str ) for val in z_result: # if value is greater then length of the pattern string # that means this index is starting position of substring # which is equal to pattern string if val >= len(__snake_case ): answer += 1 return answer if __name__ == "__main__": import doctest doctest.testmod()
269
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices __snake_case : Optional[int] = logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'maskformer-swin' SCREAMING_SNAKE_CASE = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self: Optional[int] , _SCREAMING_SNAKE_CASE: int=224 , _SCREAMING_SNAKE_CASE: Tuple=4 , _SCREAMING_SNAKE_CASE: int=3 , _SCREAMING_SNAKE_CASE: List[Any]=96 , _SCREAMING_SNAKE_CASE: Union[str, Any]=[2, 2, 6, 2] , _SCREAMING_SNAKE_CASE: Any=[3, 6, 12, 24] , _SCREAMING_SNAKE_CASE: List[str]=7 , _SCREAMING_SNAKE_CASE: List[str]=4.0 , _SCREAMING_SNAKE_CASE: Optional[int]=True , _SCREAMING_SNAKE_CASE: Tuple=0.0 , _SCREAMING_SNAKE_CASE: Any=0.0 , _SCREAMING_SNAKE_CASE: Any=0.1 , _SCREAMING_SNAKE_CASE: str="gelu" , _SCREAMING_SNAKE_CASE: Tuple=False , _SCREAMING_SNAKE_CASE: Union[str, Any]=0.02 , _SCREAMING_SNAKE_CASE: str=1e-5 , _SCREAMING_SNAKE_CASE: Optional[int]=None , _SCREAMING_SNAKE_CASE: str=None , **_SCREAMING_SNAKE_CASE: Union[str, Any] , ) -> List[str]: """simple docstring""" super().__init__(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = image_size __lowerCAmelCase : Any = patch_size __lowerCAmelCase : Tuple = num_channels __lowerCAmelCase : Any = embed_dim __lowerCAmelCase : Any = depths __lowerCAmelCase : Dict = len(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = num_heads __lowerCAmelCase : Tuple = window_size __lowerCAmelCase : Dict = mlp_ratio __lowerCAmelCase : Any = qkv_bias __lowerCAmelCase : Union[str, Any] = hidden_dropout_prob __lowerCAmelCase : int = attention_probs_dropout_prob __lowerCAmelCase : Tuple = drop_path_rate __lowerCAmelCase : int = hidden_act __lowerCAmelCase : Optional[int] = use_absolute_embeddings __lowerCAmelCase : List[str] = layer_norm_eps __lowerCAmelCase : Any = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model __lowerCAmelCase : Optional[Any] = int(embed_dim * 2 ** (len(_SCREAMING_SNAKE_CASE) - 1)) __lowerCAmelCase : Any = ["stem"] + [F"""stage{idx}""" for idx in range(1 , len(_SCREAMING_SNAKE_CASE) + 1)] __lowerCAmelCase , __lowerCAmelCase : List[str] = get_aligned_output_features_output_indices( out_features=_SCREAMING_SNAKE_CASE , out_indices=_SCREAMING_SNAKE_CASE , stage_names=self.stage_names)
269
1
"""simple docstring""" import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow __snake_case : Tuple = logging.getLogger() @unittest.skip('Temporarily disable the doc tests.' ) @require_torch @require_tf @slow class A__ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Path , _SCREAMING_SNAKE_CASE: Union[str, None] = None , _SCREAMING_SNAKE_CASE: Union[List[str], None] = None , _SCREAMING_SNAKE_CASE: Union[str, List[str], None] = None , _SCREAMING_SNAKE_CASE: bool = True , ) -> int: """simple docstring""" __lowerCAmelCase : str = [file for file in os.listdir(_SCREAMING_SNAKE_CASE) if os.path.isfile(os.path.join(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE))] if identifier is not None: __lowerCAmelCase : Tuple = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE): for n_ in n_identifier: __lowerCAmelCase : Dict = [file for file in files if n_ not in file] else: __lowerCAmelCase : List[str] = [file for file in files if n_identifier not in file] __lowerCAmelCase : List[Any] = ignore_files or [] ignore_files.append("__init__.py") __lowerCAmelCase : List[str] = [file for file in files if file not in ignore_files] for file in files: # Open all files print("Testing" , _SCREAMING_SNAKE_CASE) if only_modules: __lowerCAmelCase : Any = file.split(".")[0] try: __lowerCAmelCase : Any = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = doctest.DocTestSuite(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = unittest.TextTestRunner().run(_SCREAMING_SNAKE_CASE) self.assertIs(len(result.failures) , 0) except AttributeError: logger.info(F"""{module_identifier} is not a module.""") else: __lowerCAmelCase : List[Any] = doctest.testfile(str(".." / directory / file) , optionflags=doctest.ELLIPSIS) self.assertIs(result.failed , 0) def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> str: """simple docstring""" __lowerCAmelCase : Union[str, Any] = Path("src/transformers") __lowerCAmelCase : Optional[Any] = "modeling" __lowerCAmelCase : List[str] = [ "modeling_ctrl.py", "modeling_tf_ctrl.py", ] self.analyze_directory(_SCREAMING_SNAKE_CASE , identifier=_SCREAMING_SNAKE_CASE , ignore_files=_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> Any: """simple docstring""" __lowerCAmelCase : Any = Path("src/transformers") __lowerCAmelCase : Optional[Any] = "tokenization" self.analyze_directory(_SCREAMING_SNAKE_CASE , identifier=_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Dict) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Union[str, Any] = Path("src/transformers") __lowerCAmelCase : Union[str, Any] = "configuration" self.analyze_directory(_SCREAMING_SNAKE_CASE , identifier=_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[str]) -> List[str]: """simple docstring""" __lowerCAmelCase : Union[str, Any] = Path("src/transformers") __lowerCAmelCase : Any = ["configuration", "modeling", "tokenization"] self.analyze_directory(_SCREAMING_SNAKE_CASE , n_identifier=_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> List[Any]: """simple docstring""" __lowerCAmelCase : Union[str, Any] = Path("docs/source") __lowerCAmelCase : Dict = ["favicon.ico"] self.analyze_directory(_SCREAMING_SNAKE_CASE , ignore_files=_SCREAMING_SNAKE_CASE , only_modules=_SCREAMING_SNAKE_CASE)
269
"""simple docstring""" import gc import threading import time import psutil import torch class A__ : '''simple docstring''' def __init__( self: str) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Optional[Any] = psutil.Process() __lowerCAmelCase : str = False def _SCREAMING_SNAKE_CASE ( self: int) -> Optional[int]: """simple docstring""" __lowerCAmelCase : Optional[Any] = -1 while True: __lowerCAmelCase : str = max(self.process.memory_info().rss , self.cpu_memory_peak) # can't sleep or will not catch the peak right (this comment is here on purpose) if not self.peak_monitoring: break def _SCREAMING_SNAKE_CASE ( self: Tuple) -> int: """simple docstring""" __lowerCAmelCase : List[str] = True __lowerCAmelCase : str = threading.Thread(target=self.peak_monitor) __lowerCAmelCase : Tuple = True self.thread.start() def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> List[str]: """simple docstring""" __lowerCAmelCase : Tuple = False self.thread.join() return self.cpu_memory_peak __snake_case : Tuple = PeakCPUMemory() def _lowercase ( ) -> str: # Time __lowerCAmelCase : str = {"time": time.time()} gc.collect() torch.cuda.empty_cache() # CPU mem __lowerCAmelCase : Optional[Any] = psutil.Process().memory_info().rss cpu_peak_tracker.start() # GPU mem for i in range(torch.cuda.device_count() ): __lowerCAmelCase : Union[str, Any] = torch.cuda.memory_allocated(__snake_case ) torch.cuda.reset_peak_memory_stats() return measures def _lowercase ( __snake_case ) -> Optional[Any]: # Time __lowerCAmelCase : str = {"time": time.time() - start_measures["time"]} gc.collect() torch.cuda.empty_cache() # CPU mem __lowerCAmelCase : str = (psutil.Process().memory_info().rss - start_measures["cpu"]) / 2**20 __lowerCAmelCase : List[str] = (cpu_peak_tracker.stop() - start_measures["cpu"]) / 2**20 # GPU mem for i in range(torch.cuda.device_count() ): __lowerCAmelCase : Union[str, Any] = (torch.cuda.memory_allocated(__snake_case ) - start_measures[str(__snake_case )]) / 2**20 __lowerCAmelCase : Any = (torch.cuda.max_memory_allocated(__snake_case ) - start_measures[str(__snake_case )]) / 2**20 return measures def _lowercase ( __snake_case ,__snake_case ) -> Dict: print(F"""{description}:""" ) print(F"""- Time: {measures['time']:.2f}s""" ) for i in range(torch.cuda.device_count() ): print(F"""- GPU {i} allocated: {measures[str(__snake_case )]:.2f}MiB""" ) __lowerCAmelCase : Optional[Any] = measures[F"""{i}-peak"""] print(F"""- GPU {i} peak: {peak:.2f}MiB""" ) print(F"""- CPU RAM allocated: {measures['cpu']:.2f}MiB""" ) print(F"""- CPU RAM peak: {measures['cpu-peak']:.2f}MiB""" )
269
1
"""simple docstring""" from scipy.stats import pearsonr import datasets __snake_case : Dict = '\nPearson correlation coefficient and p-value for testing non-correlation.\nThe Pearson correlation coefficient measures the linear relationship between two datasets. The calculation of the p-value relies on the assumption that each dataset is normally distributed. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Correlations of -1 or +1 imply an exact linear relationship. Positive correlations imply that as x increases, so does y. Negative correlations imply that as x increases, y decreases.\nThe p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets.\n' __snake_case : int = '\nArgs:\n predictions (`list` of `int`): Predicted class labels, as returned by a model.\n references (`list` of `int`): Ground truth labels.\n return_pvalue (`boolean`): If `True`, returns the p-value, along with the correlation coefficient. If `False`, returns only the correlation coefficient. Defaults to `False`.\n\nReturns:\n pearsonr (`float`): Pearson correlation coefficient. Minimum possible value is -1. Maximum possible value is 1. Values of 1 and -1 indicate exact linear positive and negative relationships, respectively. A value of 0 implies no correlation.\n p-value (`float`): P-value, which roughly indicates the probability of an The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. Minimum possible value is 0. Maximum possible value is 1. Higher values indicate higher probabilities.\n\nExamples:\n\n Example 1-A simple example using only predictions and references.\n >>> pearsonr_metric = datasets.load_metric("pearsonr")\n >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5])\n >>> print(round(results[\'pearsonr\'], 2))\n -0.74\n\n Example 2-The same as Example 1, but that also returns the `p-value`.\n >>> pearsonr_metric = datasets.load_metric("pearsonr")\n >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5], return_pvalue=True)\n >>> print(sorted(list(results.keys())))\n [\'p-value\', \'pearsonr\']\n >>> print(round(results[\'pearsonr\'], 2))\n -0.74\n >>> print(round(results[\'p-value\'], 2))\n 0.15\n' __snake_case : List[Any] = '\n@article{2020SciPy-NMeth,\nauthor = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and\n Haberland, Matt and Reddy, Tyler and Cournapeau, David and\n Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and\n Bright, Jonathan and {van der Walt}, St{\'e}fan J. and\n Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and\n Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and\n Kern, Robert and Larson, Eric and Carey, C J and\n Polat, Ilhan and Feng, Yu and Moore, Eric W. and\n {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and\n Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and\n Harris, Charles R. and Archibald, Anne M. and\n Ribeiro, Antonio H. and Pedregosa, Fabian and\n {van Mulbregt}, Paul and {SciPy 1.0 Contributors}},\ntitle = {{{SciPy} 1.0: Fundamental Algorithms for Scientific\n Computing in Python}},\njournal = {Nature Methods},\nyear = {2020},\nvolume = {17},\npages = {261--272},\nadsurl = {https://rdcu.be/b08Wh},\ndoi = {10.1038/s41592-019-0686-2},\n}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A__ ( datasets.Metric ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: str) -> Tuple: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("float"), "references": datasets.Value("float"), }) , reference_urls=["https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.pearsonr.html"] , ) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Any=False) -> Union[str, Any]: """simple docstring""" if return_pvalue: __lowerCAmelCase : int = pearsonr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) return {"pearsonr": results[0], "p-value": results[1]} else: return {"pearsonr": float(pearsonr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE)[0])}
269
"""simple docstring""" from ....configuration_utils import PretrainedConfig from ....utils import logging __snake_case : List[Any] = logging.get_logger(__name__) __snake_case : Any = { 'speechbrain/m-ctc-t-large': 'https://huggingface.co/speechbrain/m-ctc-t-large/resolve/main/config.json', # See all M-CTC-T models at https://huggingface.co/models?filter=mctct } class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'mctct' def __init__( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: str=8065 , _SCREAMING_SNAKE_CASE: str=1536 , _SCREAMING_SNAKE_CASE: str=36 , _SCREAMING_SNAKE_CASE: Optional[Any]=6144 , _SCREAMING_SNAKE_CASE: Optional[Any]=4 , _SCREAMING_SNAKE_CASE: Union[str, Any]=384 , _SCREAMING_SNAKE_CASE: Optional[Any]=920 , _SCREAMING_SNAKE_CASE: Union[str, Any]=1e-5 , _SCREAMING_SNAKE_CASE: List[Any]=0.3 , _SCREAMING_SNAKE_CASE: Optional[Any]="relu" , _SCREAMING_SNAKE_CASE: Optional[int]=0.02 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.3 , _SCREAMING_SNAKE_CASE: Dict=0.3 , _SCREAMING_SNAKE_CASE: List[Any]=1 , _SCREAMING_SNAKE_CASE: Optional[Any]=0 , _SCREAMING_SNAKE_CASE: List[str]=2 , _SCREAMING_SNAKE_CASE: Union[str, Any]=1 , _SCREAMING_SNAKE_CASE: Tuple=0.3 , _SCREAMING_SNAKE_CASE: Dict=1 , _SCREAMING_SNAKE_CASE: int=(7,) , _SCREAMING_SNAKE_CASE: str=(3,) , _SCREAMING_SNAKE_CASE: Union[str, Any]=80 , _SCREAMING_SNAKE_CASE: Tuple=1 , _SCREAMING_SNAKE_CASE: Dict=None , _SCREAMING_SNAKE_CASE: Tuple="sum" , _SCREAMING_SNAKE_CASE: List[str]=False , **_SCREAMING_SNAKE_CASE: Tuple , ) -> Tuple: """simple docstring""" super().__init__(**_SCREAMING_SNAKE_CASE , pad_token_id=_SCREAMING_SNAKE_CASE , bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = vocab_size __lowerCAmelCase : str = hidden_size __lowerCAmelCase : str = num_hidden_layers __lowerCAmelCase : str = intermediate_size __lowerCAmelCase : List[Any] = num_attention_heads __lowerCAmelCase : Dict = attention_head_dim __lowerCAmelCase : Optional[int] = max_position_embeddings __lowerCAmelCase : str = layer_norm_eps __lowerCAmelCase : Tuple = layerdrop __lowerCAmelCase : str = hidden_act __lowerCAmelCase : List[Any] = initializer_range __lowerCAmelCase : int = hidden_dropout_prob __lowerCAmelCase : Optional[Any] = attention_probs_dropout_prob __lowerCAmelCase : str = pad_token_id __lowerCAmelCase : Optional[int] = bos_token_id __lowerCAmelCase : Union[str, Any] = eos_token_id __lowerCAmelCase : Any = conv_glu_dim __lowerCAmelCase : Optional[int] = conv_dropout __lowerCAmelCase : Union[str, Any] = num_conv_layers __lowerCAmelCase : Optional[int] = input_feat_per_channel __lowerCAmelCase : Union[str, Any] = input_channels __lowerCAmelCase : Optional[Any] = conv_channels __lowerCAmelCase : Dict = ctc_loss_reduction __lowerCAmelCase : int = ctc_zero_infinity # prevents config testing fail with exporting to json __lowerCAmelCase : List[str] = list(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = list(_SCREAMING_SNAKE_CASE) if len(self.conv_kernel) != self.num_conv_layers: raise ValueError( "Configuration for convolutional module is incorrect. " "It is required that `len(config.conv_kernel)` == `config.num_conv_layers` " F"""but is `len(config.conv_kernel) = {len(self.conv_kernel)}`, """ F"""`config.num_conv_layers = {self.num_conv_layers}`.""")
269
1
"""simple docstring""" from sklearn.metrics import fa_score import datasets __snake_case : Tuple = '\nThe F1 score is the harmonic mean of the precision and recall. It can be computed with the equation:\nF1 = 2 * (precision * recall) / (precision + recall)\n' __snake_case : int = '\nArgs:\n predictions (`list` of `int`): Predicted labels.\n references (`list` of `int`): Ground truth labels.\n labels (`list` of `int`): The set of labels to include when `average` is not set to `\'binary\'`, and the order of the labels if `average` is `None`. Labels present in the data can be excluded, for example to calculate a multiclass average ignoring a majority negative class. Labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in `predictions` and `references` are used in sorted order. Defaults to None.\n pos_label (`int`): The class to be considered the positive class, in the case where `average` is set to `binary`. Defaults to 1.\n average (`string`): This parameter is required for multiclass/multilabel targets. If set to `None`, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `\'binary\'`.\n\n - \'binary\': Only report results for the class specified by `pos_label`. This is applicable only if the classes found in `predictions` and `references` are binary.\n - \'micro\': Calculate metrics globally by counting the total true positives, false negatives and false positives.\n - \'macro\': Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account.\n - \'weighted\': Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `\'macro\'` to account for label imbalance. This option can result in an F-score that is not between precision and recall.\n - \'samples\': Calculate metrics for each instance, and find their average (only meaningful for multilabel classification).\n sample_weight (`list` of `float`): Sample weights Defaults to None.\n\nReturns:\n f1 (`float` or `array` of `float`): F1 score or list of f1 scores, depending on the value passed to `average`. Minimum possible value is 0. Maximum possible value is 1. Higher f1 scores are better.\n\nExamples:\n\n Example 1-A simple binary example\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0])\n >>> print(results)\n {\'f1\': 0.5}\n\n Example 2-The same simple binary example as in Example 1, but with `pos_label` set to `0`.\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], pos_label=0)\n >>> print(round(results[\'f1\'], 2))\n 0.67\n\n Example 3-The same simple binary example as in Example 1, but with `sample_weight` included.\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], sample_weight=[0.9, 0.5, 3.9, 1.2, 0.3])\n >>> print(round(results[\'f1\'], 2))\n 0.35\n\n Example 4-A multiclass example, with different values for the `average` input.\n >>> predictions = [0, 2, 1, 0, 0, 1]\n >>> references = [0, 1, 2, 0, 1, 2]\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="macro")\n >>> print(round(results[\'f1\'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="micro")\n >>> print(round(results[\'f1\'], 2))\n 0.33\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="weighted")\n >>> print(round(results[\'f1\'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average=None)\n >>> print(results)\n {\'f1\': array([0.8, 0. , 0. ])}\n' __snake_case : Dict = '\n@article{scikit-learn,\n title={Scikit-learn: Machine Learning in {P}ython},\n author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V.\n and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P.\n and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and\n Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.},\n journal={Journal of Machine Learning Research},\n volume={12},\n pages={2825--2830},\n year={2011}\n}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A__ ( datasets.Metric ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> int: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Sequence(datasets.Value("int32")), "references": datasets.Sequence(datasets.Value("int32")), } if self.config_name == "multilabel" else { "predictions": datasets.Value("int32"), "references": datasets.Value("int32"), }) , reference_urls=["https://scikit-learn.org/stable/modules/generated/sklearn.metrics.f1_score.html"] , ) def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: List[Any]=None , _SCREAMING_SNAKE_CASE: Union[str, Any]=1 , _SCREAMING_SNAKE_CASE: Optional[int]="binary" , _SCREAMING_SNAKE_CASE: List[Any]=None) -> str: """simple docstring""" __lowerCAmelCase : str = fa_score( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE , pos_label=_SCREAMING_SNAKE_CASE , average=_SCREAMING_SNAKE_CASE , sample_weight=_SCREAMING_SNAKE_CASE) return {"f1": float(_SCREAMING_SNAKE_CASE) if score.size == 1 else score}
269
"""simple docstring""" from __future__ import annotations import time import numpy as np __snake_case : Optional[Any] = [8, 5, 9, 7] __snake_case : List[Any] = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] __snake_case : Optional[int] = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class A__ : '''simple docstring''' def __init__( self: Any , _SCREAMING_SNAKE_CASE: list[int] , _SCREAMING_SNAKE_CASE: list[list[int]] , _SCREAMING_SNAKE_CASE: list[list[int]] , ) -> None: """simple docstring""" __lowerCAmelCase : Any = claim_vector __lowerCAmelCase : Tuple = allocated_resources_table __lowerCAmelCase : Tuple = maximum_claim_table def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> list[int]: """simple docstring""" return [ sum(p_item[i] for p_item in self.__allocated_resources_table) for i in range(len(self.__allocated_resources_table[0])) ] def _SCREAMING_SNAKE_CASE ( self: int) -> list[int]: """simple docstring""" return np.array(self.__claim_vector) - np.array( self.__processes_resource_summation()) def _SCREAMING_SNAKE_CASE ( self: int) -> list[list[int]]: """simple docstring""" return [ list(np.array(self.__maximum_claim_table[i]) - np.array(_SCREAMING_SNAKE_CASE)) for i, allocated_resource in enumerate(self.__allocated_resources_table) ] def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> dict[int, list[int]]: """simple docstring""" return {self.__need().index(_SCREAMING_SNAKE_CASE): i for i in self.__need()} def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , **_SCREAMING_SNAKE_CASE: List[Any]) -> None: """simple docstring""" __lowerCAmelCase : Optional[int] = self.__need() __lowerCAmelCase : int = self.__allocated_resources_table __lowerCAmelCase : Dict = self.__available_resources() __lowerCAmelCase : str = self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print("_" * 50 + "\n") while need_list: __lowerCAmelCase : int = False for each_need in need_list: __lowerCAmelCase : Dict = True for index, need in enumerate(_SCREAMING_SNAKE_CASE): if need > available_resources[index]: __lowerCAmelCase : Dict = False break if execution: __lowerCAmelCase : Any = True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: __lowerCAmelCase : Union[str, Any] = original_need_index print(F"""Process {process_number + 1} is executing.""") # remove the process run from stack need_list.remove(_SCREAMING_SNAKE_CASE) # update available/freed resources stack __lowerCAmelCase : Dict = np.array(_SCREAMING_SNAKE_CASE) + np.array( alloc_resources_table[process_number]) print( "Updated available resource stack for processes: " + " ".join([str(_SCREAMING_SNAKE_CASE) for x in available_resources])) break if safe: print("The process is in a safe state.\n") else: print("System in unsafe state. Aborting...\n") break def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> List[Any]: """simple docstring""" print(" " * 9 + "Allocated Resource Table") for item in self.__allocated_resources_table: print( F"""P{self.__allocated_resources_table.index(_SCREAMING_SNAKE_CASE) + 1}""" + " ".join(F"""{it:>8}""" for it in item) + "\n") print(" " * 9 + "System Resource Table") for item in self.__maximum_claim_table: print( F"""P{self.__maximum_claim_table.index(_SCREAMING_SNAKE_CASE) + 1}""" + " ".join(F"""{it:>8}""" for it in item) + "\n") print( "Current Usage by Active Processes: " + " ".join(str(_SCREAMING_SNAKE_CASE) for x in self.__claim_vector)) print( "Initial Available Resources: " + " ".join(str(_SCREAMING_SNAKE_CASE) for x in self.__available_resources())) time.sleep(1) if __name__ == "__main__": import doctest doctest.testmod()
269
1
"""simple docstring""" from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging __snake_case : List[Any] = logging.get_logger(__name__) __snake_case : Dict = { 'Salesforce/codegen-350M-nl': 'https://huggingface.co/Salesforce/codegen-350M-nl/resolve/main/config.json', 'Salesforce/codegen-350M-multi': 'https://huggingface.co/Salesforce/codegen-350M-multi/resolve/main/config.json', 'Salesforce/codegen-350M-mono': 'https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/config.json', 'Salesforce/codegen-2B-nl': 'https://huggingface.co/Salesforce/codegen-2B-nl/resolve/main/config.json', 'Salesforce/codegen-2B-multi': 'https://huggingface.co/Salesforce/codegen-2B-multi/resolve/main/config.json', 'Salesforce/codegen-2B-mono': 'https://huggingface.co/Salesforce/codegen-2B-mono/resolve/main/config.json', 'Salesforce/codegen-6B-nl': 'https://huggingface.co/Salesforce/codegen-6B-nl/resolve/main/config.json', 'Salesforce/codegen-6B-multi': 'https://huggingface.co/Salesforce/codegen-6B-multi/resolve/main/config.json', 'Salesforce/codegen-6B-mono': 'https://huggingface.co/Salesforce/codegen-6B-mono/resolve/main/config.json', 'Salesforce/codegen-16B-nl': 'https://huggingface.co/Salesforce/codegen-16B-nl/resolve/main/config.json', 'Salesforce/codegen-16B-multi': 'https://huggingface.co/Salesforce/codegen-16B-multi/resolve/main/config.json', 'Salesforce/codegen-16B-mono': 'https://huggingface.co/Salesforce/codegen-16B-mono/resolve/main/config.json', } class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'codegen' SCREAMING_SNAKE_CASE = { 'max_position_embeddings': 'n_positions', 'hidden_size': 'n_embd', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self: List[str] , _SCREAMING_SNAKE_CASE: Any=5_0400 , _SCREAMING_SNAKE_CASE: str=2048 , _SCREAMING_SNAKE_CASE: Optional[int]=2048 , _SCREAMING_SNAKE_CASE: List[str]=4096 , _SCREAMING_SNAKE_CASE: int=28 , _SCREAMING_SNAKE_CASE: Optional[int]=16 , _SCREAMING_SNAKE_CASE: int=64 , _SCREAMING_SNAKE_CASE: Optional[Any]=None , _SCREAMING_SNAKE_CASE: Optional[int]="gelu_new" , _SCREAMING_SNAKE_CASE: Any=0.0 , _SCREAMING_SNAKE_CASE: Tuple=0.0 , _SCREAMING_SNAKE_CASE: Optional[int]=0.0 , _SCREAMING_SNAKE_CASE: int=1e-5 , _SCREAMING_SNAKE_CASE: Optional[int]=0.02 , _SCREAMING_SNAKE_CASE: Optional[Any]=True , _SCREAMING_SNAKE_CASE: List[str]=5_0256 , _SCREAMING_SNAKE_CASE: int=5_0256 , _SCREAMING_SNAKE_CASE: Tuple=False , **_SCREAMING_SNAKE_CASE: Optional[Any] , ) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : List[str] = vocab_size __lowerCAmelCase : List[Any] = n_ctx __lowerCAmelCase : Union[str, Any] = n_positions __lowerCAmelCase : int = n_embd __lowerCAmelCase : Any = n_layer __lowerCAmelCase : Dict = n_head __lowerCAmelCase : List[Any] = n_inner __lowerCAmelCase : int = rotary_dim __lowerCAmelCase : List[str] = activation_function __lowerCAmelCase : str = resid_pdrop __lowerCAmelCase : List[Any] = embd_pdrop __lowerCAmelCase : Union[str, Any] = attn_pdrop __lowerCAmelCase : Tuple = layer_norm_epsilon __lowerCAmelCase : Union[str, Any] = initializer_range __lowerCAmelCase : Union[str, Any] = use_cache __lowerCAmelCase : Optional[Any] = bos_token_id __lowerCAmelCase : List[Any] = eos_token_id super().__init__( bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE , tie_word_embeddings=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self: Optional[int] , _SCREAMING_SNAKE_CASE: PretrainedConfig , _SCREAMING_SNAKE_CASE: str = "default" , _SCREAMING_SNAKE_CASE: List[PatchingSpec] = None , _SCREAMING_SNAKE_CASE: bool = False , ) -> Union[str, Any]: """simple docstring""" super().__init__(_SCREAMING_SNAKE_CASE , task=_SCREAMING_SNAKE_CASE , patching_specs=_SCREAMING_SNAKE_CASE , use_past=_SCREAMING_SNAKE_CASE) if not getattr(self._config , "pad_token_id" , _SCREAMING_SNAKE_CASE): # TODO: how to do that better? __lowerCAmelCase : Any = 0 @property def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Mapping[str, Mapping[int, str]]: """simple docstring""" __lowerCAmelCase : Tuple = OrderedDict({"input_ids": {0: "batch", 1: "sequence"}}) if self.use_past: self.fill_with_past_key_values_(_SCREAMING_SNAKE_CASE , direction="inputs") __lowerCAmelCase : Tuple = {0: "batch", 1: "past_sequence + sequence"} else: __lowerCAmelCase : Any = {0: "batch", 1: "sequence"} return common_inputs @property def _SCREAMING_SNAKE_CASE ( self: int) -> int: """simple docstring""" return self._config.n_layer @property def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> int: """simple docstring""" return self._config.n_head def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: PreTrainedTokenizer , _SCREAMING_SNAKE_CASE: int = -1 , _SCREAMING_SNAKE_CASE: int = -1 , _SCREAMING_SNAKE_CASE: bool = False , _SCREAMING_SNAKE_CASE: Optional[TensorType] = None , ) -> Mapping[str, Any]: """simple docstring""" __lowerCAmelCase : Tuple = super(_SCREAMING_SNAKE_CASE , self).generate_dummy_inputs( _SCREAMING_SNAKE_CASE , batch_size=_SCREAMING_SNAKE_CASE , seq_length=_SCREAMING_SNAKE_CASE , is_pair=_SCREAMING_SNAKE_CASE , framework=_SCREAMING_SNAKE_CASE) # We need to order the input in the way they appears in the forward() __lowerCAmelCase : Union[str, Any] = OrderedDict({"input_ids": common_inputs["input_ids"]}) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed.") else: import torch __lowerCAmelCase , __lowerCAmelCase : Tuple = common_inputs["input_ids"].shape # Not using the same length for past_key_values __lowerCAmelCase : List[str] = seqlen + 2 __lowerCAmelCase : str = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) __lowerCAmelCase : Tuple = [ (torch.zeros(_SCREAMING_SNAKE_CASE), torch.zeros(_SCREAMING_SNAKE_CASE)) for _ in range(self.num_layers) ] __lowerCAmelCase : Dict = common_inputs["attention_mask"] if self.use_past: __lowerCAmelCase : List[Any] = ordered_inputs["attention_mask"].dtype __lowerCAmelCase : List[Any] = torch.cat( [ordered_inputs["attention_mask"], torch.ones(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , dtype=_SCREAMING_SNAKE_CASE)] , dim=1) return ordered_inputs @property def _SCREAMING_SNAKE_CASE ( self: List[str]) -> int: """simple docstring""" return 13
269
"""simple docstring""" # # This a `torch.distributed` diagnostics script that checks that all GPUs in the cluster (one or # many nodes) can talk to each other via nccl and allocate gpu memory. # # To run first adjust the number of processes and nodes: # # python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # You may need to add --master_addr $MASTER_ADDR --master_port $MASTER_PORT if using a custom addr:port # # You can also use the rdzv API: --rdzv_endpoint $MASTER_ADDR:$MASTER_PORT --rdzv_backend c10d # # use torch.distributed.launch instead of torch.distributed.run for torch < 1.9 # # If you get a hanging in `barrier` calls you have some network issues, you may try to debug this with: # # NCCL_DEBUG=INFO python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # which should tell you what's going on behind the scenes. # # # This script can be run via `srun` in the SLURM environment as well. Here is a SLURM script that # runs on 2 nodes of 4 gpus per node: # # #SBATCH --job-name=test-nodes # name # #SBATCH --nodes=2 # nodes # #SBATCH --ntasks-per-node=1 # crucial - only 1 task per dist per node! # #SBATCH --cpus-per-task=10 # number of cores per tasks # #SBATCH --gres=gpu:4 # number of gpus # #SBATCH --time 0:05:00 # maximum execution time (HH:MM:SS) # #SBATCH --output=%x-%j.out # output file name # # GPUS_PER_NODE=4 # MASTER_ADDR=$(scontrol show hostnames $SLURM_JOB_NODELIST | head -n 1) # MASTER_PORT=6000 # # srun --jobid $SLURM_JOBID bash -c 'python -m torch.distributed.run \ # --nproc_per_node $GPUS_PER_NODE --nnodes $SLURM_NNODES --node_rank $SLURM_PROCID \ # --master_addr $MASTER_ADDR --master_port $MASTER_PORT \ # torch-distributed-gpu-test.py' # import fcntl import os import socket import torch import torch.distributed as dist def _lowercase ( *__snake_case ) -> Optional[Any]: with open(__snake_case ,"r" ) as fh: fcntl.flock(__snake_case ,fcntl.LOCK_EX ) try: print(*__snake_case ) finally: fcntl.flock(__snake_case ,fcntl.LOCK_UN ) __snake_case : List[Any] = int(os.environ['LOCAL_RANK']) torch.cuda.set_device(local_rank) __snake_case : List[str] = torch.device('cuda', local_rank) __snake_case : Optional[Any] = socket.gethostname() __snake_case : str = F"""[{hostname}-{local_rank}]""" try: # test distributed dist.init_process_group('nccl') dist.all_reduce(torch.ones(1).to(device), op=dist.ReduceOp.SUM) dist.barrier() # test cuda is available and can allocate memory torch.cuda.is_available() torch.ones(1).cuda(local_rank) # global rank __snake_case : Tuple = dist.get_rank() __snake_case : Any = dist.get_world_size() printflock(F"""{gpu} is OK (global rank: {rank}/{world_size})""") dist.barrier() if rank == 0: printflock(F"""pt={torch.__version__}, cuda={torch.version.cuda}, nccl={torch.cuda.nccl.version()}""") except Exception: printflock(F"""{gpu} is broken""") raise
269
1
"""simple docstring""" __snake_case : List[Any] = [ 999, 800, 799, 600, 599, 500, 400, 399, 377, 355, 333, 311, 288, 266, 244, 222, 200, 199, 177, 155, 133, 111, 88, 66, 44, 22, 0, ] __snake_case : Union[str, Any] = [ 999, 976, 952, 928, 905, 882, 858, 857, 810, 762, 715, 714, 572, 429, 428, 286, 285, 238, 190, 143, 142, 118, 95, 71, 47, 24, 0, ] __snake_case : str = [ 999, 988, 977, 966, 955, 944, 933, 922, 911, 900, 899, 879, 859, 840, 820, 800, 799, 766, 733, 700, 699, 650, 600, 599, 500, 499, 400, 399, 350, 300, 299, 266, 233, 200, 199, 179, 159, 140, 120, 100, 99, 88, 77, 66, 55, 44, 33, 22, 11, 0, ] __snake_case : Optional[int] = [ 999, 995, 992, 989, 985, 981, 978, 975, 971, 967, 964, 961, 957, 956, 951, 947, 942, 937, 933, 928, 923, 919, 914, 913, 908, 903, 897, 892, 887, 881, 876, 871, 870, 864, 858, 852, 846, 840, 834, 828, 827, 820, 813, 806, 799, 792, 785, 784, 777, 770, 763, 756, 749, 742, 741, 733, 724, 716, 707, 699, 698, 688, 677, 666, 656, 655, 645, 634, 623, 613, 612, 598, 584, 570, 569, 555, 541, 527, 526, 505, 484, 483, 462, 440, 439, 396, 395, 352, 351, 308, 307, 264, 263, 220, 219, 176, 132, 88, 44, 0, ] __snake_case : List[Any] = [ 999, 997, 995, 992, 990, 988, 986, 984, 981, 979, 977, 975, 972, 970, 968, 966, 964, 961, 959, 957, 956, 954, 951, 949, 946, 944, 941, 939, 936, 934, 931, 929, 926, 924, 921, 919, 916, 914, 913, 910, 907, 905, 902, 899, 896, 893, 891, 888, 885, 882, 879, 877, 874, 871, 870, 867, 864, 861, 858, 855, 852, 849, 846, 843, 840, 837, 834, 831, 828, 827, 824, 821, 817, 814, 811, 808, 804, 801, 798, 795, 791, 788, 785, 784, 780, 777, 774, 770, 766, 763, 760, 756, 752, 749, 746, 742, 741, 737, 733, 730, 726, 722, 718, 714, 710, 707, 703, 699, 698, 694, 690, 685, 681, 677, 673, 669, 664, 660, 656, 655, 650, 646, 641, 636, 632, 627, 622, 618, 613, 612, 607, 602, 596, 591, 586, 580, 575, 570, 569, 563, 557, 551, 545, 539, 533, 527, 526, 519, 512, 505, 498, 491, 484, 483, 474, 466, 457, 449, 440, 439, 428, 418, 407, 396, 395, 381, 366, 352, 351, 330, 308, 307, 286, 264, 263, 242, 220, 219, 176, 175, 132, 131, 88, 44, 0, ] __snake_case : Any = [ 999, 991, 982, 974, 966, 958, 950, 941, 933, 925, 916, 908, 900, 899, 874, 850, 825, 800, 799, 700, 600, 500, 400, 300, 200, 100, 0, ] __snake_case : Optional[Any] = [ 999, 992, 985, 978, 971, 964, 957, 949, 942, 935, 928, 921, 914, 907, 900, 899, 879, 859, 840, 820, 800, 799, 766, 733, 700, 699, 650, 600, 599, 500, 499, 400, 399, 300, 299, 200, 199, 100, 99, 0, ] __snake_case : List[str] = [ 999, 996, 992, 989, 985, 982, 979, 975, 972, 968, 965, 961, 958, 955, 951, 948, 944, 941, 938, 934, 931, 927, 924, 920, 917, 914, 910, 907, 903, 900, 899, 891, 884, 876, 869, 861, 853, 846, 838, 830, 823, 815, 808, 800, 799, 788, 777, 766, 755, 744, 733, 722, 711, 700, 699, 688, 677, 666, 655, 644, 633, 622, 611, 600, 599, 585, 571, 557, 542, 528, 514, 500, 499, 485, 471, 457, 442, 428, 414, 400, 399, 379, 359, 340, 320, 300, 299, 279, 259, 240, 220, 200, 199, 166, 133, 100, 99, 66, 33, 0, ]
269
"""simple docstring""" import numpy as np from scipy.spatial.distance import cdist from sklearn.metrics import fa_score import datasets __snake_case : Optional[int] = '\\n @inproceedings{kakwani2020indicnlpsuite,\n title={{IndicNLPSuite: Monolingual Corpora, Evaluation Benchmarks and Pre-trained Multilingual Language Models for Indian Languages}},\n author={Divyanshu Kakwani and Anoop Kunchukuttan and Satish Golla and Gokul N.C. and Avik Bhattacharyya and Mitesh M. Khapra and Pratyush Kumar},\n year={2020},\n booktitle={Findings of EMNLP},\n}\n' __snake_case : str = '\\n IndicGLUE is a natural language understanding benchmark for Indian languages. It contains a wide\n variety of tasks and covers 11 major Indian languages - as, bn, gu, hi, kn, ml, mr, or, pa, ta, te.\n' __snake_case : str = '\nCompute IndicGLUE evaluation metric associated to each IndicGLUE dataset.\nArgs:\n predictions: list of predictions to score (as int64),\n except for \'cvit-mkb-clsr\' where each prediction is a vector (of float32).\n references: list of ground truth labels corresponding to the predictions (as int64),\n except for \'cvit-mkb-clsr\' where each reference is a vector (of float32).\nReturns: depending on the IndicGLUE subset, one or several of:\n "accuracy": Accuracy\n "f1": F1 score\n "precision": Precision@10\nExamples:\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wnli\') # \'wnli\' or any of ["copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md"]\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0}\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wiki-ner\')\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0, \'f1\': 1.0}\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'cvit-mkb-clsr\')\n >>> references = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]]\n >>> predictions = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'precision@10\': 1.0}\n\n' def _lowercase ( __snake_case ,__snake_case ) -> Union[str, Any]: return float((preds == labels).mean() ) def _lowercase ( __snake_case ,__snake_case ) -> str: __lowerCAmelCase : str = simple_accuracy(__snake_case ,__snake_case ) __lowerCAmelCase : Any = float(fa_score(y_true=__snake_case ,y_pred=__snake_case ) ) return { "accuracy": acc, "f1": fa, } def _lowercase ( __snake_case ,__snake_case ) -> int: __lowerCAmelCase : Union[str, Any] = np.array(__snake_case ) __lowerCAmelCase : Tuple = np.array(__snake_case ) __lowerCAmelCase : List[Any] = en_sentvecs.shape[0] # mean centering __lowerCAmelCase : Union[str, Any] = en_sentvecs - np.mean(__snake_case ,axis=0 ) __lowerCAmelCase : int = in_sentvecs - np.mean(__snake_case ,axis=0 ) __lowerCAmelCase : Optional[Any] = cdist(__snake_case ,__snake_case ,"cosine" ) __lowerCAmelCase : int = np.array(range(__snake_case ) ) __lowerCAmelCase : int = sim.argsort(axis=1 )[:, :10] __lowerCAmelCase : Optional[Any] = np.any(preds == actual[:, None] ,axis=1 ) return float(matches.mean() ) @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A__ ( datasets.Metric ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: int) -> str: """simple docstring""" if self.config_name not in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", "wiki-ner", ]: raise KeyError( "You should supply a configuration name selected in " "[\"wnli\", \"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", " "\"cvit-mkb-clsr\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\", " "\"wiki-ner\"]") return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("int64") if self.config_name != "cvit-mkb-clsr" else datasets.Sequence(datasets.Value("float32")), "references": datasets.Value("int64") if self.config_name != "cvit-mkb-clsr" else datasets.Sequence(datasets.Value("float32")), }) , codebase_urls=[] , reference_urls=[] , format="numpy" if self.config_name != "cvit-mkb-clsr" else None , ) def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Optional[Any]) -> int: """simple docstring""" if self.config_name == "cvit-mkb-clsr": return {"precision@10": precision_at_aa(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE)} elif self.config_name in ["wiki-ner"]: return acc_and_fa(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) elif self.config_name in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md", ]: return {"accuracy": simple_accuracy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE)} else: raise KeyError( "You should supply a configuration name selected in " "[\"wnli\", \"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", " "\"cvit-mkb-clsr\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\", " "\"wiki-ner\"]")
269
1
"""simple docstring""" def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> float: __lowerCAmelCase : List[str] = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) # formula for sum of series return total def _lowercase ( ) -> Optional[int]: print(sum_of_series(1 ,1 ,10 ) ) if __name__ == "__main__": import doctest doctest.testmod()
269
"""simple docstring""" def _lowercase ( __snake_case ,__snake_case ) -> float: if digit_amount > 0: return round(number - int(__snake_case ) ,__snake_case ) return number - int(__snake_case ) if __name__ == "__main__": print(decimal_isolate(1.53, 0)) print(decimal_isolate(35.345, 1)) print(decimal_isolate(35.345, 2)) print(decimal_isolate(35.345, 3)) print(decimal_isolate(-14.789, 3)) print(decimal_isolate(0, 2)) print(decimal_isolate(-14.123, 1)) print(decimal_isolate(-14.123, 2)) print(decimal_isolate(-14.123, 3))
269
1
"""simple docstring""" from collections.abc import Generator def _lowercase ( ) -> Generator[int, None, None]: __lowerCAmelCase , __lowerCAmelCase : Union[str, Any] = 0, 1 while True: __lowerCAmelCase , __lowerCAmelCase : Optional[int] = b, a + b yield b def _lowercase ( __snake_case = 1_000 ) -> int: __lowerCAmelCase : Dict = 1 __lowerCAmelCase : int = fibonacci_generator() while len(str(next(__snake_case ) ) ) < n: answer += 1 return answer + 1 if __name__ == "__main__": print(solution(int(str(input()).strip())))
269
"""simple docstring""" from ..utils import DummyObject, requires_backends class A__ ( metaclass=__SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = ['torch', 'torchsde'] def __init__( self: int , *_SCREAMING_SNAKE_CASE: Optional[Any] , **_SCREAMING_SNAKE_CASE: str) -> Dict: """simple docstring""" requires_backends(self , ["torch", "torchsde"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls: Optional[Any] , *_SCREAMING_SNAKE_CASE: Optional[int] , **_SCREAMING_SNAKE_CASE: Optional[int]) -> Optional[Any]: """simple docstring""" requires_backends(cls , ["torch", "torchsde"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls: Dict , *_SCREAMING_SNAKE_CASE: List[Any] , **_SCREAMING_SNAKE_CASE: Any) -> Optional[int]: """simple docstring""" requires_backends(cls , ["torch", "torchsde"])
269
1
"""simple docstring""" from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer __snake_case : Optional[Any] = logging.get_logger(__name__) __snake_case : Tuple = { 'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_config_file': 'tokenizer_config.json', } __snake_case : str = { 'vocab_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json' }, 'merges_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt' }, 'tokenizer_config_file': { 'facebook/blenderbot_small-90M': ( 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json' ) }, } __snake_case : Dict = { 'facebook/blenderbot_small-90M': 512, } class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE = BlenderbotSmallTokenizer def __init__( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: List[Any]=None , _SCREAMING_SNAKE_CASE: Optional[Any]=None , _SCREAMING_SNAKE_CASE: List[Any]="<|endoftext|>" , _SCREAMING_SNAKE_CASE: Optional[int]="<|endoftext|>" , _SCREAMING_SNAKE_CASE: str="<|endoftext|>" , _SCREAMING_SNAKE_CASE: Union[str, Any]=False , _SCREAMING_SNAKE_CASE: Any=True , **_SCREAMING_SNAKE_CASE: Optional[Any] , ) -> str: """simple docstring""" super().__init__( ByteLevelBPETokenizer( vocab=_SCREAMING_SNAKE_CASE , merges=_SCREAMING_SNAKE_CASE , add_prefix_space=_SCREAMING_SNAKE_CASE , trim_offsets=_SCREAMING_SNAKE_CASE , ) , bos_token=_SCREAMING_SNAKE_CASE , eos_token=_SCREAMING_SNAKE_CASE , unk_token=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : List[str] = add_prefix_space def _SCREAMING_SNAKE_CASE ( self: Dict , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: Dict=None) -> str: """simple docstring""" __lowerCAmelCase : int = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: List[int] , _SCREAMING_SNAKE_CASE: Optional[List[int]] = None) -> List[int]: """simple docstring""" __lowerCAmelCase : List[str] = [self.sep_token_id] __lowerCAmelCase : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0]
269
"""simple docstring""" def _lowercase ( ) -> int: return 1 def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else two_pence(x - 2 ) + one_pence() def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else five_pence(x - 5 ) + two_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else ten_pence(x - 10 ) + five_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else twenty_pence(x - 20 ) + ten_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else fifty_pence(x - 50 ) + twenty_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else one_pound(x - 100 ) + fifty_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else two_pound(x - 200 ) + one_pound(__snake_case ) def _lowercase ( __snake_case = 200 ) -> int: return two_pound(__snake_case ) if __name__ == "__main__": print(solution(int(input().strip())))
269
1
"""simple docstring""" import math import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from .attention_processor import Attention from .embeddings import get_timestep_embedding from .modeling_utils import ModelMixin class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): '''simple docstring''' @register_to_config def __init__( self: Optional[Any] , _SCREAMING_SNAKE_CASE: int = 128 , _SCREAMING_SNAKE_CASE: int = 256 , _SCREAMING_SNAKE_CASE: float = 2000.0 , _SCREAMING_SNAKE_CASE: int = 768 , _SCREAMING_SNAKE_CASE: int = 12 , _SCREAMING_SNAKE_CASE: int = 12 , _SCREAMING_SNAKE_CASE: int = 64 , _SCREAMING_SNAKE_CASE: int = 2048 , _SCREAMING_SNAKE_CASE: float = 0.1 , ) -> Optional[Any]: """simple docstring""" super().__init__() __lowerCAmelCase : Optional[Any] = nn.Sequential( nn.Linear(_SCREAMING_SNAKE_CASE , d_model * 4 , bias=_SCREAMING_SNAKE_CASE) , nn.SiLU() , nn.Linear(d_model * 4 , d_model * 4 , bias=_SCREAMING_SNAKE_CASE) , nn.SiLU() , ) __lowerCAmelCase : List[str] = nn.Embedding(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = False __lowerCAmelCase : Optional[Any] = nn.Linear(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , bias=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = nn.Dropout(p=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = nn.ModuleList() for lyr_num in range(_SCREAMING_SNAKE_CASE): # FiLM conditional T5 decoder __lowerCAmelCase : List[Any] = DecoderLayer(d_model=_SCREAMING_SNAKE_CASE , d_kv=_SCREAMING_SNAKE_CASE , num_heads=_SCREAMING_SNAKE_CASE , d_ff=_SCREAMING_SNAKE_CASE , dropout_rate=_SCREAMING_SNAKE_CASE) self.decoders.append(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = TaLayerNorm(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = nn.Dropout(p=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = nn.Linear(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , bias=_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: str) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Optional[Any] = torch.mul(query_input.unsqueeze(-1) , key_input.unsqueeze(-2)) return mask.unsqueeze(-3) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: Tuple , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Optional[Any]) -> List[str]: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : str = decoder_input_tokens.shape assert decoder_noise_time.shape == (batch,) # decoder_noise_time is in [0, 1), so rescale to expected timing range. __lowerCAmelCase : Dict = get_timestep_embedding( decoder_noise_time * self.config.max_decoder_noise_time , embedding_dim=self.config.d_model , max_period=self.config.max_decoder_noise_time , ).to(dtype=self.dtype) __lowerCAmelCase : int = self.conditioning_emb(_SCREAMING_SNAKE_CASE).unsqueeze(1) assert conditioning_emb.shape == (batch, 1, self.config.d_model * 4) __lowerCAmelCase : List[Any] = decoder_input_tokens.shape[1] # If we want to use relative positions for audio context, we can just offset # this sequence by the length of encodings_and_masks. __lowerCAmelCase : Optional[Any] = torch.broadcast_to( torch.arange(_SCREAMING_SNAKE_CASE , device=decoder_input_tokens.device) , (batch, seq_length) , ) __lowerCAmelCase : Optional[Any] = self.position_encoding(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = self.continuous_inputs_projection(_SCREAMING_SNAKE_CASE) inputs += position_encodings __lowerCAmelCase : Union[str, Any] = self.dropout(_SCREAMING_SNAKE_CASE) # decoder: No padding present. __lowerCAmelCase : List[str] = torch.ones( decoder_input_tokens.shape[:2] , device=decoder_input_tokens.device , dtype=inputs.dtype) # Translate encoding masks to encoder-decoder masks. __lowerCAmelCase : Any = [(x, self.encoder_decoder_mask(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE)) for x, y in encodings_and_masks] # cross attend style: concat encodings __lowerCAmelCase : List[Any] = torch.cat([x[0] for x in encodings_and_encdec_masks] , dim=1) __lowerCAmelCase : Union[str, Any] = torch.cat([x[1] for x in encodings_and_encdec_masks] , dim=-1) for lyr in self.decoders: __lowerCAmelCase : List[Any] = lyr( _SCREAMING_SNAKE_CASE , conditioning_emb=_SCREAMING_SNAKE_CASE , encoder_hidden_states=_SCREAMING_SNAKE_CASE , encoder_attention_mask=_SCREAMING_SNAKE_CASE , )[0] __lowerCAmelCase : Optional[int] = self.decoder_norm(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = self.post_dropout(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = self.spec_out(_SCREAMING_SNAKE_CASE) return spec_out class A__ ( nn.Module ): '''simple docstring''' def __init__( self: Dict , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Tuple , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Tuple , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: List[Any]=1e-6) -> Optional[Any]: """simple docstring""" super().__init__() __lowerCAmelCase : str = nn.ModuleList() # cond self attention: layer 0 self.layer.append( TaLayerSelfAttentionCond(d_model=_SCREAMING_SNAKE_CASE , d_kv=_SCREAMING_SNAKE_CASE , num_heads=_SCREAMING_SNAKE_CASE , dropout_rate=_SCREAMING_SNAKE_CASE)) # cross attention: layer 1 self.layer.append( TaLayerCrossAttention( d_model=_SCREAMING_SNAKE_CASE , d_kv=_SCREAMING_SNAKE_CASE , num_heads=_SCREAMING_SNAKE_CASE , dropout_rate=_SCREAMING_SNAKE_CASE , layer_norm_epsilon=_SCREAMING_SNAKE_CASE , )) # Film Cond MLP + dropout: last layer self.layer.append( TaLayerFFCond(d_model=_SCREAMING_SNAKE_CASE , d_ff=_SCREAMING_SNAKE_CASE , dropout_rate=_SCREAMING_SNAKE_CASE , layer_norm_epsilon=_SCREAMING_SNAKE_CASE)) def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Any=None , _SCREAMING_SNAKE_CASE: Any=None , _SCREAMING_SNAKE_CASE: int=None , _SCREAMING_SNAKE_CASE: Optional[Any]=None , _SCREAMING_SNAKE_CASE: Any=None , ) -> Optional[int]: """simple docstring""" __lowerCAmelCase : str = self.layer[0]( _SCREAMING_SNAKE_CASE , conditioning_emb=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , ) if encoder_hidden_states is not None: __lowerCAmelCase : Tuple = torch.where(encoder_attention_mask > 0 , 0 , -1e10).to( encoder_hidden_states.dtype) __lowerCAmelCase : Optional[int] = self.layer[1]( _SCREAMING_SNAKE_CASE , key_value_states=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , ) # Apply Film Conditional Feed Forward layer __lowerCAmelCase : List[str] = self.layer[-1](_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) return (hidden_states,) class A__ ( nn.Module ): '''simple docstring''' def __init__( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: List[str]) -> Union[str, Any]: """simple docstring""" super().__init__() __lowerCAmelCase : Optional[Any] = TaLayerNorm(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = TaFiLMLayer(in_features=d_model * 4 , out_features=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = Attention(query_dim=_SCREAMING_SNAKE_CASE , heads=_SCREAMING_SNAKE_CASE , dim_head=_SCREAMING_SNAKE_CASE , out_bias=_SCREAMING_SNAKE_CASE , scale_qk=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = nn.Dropout(_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[int]=None , _SCREAMING_SNAKE_CASE: str=None , ) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : Tuple = self.layer_norm(_SCREAMING_SNAKE_CASE) if conditioning_emb is not None: __lowerCAmelCase : Optional[Any] = self.FiLMLayer(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) # Self-attention block __lowerCAmelCase : str = self.attention(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = hidden_states + self.dropout(_SCREAMING_SNAKE_CASE) return hidden_states class A__ ( nn.Module ): '''simple docstring''' def __init__( self: Optional[int] , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: Dict) -> int: """simple docstring""" super().__init__() __lowerCAmelCase : Optional[Any] = Attention(query_dim=_SCREAMING_SNAKE_CASE , heads=_SCREAMING_SNAKE_CASE , dim_head=_SCREAMING_SNAKE_CASE , out_bias=_SCREAMING_SNAKE_CASE , scale_qk=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = TaLayerNorm(_SCREAMING_SNAKE_CASE , eps=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = nn.Dropout(_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: str=None , _SCREAMING_SNAKE_CASE: Any=None , ) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : str = self.layer_norm(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = self.attention( _SCREAMING_SNAKE_CASE , encoder_hidden_states=_SCREAMING_SNAKE_CASE , attention_mask=attention_mask.squeeze(1) , ) __lowerCAmelCase : List[Any] = hidden_states + self.dropout(_SCREAMING_SNAKE_CASE) return layer_output class A__ ( nn.Module ): '''simple docstring''' def __init__( self: List[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: Any) -> List[Any]: """simple docstring""" super().__init__() __lowerCAmelCase : List[Any] = TaDenseGatedActDense(d_model=_SCREAMING_SNAKE_CASE , d_ff=_SCREAMING_SNAKE_CASE , dropout_rate=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = TaFiLMLayer(in_features=d_model * 4 , out_features=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = TaLayerNorm(_SCREAMING_SNAKE_CASE , eps=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = nn.Dropout(_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Union[str, Any]=None) -> Tuple: """simple docstring""" __lowerCAmelCase : Any = self.layer_norm(_SCREAMING_SNAKE_CASE) if conditioning_emb is not None: __lowerCAmelCase : int = self.film(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = self.DenseReluDense(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = hidden_states + self.dropout(_SCREAMING_SNAKE_CASE) return hidden_states class A__ ( nn.Module ): '''simple docstring''' def __init__( self: Tuple , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: List[str]) -> Union[str, Any]: """simple docstring""" super().__init__() __lowerCAmelCase : Tuple = nn.Linear(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , bias=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = nn.Linear(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , bias=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = nn.Linear(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , bias=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = nn.Dropout(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = NewGELUActivation() def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: Optional[int]) -> Any: """simple docstring""" __lowerCAmelCase : Union[str, Any] = self.act(self.wi_a(_SCREAMING_SNAKE_CASE)) __lowerCAmelCase : Dict = self.wi_a(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = hidden_gelu * hidden_linear __lowerCAmelCase : List[Any] = self.dropout(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = self.wo(_SCREAMING_SNAKE_CASE) return hidden_states class A__ ( nn.Module ): '''simple docstring''' def __init__( self: Optional[Any] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Optional[int]=1e-6) -> str: """simple docstring""" super().__init__() __lowerCAmelCase : List[Any] = nn.Parameter(torch.ones(_SCREAMING_SNAKE_CASE)) __lowerCAmelCase : Union[str, Any] = eps def _SCREAMING_SNAKE_CASE ( self: Optional[int] , _SCREAMING_SNAKE_CASE: str) -> Tuple: """simple docstring""" __lowerCAmelCase : Optional[Any] = hidden_states.to(torch.floataa).pow(2).mean(-1 , keepdim=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = hidden_states * torch.rsqrt(variance + self.variance_epsilon) # convert into half-precision if necessary if self.weight.dtype in [torch.floataa, torch.bfloataa]: __lowerCAmelCase : Dict = hidden_states.to(self.weight.dtype) return self.weight * hidden_states class A__ ( nn.Module ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: torch.Tensor) -> torch.Tensor: """simple docstring""" return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.04_4715 * torch.pow(_SCREAMING_SNAKE_CASE , 3.0)))) class A__ ( nn.Module ): '''simple docstring''' def __init__( self: Tuple , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Optional[int]) -> int: """simple docstring""" super().__init__() __lowerCAmelCase : List[Any] = nn.Linear(_SCREAMING_SNAKE_CASE , out_features * 2 , bias=_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: List[Any]) -> List[Any]: """simple docstring""" __lowerCAmelCase : str = self.scale_bias(_SCREAMING_SNAKE_CASE) __lowerCAmelCase , __lowerCAmelCase : List[Any] = torch.chunk(_SCREAMING_SNAKE_CASE , 2 , -1) __lowerCAmelCase : Tuple = x * (1 + scale) + shift return x
269
"""simple docstring""" import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Optional[int] , _SCREAMING_SNAKE_CASE: str) -> Dict: """simple docstring""" with open(_SCREAMING_SNAKE_CASE , encoding="utf-8") as input_file: __lowerCAmelCase : List[str] = re.compile(r"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)") __lowerCAmelCase : List[Any] = input_file.read() __lowerCAmelCase : Any = regexp.search(_SCREAMING_SNAKE_CASE) return match def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: str) -> Optional[Any]: """simple docstring""" with open(_SCREAMING_SNAKE_CASE , encoding="utf-8") as input_file: __lowerCAmelCase : Any = re.compile(r"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL) __lowerCAmelCase : Optional[int] = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` __lowerCAmelCase : int = regexp.finditer(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = [match for match in matches if match is not None and match.group(1) is not None] return matches[0] if matches else None def _SCREAMING_SNAKE_CASE ( self: str) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[Any] = Path("./datasets") __lowerCAmelCase : Optional[int] = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_encoding_on_file_open(str(_SCREAMING_SNAKE_CASE)): raise AssertionError(F"""open(...) must use utf-8 encoding in {dataset}""") def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> Optional[int]: """simple docstring""" __lowerCAmelCase : Dict = Path("./datasets") __lowerCAmelCase : Union[str, Any] = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_print_statements(str(_SCREAMING_SNAKE_CASE)): raise AssertionError(F"""print statement found in {dataset}. Use datasets.logger/logging instead.""")
269
1
"""simple docstring""" import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Optional[int] , _SCREAMING_SNAKE_CASE: str) -> Dict: """simple docstring""" with open(_SCREAMING_SNAKE_CASE , encoding="utf-8") as input_file: __lowerCAmelCase : List[str] = re.compile(r"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)") __lowerCAmelCase : List[Any] = input_file.read() __lowerCAmelCase : Any = regexp.search(_SCREAMING_SNAKE_CASE) return match def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: str) -> Optional[Any]: """simple docstring""" with open(_SCREAMING_SNAKE_CASE , encoding="utf-8") as input_file: __lowerCAmelCase : Any = re.compile(r"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL) __lowerCAmelCase : Optional[int] = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` __lowerCAmelCase : int = regexp.finditer(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = [match for match in matches if match is not None and match.group(1) is not None] return matches[0] if matches else None def _SCREAMING_SNAKE_CASE ( self: str) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[Any] = Path("./datasets") __lowerCAmelCase : Optional[int] = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_encoding_on_file_open(str(_SCREAMING_SNAKE_CASE)): raise AssertionError(F"""open(...) must use utf-8 encoding in {dataset}""") def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> Optional[int]: """simple docstring""" __lowerCAmelCase : Dict = Path("./datasets") __lowerCAmelCase : Union[str, Any] = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_print_statements(str(_SCREAMING_SNAKE_CASE)): raise AssertionError(F"""print statement found in {dataset}. Use datasets.logger/logging instead.""")
269
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxSeqaSeqConfigWithPast from ...utils import logging if TYPE_CHECKING: from ...feature_extraction_utils import FeatureExtractionMixin from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType __snake_case : Optional[Any] = logging.get_logger(__name__) __snake_case : Optional[Any] = { 'openai/whisper-base': 'https://huggingface.co/openai/whisper-base/resolve/main/config.json', } # fmt: off __snake_case : Any = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 357, 366, 438, 532, 685, 705, 796, 930, 1_058, 1_220, 1_267, 1_279, 1_303, 1_343, 1_377, 1_391, 1_635, 1_782, 1_875, 2_162, 2_361, 2_488, 3_467, 4_008, 4_211, 4_600, 4_808, 5_299, 5_855, 6_329, 7_203, 9_609, 9_959, 10_563, 10_786, 11_420, 11_709, 11_907, 13_163, 13_697, 13_700, 14_808, 15_306, 16_410, 16_791, 17_992, 19_203, 19_510, 20_724, 22_305, 22_935, 27_007, 30_109, 30_420, 33_409, 34_949, 40_283, 40_493, 40_549, 47_282, 49_146, 50_257, 50_359, 50_360, 50_361 ] __snake_case : str = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 359, 503, 522, 542, 873, 893, 902, 918, 922, 931, 1_350, 1_853, 1_982, 2_460, 2_627, 3_246, 3_253, 3_268, 3_536, 3_846, 3_961, 4_183, 4_667, 6_585, 6_647, 7_273, 9_061, 9_383, 10_428, 10_929, 11_938, 12_033, 12_331, 12_562, 13_793, 14_157, 14_635, 15_265, 15_618, 16_553, 16_604, 18_362, 18_956, 20_075, 21_675, 22_520, 26_130, 26_161, 26_435, 28_279, 29_464, 31_650, 32_302, 32_470, 36_865, 42_863, 47_425, 49_870, 50_254, 50_258, 50_360, 50_361, 50_362 ] class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'whisper' SCREAMING_SNAKE_CASE = ['past_key_values'] SCREAMING_SNAKE_CASE = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self: Optional[int] , _SCREAMING_SNAKE_CASE: Any=5_1865 , _SCREAMING_SNAKE_CASE: Optional[Any]=80 , _SCREAMING_SNAKE_CASE: Optional[int]=6 , _SCREAMING_SNAKE_CASE: Any=4 , _SCREAMING_SNAKE_CASE: Dict=6 , _SCREAMING_SNAKE_CASE: Dict=4 , _SCREAMING_SNAKE_CASE: Optional[Any]=1536 , _SCREAMING_SNAKE_CASE: List[str]=1536 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.0 , _SCREAMING_SNAKE_CASE: str=0.0 , _SCREAMING_SNAKE_CASE: List[str]=5_0257 , _SCREAMING_SNAKE_CASE: Optional[Any]=True , _SCREAMING_SNAKE_CASE: List[str]=True , _SCREAMING_SNAKE_CASE: Optional[int]="gelu" , _SCREAMING_SNAKE_CASE: Tuple=256 , _SCREAMING_SNAKE_CASE: str=0.0 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.0 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.0 , _SCREAMING_SNAKE_CASE: List[Any]=0.02 , _SCREAMING_SNAKE_CASE: Tuple=False , _SCREAMING_SNAKE_CASE: Union[str, Any]=1500 , _SCREAMING_SNAKE_CASE: str=448 , _SCREAMING_SNAKE_CASE: Any=5_0256 , _SCREAMING_SNAKE_CASE: Any=5_0256 , _SCREAMING_SNAKE_CASE: List[str]=5_0256 , _SCREAMING_SNAKE_CASE: Dict=None , _SCREAMING_SNAKE_CASE: List[Any]=[220, 5_0256] , _SCREAMING_SNAKE_CASE: Dict=False , _SCREAMING_SNAKE_CASE: str=256 , _SCREAMING_SNAKE_CASE: List[str]=False , _SCREAMING_SNAKE_CASE: Tuple=0.05 , _SCREAMING_SNAKE_CASE: List[str]=10 , _SCREAMING_SNAKE_CASE: str=2 , _SCREAMING_SNAKE_CASE: Any=0.0 , _SCREAMING_SNAKE_CASE: Optional[int]=10 , _SCREAMING_SNAKE_CASE: int=0 , _SCREAMING_SNAKE_CASE: Any=7 , **_SCREAMING_SNAKE_CASE: List[str] , ) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[int] = vocab_size __lowerCAmelCase : Optional[Any] = num_mel_bins __lowerCAmelCase : int = d_model __lowerCAmelCase : List[Any] = encoder_layers __lowerCAmelCase : List[Any] = encoder_attention_heads __lowerCAmelCase : List[str] = decoder_layers __lowerCAmelCase : Tuple = decoder_attention_heads __lowerCAmelCase : Any = decoder_ffn_dim __lowerCAmelCase : Tuple = encoder_ffn_dim __lowerCAmelCase : List[str] = dropout __lowerCAmelCase : Union[str, Any] = attention_dropout __lowerCAmelCase : Union[str, Any] = activation_dropout __lowerCAmelCase : Dict = activation_function __lowerCAmelCase : Tuple = init_std __lowerCAmelCase : str = encoder_layerdrop __lowerCAmelCase : int = decoder_layerdrop __lowerCAmelCase : Optional[int] = use_cache __lowerCAmelCase : Union[str, Any] = encoder_layers __lowerCAmelCase : List[str] = scale_embedding # scale factor will be sqrt(d_model) if True __lowerCAmelCase : int = max_source_positions __lowerCAmelCase : Any = max_target_positions # Audio Classification-specific parameters. Feel free to ignore for other classes. __lowerCAmelCase : Dict = classifier_proj_size __lowerCAmelCase : Dict = use_weighted_layer_sum # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __lowerCAmelCase : int = apply_spec_augment __lowerCAmelCase : Union[str, Any] = mask_time_prob __lowerCAmelCase : str = mask_time_length __lowerCAmelCase : int = mask_time_min_masks __lowerCAmelCase : List[Any] = mask_feature_prob __lowerCAmelCase : Tuple = mask_feature_length __lowerCAmelCase : Any = mask_feature_min_masks __lowerCAmelCase : Union[str, Any] = median_filter_width super().__init__( pad_token_id=_SCREAMING_SNAKE_CASE , bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE , is_encoder_decoder=_SCREAMING_SNAKE_CASE , decoder_start_token_id=_SCREAMING_SNAKE_CASE , suppress_tokens=_SCREAMING_SNAKE_CASE , begin_suppress_tokens=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' @property def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> Mapping[str, Mapping[int, str]]: """simple docstring""" __lowerCAmelCase : List[str] = OrderedDict( [ ("input_features", {0: "batch", 1: "feature_size", 2: "encoder_sequence"}), ]) if self.use_past: __lowerCAmelCase : Tuple = {0: "batch"} else: __lowerCAmelCase : Union[str, Any] = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(_SCREAMING_SNAKE_CASE , direction="inputs") return common_inputs def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: Union["PreTrainedTokenizerBase", "FeatureExtractionMixin"] , _SCREAMING_SNAKE_CASE: int = -1 , _SCREAMING_SNAKE_CASE: int = -1 , _SCREAMING_SNAKE_CASE: bool = False , _SCREAMING_SNAKE_CASE: Optional["TensorType"] = None , _SCREAMING_SNAKE_CASE: int = 2_2050 , _SCREAMING_SNAKE_CASE: float = 5.0 , _SCREAMING_SNAKE_CASE: int = 220 , ) -> Mapping[str, Any]: """simple docstring""" __lowerCAmelCase : int = OrderedDict() __lowerCAmelCase : Optional[Any] = OnnxConfig.generate_dummy_inputs( self , preprocessor=preprocessor.feature_extractor , batch_size=_SCREAMING_SNAKE_CASE , framework=_SCREAMING_SNAKE_CASE , sampling_rate=_SCREAMING_SNAKE_CASE , time_duration=_SCREAMING_SNAKE_CASE , frequency=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : List[str] = encoder_inputs["input_features"].shape[2] __lowerCAmelCase : List[str] = encoder_sequence_length // 2 if self.use_past else seq_length __lowerCAmelCase : List[Any] = super().generate_dummy_inputs( preprocessor.tokenizer , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = encoder_inputs.pop("input_features") __lowerCAmelCase : List[Any] = decoder_inputs.pop("decoder_input_ids") if "past_key_values" in decoder_inputs: __lowerCAmelCase : int = decoder_inputs.pop("past_key_values") return dummy_inputs @property def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> float: """simple docstring""" return 1e-3
269
1
"""simple docstring""" from __future__ import annotations from typing import Any class A__ : '''simple docstring''' def __init__( self: List[str] , _SCREAMING_SNAKE_CASE: int) -> None: """simple docstring""" __lowerCAmelCase : List[Any] = num_of_nodes __lowerCAmelCase : list[list[int]] = [] __lowerCAmelCase : dict[int, int] = {} def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: int) -> None: """simple docstring""" self.m_edges.append([u_node, v_node, weight]) def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: int) -> int: """simple docstring""" if self.m_component[u_node] == u_node: return u_node return self.find_component(self.m_component[u_node]) def _SCREAMING_SNAKE_CASE ( self: Dict , _SCREAMING_SNAKE_CASE: int) -> None: """simple docstring""" if self.m_component[u_node] != u_node: for k in self.m_component: __lowerCAmelCase : Tuple = self.find_component(_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Dict , _SCREAMING_SNAKE_CASE: list[int] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: int) -> None: """simple docstring""" if component_size[u_node] <= component_size[v_node]: __lowerCAmelCase : int = v_node component_size[v_node] += component_size[u_node] self.set_component(_SCREAMING_SNAKE_CASE) elif component_size[u_node] >= component_size[v_node]: __lowerCAmelCase : Any = self.find_component(_SCREAMING_SNAKE_CASE) component_size[u_node] += component_size[v_node] self.set_component(_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> None: """simple docstring""" __lowerCAmelCase : int = [] __lowerCAmelCase : int = 0 __lowerCAmelCase : list[Any] = [-1] * self.m_num_of_nodes # A list of components (initialized to all of the nodes) for node in range(self.m_num_of_nodes): self.m_component.update({node: node}) component_size.append(1) __lowerCAmelCase : Union[str, Any] = self.m_num_of_nodes while num_of_components > 1: for edge in self.m_edges: __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Tuple = edge __lowerCAmelCase : Dict = self.m_component[u] __lowerCAmelCase : Union[str, Any] = self.m_component[v] if u_component != v_component: for component in (u_component, v_component): if ( minimum_weight_edge[component] == -1 or minimum_weight_edge[component][2] > w ): __lowerCAmelCase : List[Any] = [u, v, w] for edge in minimum_weight_edge: if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE): __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Optional[Any] = edge __lowerCAmelCase : Optional[Any] = self.m_component[u] __lowerCAmelCase : Dict = self.m_component[v] if u_component != v_component: mst_weight += w self.union(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) print(F"""Added edge [{u} - {v}]\nAdded weight: {w}\n""") num_of_components -= 1 __lowerCAmelCase : str = [-1] * self.m_num_of_nodes print(F"""The total weight of the minimal spanning tree is: {mst_weight}""") def _lowercase ( ) -> None: pass if __name__ == "__main__": import doctest doctest.testmod()
269
"""simple docstring""" import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration __snake_case : Optional[int] = 50_000 __snake_case : Dict = 5_000 __snake_case , __snake_case : Union[str, Any] = os.path.split(__file__) __snake_case : Any = os.path.join(RESULTS_BASEPATH, 'results', RESULTS_FILENAME.replace('.py', '.json')) @get_duration def _lowercase ( __snake_case ,__snake_case ) -> Dict: for i in range(__snake_case ): __lowerCAmelCase : Union[str, Any] = dataset[i] @get_duration def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> Dict: for i in range(0 ,len(__snake_case ) ,__snake_case ): __lowerCAmelCase : List[str] = dataset[i : i + batch_size] @get_duration def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> Dict: with dataset.formatted_as(type=__snake_case ): for i in range(__snake_case ): __lowerCAmelCase : Union[str, Any] = dataset[i] @get_duration def _lowercase ( __snake_case ,__snake_case ,__snake_case ,__snake_case ) -> str: with dataset.formatted_as(type=__snake_case ): for i in range(0 ,__snake_case ,__snake_case ): __lowerCAmelCase : Optional[int] = dataset[i : i + batch_size] def _lowercase ( ) -> Union[str, Any]: __lowerCAmelCase : Optional[int] = {"num examples": SPEED_TEST_N_EXAMPLES} __lowerCAmelCase : Optional[int] = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted, {"type": "pandas", "length": SMALL_TEST}), (read_formatted, {"type": "torch", "length": SMALL_TEST}), (read_formatted, {"type": "tensorflow", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] __lowerCAmelCase : Any = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print("generating dataset" ) __lowerCAmelCase : int = datasets.Features( {"list": datasets.Sequence(datasets.Value("float32" ) ), "numbers": datasets.Value("float32" )} ) __lowerCAmelCase : str = generate_example_dataset( os.path.join(__snake_case ,"dataset.arrow" ) ,__snake_case ,num_examples=__snake_case ,seq_shapes={"list": (100,)} ,) print("first set of iterations" ) for func, kwargs in functions: print(func.__name__ ,str(__snake_case ) ) __lowerCAmelCase : str = func(__snake_case ,**__snake_case ) print("shuffling dataset" ) __lowerCAmelCase : Optional[int] = dataset.shuffle() print("Second set of iterations (after shuffling" ) for func, kwargs in functions_shuffled: print("shuffled " ,func.__name__ ,str(__snake_case ) ) __lowerCAmelCase : List[str] = func( __snake_case ,**__snake_case ) with open(__snake_case ,"wb" ) as f: f.write(json.dumps(__snake_case ).encode("utf-8" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
269
1
"""simple docstring""" def _lowercase ( __snake_case = 50_000_000 ) -> int: __lowerCAmelCase : List[Any] = set() __lowerCAmelCase : Union[str, Any] = int((limit - 24) ** (1 / 2) ) __lowerCAmelCase : Dict = set(range(3 ,prime_square_limit + 1 ,2 ) ) primes.add(2 ) for p in range(3 ,prime_square_limit + 1 ,2 ): if p not in primes: continue primes.difference_update(set(range(p * p ,prime_square_limit + 1 ,__snake_case ) ) ) for primea in primes: __lowerCAmelCase : List[str] = primea * primea for primea in primes: __lowerCAmelCase : List[Any] = primea * primea * primea if square + cube >= limit - 16: break for primea in primes: __lowerCAmelCase : Tuple = primea * primea * primea * primea __lowerCAmelCase : Dict = square + cube + tetr if total >= limit: break ret.add(__snake_case ) return len(__snake_case ) if __name__ == "__main__": print(F"""{solution() = }""")
269
"""simple docstring""" import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self: Tuple , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: List[str]=13 , _SCREAMING_SNAKE_CASE: Tuple=7 , _SCREAMING_SNAKE_CASE: int=True , _SCREAMING_SNAKE_CASE: Optional[Any]=True , _SCREAMING_SNAKE_CASE: str=False , _SCREAMING_SNAKE_CASE: Optional[Any]=True , _SCREAMING_SNAKE_CASE: int=99 , _SCREAMING_SNAKE_CASE: int=32 , _SCREAMING_SNAKE_CASE: List[str]=5 , _SCREAMING_SNAKE_CASE: Union[str, Any]=4 , _SCREAMING_SNAKE_CASE: int=64 , _SCREAMING_SNAKE_CASE: List[str]="gelu" , _SCREAMING_SNAKE_CASE: str=0.1 , _SCREAMING_SNAKE_CASE: Any=0.1 , _SCREAMING_SNAKE_CASE: Optional[int]=512 , _SCREAMING_SNAKE_CASE: Tuple=16 , _SCREAMING_SNAKE_CASE: Any=2 , _SCREAMING_SNAKE_CASE: List[str]=0.02 , _SCREAMING_SNAKE_CASE: Tuple=3 , _SCREAMING_SNAKE_CASE: Optional[Any]=4 , _SCREAMING_SNAKE_CASE: int=None , _SCREAMING_SNAKE_CASE: int=2 , _SCREAMING_SNAKE_CASE: str=2 , _SCREAMING_SNAKE_CASE: Union[str, Any]=2 , _SCREAMING_SNAKE_CASE: List[Any]=2 , _SCREAMING_SNAKE_CASE: int=4 , _SCREAMING_SNAKE_CASE: List[str]=1 , ) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : List[str] = parent __lowerCAmelCase : Optional[Any] = batch_size __lowerCAmelCase : Union[str, Any] = seq_length __lowerCAmelCase : Optional[Any] = is_training __lowerCAmelCase : Optional[int] = use_input_mask __lowerCAmelCase : Dict = use_token_type_ids __lowerCAmelCase : Dict = use_labels __lowerCAmelCase : Dict = vocab_size __lowerCAmelCase : Tuple = hidden_size __lowerCAmelCase : List[Any] = num_hidden_layers __lowerCAmelCase : Union[str, Any] = num_attention_heads __lowerCAmelCase : Tuple = intermediate_size __lowerCAmelCase : List[Any] = hidden_act __lowerCAmelCase : Optional[Any] = hidden_dropout_prob __lowerCAmelCase : Optional[Any] = attention_probs_dropout_prob __lowerCAmelCase : Optional[int] = max_position_embeddings __lowerCAmelCase : Union[str, Any] = type_vocab_size __lowerCAmelCase : Optional[int] = type_sequence_label_size __lowerCAmelCase : Dict = initializer_range __lowerCAmelCase : Tuple = num_labels __lowerCAmelCase : Optional[Any] = num_choices __lowerCAmelCase : Union[str, Any] = scope __lowerCAmelCase : Optional[Any] = q_groups __lowerCAmelCase : Optional[int] = k_groups __lowerCAmelCase : Any = v_groups __lowerCAmelCase : int = post_attention_groups __lowerCAmelCase : List[str] = intermediate_groups __lowerCAmelCase : Optional[Any] = output_groups def _SCREAMING_SNAKE_CASE ( self: Dict) -> List[str]: """simple docstring""" __lowerCAmelCase : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) __lowerCAmelCase : Union[str, Any] = None if self.use_input_mask: __lowerCAmelCase : Tuple = random_attention_mask([self.batch_size, self.seq_length]) __lowerCAmelCase : Optional[int] = None __lowerCAmelCase : List[Any] = None __lowerCAmelCase : str = None if self.use_labels: __lowerCAmelCase : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size) __lowerCAmelCase : str = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) __lowerCAmelCase : str = ids_tensor([self.batch_size] , self.num_choices) __lowerCAmelCase : Any = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _SCREAMING_SNAKE_CASE ( self: Tuple) -> int: """simple docstring""" return SqueezeBertConfig( embedding_size=self.hidden_size , vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , attention_probs_dropout_prob=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , q_groups=self.q_groups , k_groups=self.k_groups , v_groups=self.v_groups , post_attention_groups=self.post_attention_groups , intermediate_groups=self.intermediate_groups , output_groups=self.output_groups , ) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Tuple , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Tuple) -> Tuple: """simple docstring""" __lowerCAmelCase : List[Any] = SqueezeBertModel(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : List[Any] = model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = model(_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Tuple) -> Dict: """simple docstring""" __lowerCAmelCase : int = SqueezeBertForMaskedLM(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : Dict = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def _SCREAMING_SNAKE_CASE ( self: Optional[int] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Union[str, Any]) -> int: """simple docstring""" __lowerCAmelCase : str = SqueezeBertForQuestionAnswering(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : Union[str, Any] = model( _SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , start_positions=_SCREAMING_SNAKE_CASE , end_positions=_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: Tuple , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: Tuple) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : List[Any] = self.num_labels __lowerCAmelCase : Union[str, Any] = SqueezeBertForSequenceClassification(_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : int = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[int]) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : Dict = self.num_labels __lowerCAmelCase : Optional[int] = SqueezeBertForTokenClassification(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : List[str] = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: int) -> Tuple: """simple docstring""" __lowerCAmelCase : List[str] = self.num_choices __lowerCAmelCase : str = SqueezeBertForMultipleChoice(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : int = input_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() __lowerCAmelCase : Union[str, Any] = input_mask.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() __lowerCAmelCase : str = model( _SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices)) def _SCREAMING_SNAKE_CASE ( self: str) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[int] = self.prepare_config_and_inputs() ((__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase)) : Union[str, Any] = config_and_inputs __lowerCAmelCase : int = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) SCREAMING_SNAKE_CASE = ( { 'feature-extraction': SqueezeBertModel, 'fill-mask': SqueezeBertForMaskedLM, 'question-answering': SqueezeBertForQuestionAnswering, 'text-classification': SqueezeBertForSequenceClassification, 'token-classification': SqueezeBertForTokenClassification, 'zero-shot': SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = False def _SCREAMING_SNAKE_CASE ( self: Dict) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Any = SqueezeBertModelTester(self) __lowerCAmelCase : Optional[int] = ConfigTester(self , config_class=_SCREAMING_SNAKE_CASE , dim=37) def _SCREAMING_SNAKE_CASE ( self: Any) -> Tuple: """simple docstring""" self.config_tester.run_common_tests() def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> Any: """simple docstring""" __lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> List[Any]: """simple docstring""" __lowerCAmelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> int: """simple docstring""" __lowerCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Any) -> int: """simple docstring""" __lowerCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> str: """simple docstring""" __lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*_SCREAMING_SNAKE_CASE) @slow def _SCREAMING_SNAKE_CASE ( self: Any) -> Dict: """simple docstring""" for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase : Optional[Any] = SqueezeBertModel.from_pretrained(_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) @require_sentencepiece @require_tokenizers @require_torch class A__ ( unittest.TestCase ): '''simple docstring''' @slow def _SCREAMING_SNAKE_CASE ( self: int) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[Any] = SqueezeBertForSequenceClassification.from_pretrained("squeezebert/squeezebert-mnli") __lowerCAmelCase : List[Any] = torch.tensor([[1, 2_9414, 232, 328, 740, 1140, 1_2695, 69, 13, 1588, 2]]) __lowerCAmelCase : List[Any] = model(_SCREAMING_SNAKE_CASE)[0] __lowerCAmelCase : Any = torch.Size((1, 3)) self.assertEqual(output.shape , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = torch.tensor([[0.6401, -0.0349, -0.6041]]) self.assertTrue(torch.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=1e-4))
269
1
"""simple docstring""" import importlib import os import sys # This is required to make the module import works (when the python process is running from the root of the repo) sys.path.append('.') def _lowercase ( __snake_case ) -> List[Any]: __lowerCAmelCase : Any = test_file.split(os.path.sep ) if components[0:2] != ["tests", "models"]: raise ValueError( "`test_file` should start with `tests/models/` (with `/` being the OS specific path separator). Got " F"""{test_file} instead.""" ) __lowerCAmelCase : str = components[-1] if not test_fn.endswith("py" ): raise ValueError(F"""`test_file` should be a python file. Got {test_fn} instead.""" ) if not test_fn.startswith("test_modeling_" ): raise ValueError( F"""`test_file` should point to a file name of the form `test_modeling_*.py`. Got {test_fn} instead.""" ) __lowerCAmelCase : int = components[:-1] + [test_fn.replace(".py" ,"" )] __lowerCAmelCase : Tuple = ".".join(__snake_case ) return test_module_path def _lowercase ( __snake_case ) -> Union[str, Any]: __lowerCAmelCase : List[Any] = get_module_path(__snake_case ) __lowerCAmelCase : Optional[int] = importlib.import_module(__snake_case ) return test_module def _lowercase ( __snake_case ) -> Optional[Any]: __lowerCAmelCase : Union[str, Any] = [] __lowerCAmelCase : str = get_test_module(__snake_case ) for attr in dir(__snake_case ): if attr.endswith("ModelTester" ): tester_classes.append(getattr(__snake_case ,__snake_case ) ) # sort with class names return sorted(__snake_case ,key=lambda __snake_case : x.__name__ ) def _lowercase ( __snake_case ) -> Union[str, Any]: __lowerCAmelCase : int = [] __lowerCAmelCase : List[Any] = get_test_module(__snake_case ) for attr in dir(__snake_case ): __lowerCAmelCase : List[Any] = getattr(__snake_case ,__snake_case ) # (TF/Flax)ModelTesterMixin is also an attribute in specific model test module. Let's exclude them by checking # `all_model_classes` is not empty (which also excludes other special classes). __lowerCAmelCase : Optional[int] = getattr(__snake_case ,"all_model_classes" ,[] ) if len(__snake_case ) > 0: test_classes.append(__snake_case ) # sort with class names return sorted(__snake_case ,key=lambda __snake_case : x.__name__ ) def _lowercase ( __snake_case ) -> str: __lowerCAmelCase : Optional[Any] = get_test_classes(__snake_case ) __lowerCAmelCase : List[Any] = set() for test_class in test_classes: model_classes.update(test_class.all_model_classes ) # sort with class names return sorted(__snake_case ,key=lambda __snake_case : x.__name__ ) def _lowercase ( __snake_case ) -> str: __lowerCAmelCase : Any = test_class() if hasattr(__snake_case ,"setUp" ): test.setUp() __lowerCAmelCase : int = None if hasattr(__snake_case ,"model_tester" ): # `(TF/Flax)ModelTesterMixin` has this attribute default to `None`. Let's skip this case. if test.model_tester is not None: __lowerCAmelCase : Optional[int] = test.model_tester.__class__ return model_tester def _lowercase ( __snake_case ,__snake_case ) -> str: __lowerCAmelCase : int = get_test_classes(__snake_case ) __lowerCAmelCase : Tuple = [] for test_class in test_classes: if model_class in test_class.all_model_classes: target_test_classes.append(__snake_case ) # sort with class names return sorted(__snake_case ,key=lambda __snake_case : x.__name__ ) def _lowercase ( __snake_case ,__snake_case ) -> int: __lowerCAmelCase : str = get_test_classes_for_model(__snake_case ,__snake_case ) __lowerCAmelCase : Optional[Any] = [] for test_class in test_classes: __lowerCAmelCase : str = get_model_tester_from_test_class(__snake_case ) if tester_class is not None: tester_classes.append(__snake_case ) # sort with class names return sorted(__snake_case ,key=lambda __snake_case : x.__name__ ) def _lowercase ( __snake_case ) -> Dict: __lowerCAmelCase : Union[str, Any] = get_test_classes(__snake_case ) __lowerCAmelCase : Dict = {test_class: get_model_tester_from_test_class(__snake_case ) for test_class in test_classes} return test_tester_mapping def _lowercase ( __snake_case ) -> Dict: __lowerCAmelCase : int = get_model_classes(__snake_case ) __lowerCAmelCase : List[str] = { model_class: get_test_classes_for_model(__snake_case ,__snake_case ) for model_class in model_classes } return model_test_mapping def _lowercase ( __snake_case ) -> Optional[int]: __lowerCAmelCase : Dict = get_model_classes(__snake_case ) __lowerCAmelCase : str = { model_class: get_tester_classes_for_model(__snake_case ,__snake_case ) for model_class in model_classes } return model_to_tester_mapping def _lowercase ( __snake_case ) -> Any: if isinstance(__snake_case ,__snake_case ): return o elif isinstance(__snake_case ,__snake_case ): return o.__name__ elif isinstance(__snake_case ,(list, tuple) ): return [to_json(__snake_case ) for x in o] elif isinstance(__snake_case ,__snake_case ): return {to_json(__snake_case ): to_json(__snake_case ) for k, v in o.items()} else: return o
269
"""simple docstring""" import os from math import logaa def _lowercase ( __snake_case = "base_exp.txt" ) -> int: __lowerCAmelCase : float = 0 __lowerCAmelCase : Any = 0 for i, line in enumerate(open(os.path.join(os.path.dirname(__snake_case ) ,__snake_case ) ) ): __lowerCAmelCase , __lowerCAmelCase : List[str] = list(map(__snake_case ,line.split("," ) ) ) if x * logaa(__snake_case ) > largest: __lowerCAmelCase : Tuple = x * logaa(__snake_case ) __lowerCAmelCase : Optional[Any] = i + 1 return result if __name__ == "__main__": print(solution())
269
1
"""simple docstring""" import unittest from transformers import SPIECE_UNDERLINE, XLNetTokenizer, XLNetTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin __snake_case : Optional[int] = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = XLNetTokenizer SCREAMING_SNAKE_CASE = XLNetTokenizerFast SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = True def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> str: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __lowerCAmelCase : Tuple = XLNetTokenizer(_SCREAMING_SNAKE_CASE , keep_accents=_SCREAMING_SNAKE_CASE) tokenizer.sanitize_special_tokens() tokenizer.save_pretrained(self.tmpdirname) def _SCREAMING_SNAKE_CASE ( self: Dict) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Optional[Any] = "<s>" __lowerCAmelCase : Tuple = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_SCREAMING_SNAKE_CASE) , _SCREAMING_SNAKE_CASE) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_SCREAMING_SNAKE_CASE) , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Tuple) -> str: """simple docstring""" __lowerCAmelCase : int = list(self.get_tokenizer().get_vocab().keys()) self.assertEqual(vocab_keys[0] , "<unk>") self.assertEqual(vocab_keys[1] , "<s>") self.assertEqual(vocab_keys[-1] , "<eod>") self.assertEqual(len(_SCREAMING_SNAKE_CASE) , 1006) def _SCREAMING_SNAKE_CASE ( self: str) -> Optional[Any]: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1000) def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Any: """simple docstring""" __lowerCAmelCase : List[Any] = XLNetTokenizer(_SCREAMING_SNAKE_CASE , keep_accents=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = tokenizer.tokenize("This is a test") self.assertListEqual(_SCREAMING_SNAKE_CASE , ["▁This", "▁is", "▁a", "▁t", "est"]) self.assertListEqual(tokenizer.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE) , [285, 46, 10, 170, 382]) __lowerCAmelCase : Dict = tokenizer.tokenize("I was born in 92000, and this is falsé.") self.assertListEqual( _SCREAMING_SNAKE_CASE , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) __lowerCAmelCase : Any = tokenizer.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE) self.assertListEqual(_SCREAMING_SNAKE_CASE , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4]) __lowerCAmelCase : str = tokenizer.convert_ids_to_tokens(_SCREAMING_SNAKE_CASE) self.assertListEqual( _SCREAMING_SNAKE_CASE , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : int = XLNetTokenizer(_SCREAMING_SNAKE_CASE , do_lower_case=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = tokenizer.tokenize("I was born in 92000, and this is falsé.") self.assertListEqual( _SCREAMING_SNAKE_CASE , [ SPIECE_UNDERLINE + "", "i", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "se", ".", ] , ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo") , ["▁he", "ll", "o"]) def _SCREAMING_SNAKE_CASE ( self: Dict) -> str: """simple docstring""" __lowerCAmelCase : Dict = XLNetTokenizer(_SCREAMING_SNAKE_CASE , do_lower_case=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = tokenizer.tokenize("I was born in 92000, and this is falsé.") self.assertListEqual( _SCREAMING_SNAKE_CASE , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "se", ".", ] , ) @slow def _SCREAMING_SNAKE_CASE ( self: Tuple) -> List[str]: """simple docstring""" __lowerCAmelCase : Any = XLNetTokenizer.from_pretrained("xlnet-base-cased") __lowerCAmelCase : int = tokenizer.encode("sequence builders" , add_special_tokens=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = tokenizer.encode("multi-sequence build" , add_special_tokens=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = tokenizer.build_inputs_with_special_tokens(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = tokenizer.build_inputs_with_special_tokens(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) assert encoded_sentence == text + [4, 3] assert encoded_pair == text + [4] + text_a + [4, 3] @slow def _SCREAMING_SNAKE_CASE ( self: Dict) -> int: """simple docstring""" __lowerCAmelCase : Dict = {"input_ids": [[17, 2_1442, 270, 17, 10, 1_4645, 318, 34, 17, 4546, 3145, 787, 13, 7752, 2_2018, 23, 21, 17, 4546, 3145, 787, 13, 3352, 1_4431, 13, 5500, 11, 1176, 580, 13, 1_6819, 4797, 23, 17, 10, 1_7135, 658, 19, 457, 7932, 13, 184, 19, 3154, 1_7135, 6468, 19, 1404, 1_2269, 19, 4229, 5356, 1_6264, 46, 19, 17, 2_0545, 1_0395, 9, 9, 9, 11, 28, 6421, 9531, 2_0729, 17, 10, 353, 1_7022, 11, 21, 6421, 9531, 1_6949, 17, 10, 1_1509, 753, 11, 33, 95, 2421, 7385, 956, 1_4431, 2626, 25, 842, 7385, 4836, 21, 1429, 2272, 9855, 3120, 161, 2_4738, 19, 1_3203, 658, 218, 787, 21, 430, 1_8482, 847, 2637, 9, 4, 3], [5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 322, 2_2178, 27, 1064, 22, 956, 13, 1_1101, 1429, 5854, 2_4313, 1_8953, 40, 422, 2_4366, 68, 1758, 37, 1_0483, 1_4257, 31, 207, 263, 21, 203, 3773, 25, 71, 9735, 9, 4, 3], [5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 32, 2049, 3442, 17, 1_3894, 3380, 23, 95, 18, 1_7634, 2288, 9, 4, 3]], "token_type_ids": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2], [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2], [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_SCREAMING_SNAKE_CASE , model_name="xlnet-base-cased" , revision="c841166438c31ec7ca9a106dee7bb312b73ae511" , )
269
"""simple docstring""" import collections import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_flax_cross_test, require_flax, require_torch, require_vision, slow, torch_device, ) from transformers.utils import is_flax_available, is_torch_available, is_vision_available from ...test_modeling_flax_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_flax_bert import FlaxBertModelTester from ..clip.test_modeling_flax_clip import FlaxCLIPVisionModelTester from ..vit.test_modeling_flax_vit import FlaxViTModelTester if is_flax_available(): from transformers import ( FlaxBertModel, FlaxCLIPVisionModel, FlaxVisionTextDualEncoderModel, FlaxViTModel, VisionTextDualEncoderConfig, VisionTextDualEncoderProcessor, ) from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) if is_torch_available(): import torch from transformers import VisionTextDualEncoderModel if is_vision_available(): from PIL import Image def _lowercase ( __snake_case ) -> List[str]: if isinstance(__snake_case ,collections.abc.Iterable ): return x return (x, x) @require_flax class A__ : '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Optional[Any]) -> str: """simple docstring""" pass def _SCREAMING_SNAKE_CASE ( self: str) -> int: """simple docstring""" pass def _SCREAMING_SNAKE_CASE ( self: List[str]) -> Tuple: """simple docstring""" pass def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: np.ndarray , _SCREAMING_SNAKE_CASE: np.ndarray , _SCREAMING_SNAKE_CASE: float) -> List[Any]: """simple docstring""" __lowerCAmelCase : str = np.abs((a - b)).max() self.assertLessEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , F"""Difference between torch and flax is {diff} (>= {tol}).""") def _SCREAMING_SNAKE_CASE ( self: str , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Any=None , **_SCREAMING_SNAKE_CASE: Tuple) -> List[str]: """simple docstring""" __lowerCAmelCase : Tuple = VisionTextDualEncoderConfig.from_vision_text_configs(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = FlaxVisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = model(input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], config.projection_dim)) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], config.projection_dim)) def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Optional[Any]=None , **_SCREAMING_SNAKE_CASE: List[str]) -> Dict: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : List[Any] = self.get_vision_text_model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = {"vision_model": vision_model, "text_model": text_model} __lowerCAmelCase : Optional[Any] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = model(input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], model.config.projection_dim)) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], model.config.projection_dim)) def _SCREAMING_SNAKE_CASE ( self: Dict , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: List[Any]=None , **_SCREAMING_SNAKE_CASE: Tuple) -> str: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : Optional[Any] = self.get_vision_text_model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = {"vision_model": vision_model, "text_model": text_model} __lowerCAmelCase : Any = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = model(input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = output[0] with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = FlaxVisionTextDualEncoderModel.from_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = model(input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = after_output[0] __lowerCAmelCase : Any = np.amax(np.abs(out_a - out_a)) self.assertLessEqual(_SCREAMING_SNAKE_CASE , 1e-3) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Any=None , **_SCREAMING_SNAKE_CASE: List[str]) -> Optional[int]: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : Any = self.get_vision_text_model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = {"vision_model": vision_model, "text_model": text_model} __lowerCAmelCase : Optional[Any] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = model( input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , output_attentions=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = output.vision_model_output.attentions self.assertEqual(len(_SCREAMING_SNAKE_CASE) , vision_config.num_hidden_layers) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) __lowerCAmelCase : List[str] = to_atuple(vision_model.config.image_size) __lowerCAmelCase : Any = to_atuple(vision_model.config.patch_size) __lowerCAmelCase : Union[str, Any] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) __lowerCAmelCase : Tuple = num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len)) __lowerCAmelCase : Union[str, Any] = output.text_model_output.attentions self.assertEqual(len(_SCREAMING_SNAKE_CASE) , text_config.num_hidden_layers) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: int) -> str: """simple docstring""" pt_model.to(_SCREAMING_SNAKE_CASE) pt_model.eval() # prepare inputs __lowerCAmelCase : Union[str, Any] = inputs_dict __lowerCAmelCase : Union[str, Any] = {k: torch.tensor(v.tolist()) for k, v in flax_inputs.items()} with torch.no_grad(): __lowerCAmelCase : Any = pt_model(**_SCREAMING_SNAKE_CASE).to_tuple() __lowerCAmelCase : List[Any] = fx_model(**_SCREAMING_SNAKE_CASE).to_tuple() self.assertEqual(len(_SCREAMING_SNAKE_CASE) , len(_SCREAMING_SNAKE_CASE) , "Output lengths differ between Flax and PyTorch") for fx_output, pt_output in zip(fx_outputs[:4] , pt_outputs[:4]): self.assert_almost_equals(_SCREAMING_SNAKE_CASE , pt_output.numpy() , 4e-2) # PT -> Flax with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = FlaxVisionTextDualEncoderModel.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = fx_model_loaded(**_SCREAMING_SNAKE_CASE).to_tuple() self.assertEqual(len(_SCREAMING_SNAKE_CASE) , len(_SCREAMING_SNAKE_CASE) , "Output lengths differ between Flax and PyTorch") for fx_output_loaded, pt_output in zip(fx_outputs_loaded[:4] , pt_outputs[:4]): self.assert_almost_equals(_SCREAMING_SNAKE_CASE , pt_output.numpy() , 4e-2) # Flax -> PT with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = VisionTextDualEncoderModel.from_pretrained(_SCREAMING_SNAKE_CASE , from_flax=_SCREAMING_SNAKE_CASE) pt_model_loaded.to(_SCREAMING_SNAKE_CASE) pt_model_loaded.eval() with torch.no_grad(): __lowerCAmelCase : Optional[Any] = pt_model_loaded(**_SCREAMING_SNAKE_CASE).to_tuple() self.assertEqual(len(_SCREAMING_SNAKE_CASE) , len(_SCREAMING_SNAKE_CASE) , "Output lengths differ between Flax and PyTorch") for fx_output, pt_output_loaded in zip(fx_outputs[:4] , pt_outputs_loaded[:4]): self.assert_almost_equals(_SCREAMING_SNAKE_CASE , pt_output_loaded.numpy() , 4e-2) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Dict) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Union[str, Any] = VisionTextDualEncoderConfig.from_vision_text_configs(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = VisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = FlaxVisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = fx_state self.check_pt_flax_equivalence(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Tuple , _SCREAMING_SNAKE_CASE: Dict) -> str: """simple docstring""" __lowerCAmelCase : Optional[int] = VisionTextDualEncoderConfig.from_vision_text_configs(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = VisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = FlaxVisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = load_flax_weights_in_pytorch_model(_SCREAMING_SNAKE_CASE , fx_model.params) self.check_pt_flax_equivalence(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> str: """simple docstring""" __lowerCAmelCase : List[str] = self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Dict) -> int: """simple docstring""" __lowerCAmelCase : List[Any] = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> List[Any]: """simple docstring""" __lowerCAmelCase : Dict = self.prepare_config_and_inputs() self.check_save_load(**_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: int) -> Dict: """simple docstring""" __lowerCAmelCase : List[str] = self.prepare_config_and_inputs() self.check_vision_text_output_attention(**_SCREAMING_SNAKE_CASE) @is_pt_flax_cross_test def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> Any: """simple docstring""" __lowerCAmelCase : Dict = self.prepare_config_and_inputs() __lowerCAmelCase : List[Any] = config_inputs_dict.pop("vision_config") __lowerCAmelCase : str = config_inputs_dict.pop("text_config") __lowerCAmelCase : Union[str, Any] = config_inputs_dict self.check_equivalence_pt_to_flax(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) self.check_equivalence_flax_to_pt(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) @slow def _SCREAMING_SNAKE_CASE ( self: str) -> Dict: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : Dict = self.get_pretrained_model_and_inputs() __lowerCAmelCase : Union[str, Any] = model_a(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = outputs[0] with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = FlaxVisionTextDualEncoderModel.from_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = model_a(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = after_outputs[0] __lowerCAmelCase : List[Any] = np.amax(np.abs(out_a - out_a)) self.assertLessEqual(_SCREAMING_SNAKE_CASE , 1e-5) @require_flax class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: List[str]) -> List[Any]: """simple docstring""" __lowerCAmelCase : Any = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( "hf-internal-testing/tiny-random-vit" , "hf-internal-testing/tiny-bert" , vision_from_pt=_SCREAMING_SNAKE_CASE , text_from_pt=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Union[str, Any] = 13 __lowerCAmelCase : Optional[int] = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ]) __lowerCAmelCase : List[Any] = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size) __lowerCAmelCase : List[Any] = random_attention_mask([batch_size, 4]) __lowerCAmelCase : str = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def _SCREAMING_SNAKE_CASE ( self: int , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Union[str, Any]) -> Optional[int]: """simple docstring""" __lowerCAmelCase : List[str] = FlaxViTModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = FlaxBertModel(_SCREAMING_SNAKE_CASE) return vision_model, text_model def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> int: """simple docstring""" __lowerCAmelCase : List[Any] = FlaxViTModelTester(self) __lowerCAmelCase : Optional[Any] = FlaxBertModelTester(self) __lowerCAmelCase : int = vit_model_tester.prepare_config_and_inputs() __lowerCAmelCase : List[str] = bert_model_tester.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase : Tuple = vision_config_and_inputs __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Union[str, Any] = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_torch class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Dict) -> List[str]: """simple docstring""" __lowerCAmelCase : Optional[int] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( "hf-internal-testing/tiny-random-clip" , "hf-internal-testing/tiny-bert" , vision_from_pt=_SCREAMING_SNAKE_CASE , text_from_pt=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Optional[int] = 13 __lowerCAmelCase : List[str] = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ]) __lowerCAmelCase : Any = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size) __lowerCAmelCase : str = random_attention_mask([batch_size, 4]) __lowerCAmelCase : Optional[int] = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Union[str, Any]) -> int: """simple docstring""" __lowerCAmelCase : int = FlaxCLIPVisionModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = FlaxBertModel(_SCREAMING_SNAKE_CASE) return vision_model, text_model def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : List[Any] = FlaxCLIPVisionModelTester(self) __lowerCAmelCase : str = FlaxBertModelTester(self) __lowerCAmelCase : Optional[Any] = clip_model_tester.prepare_config_and_inputs() __lowerCAmelCase : Dict = bert_model_tester.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase : Any = vision_config_and_inputs __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : List[Any] = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_flax @require_vision class A__ ( unittest.TestCase ): '''simple docstring''' @slow def _SCREAMING_SNAKE_CASE ( self: List[str]) -> List[str]: """simple docstring""" __lowerCAmelCase : Dict = FlaxVisionTextDualEncoderModel.from_pretrained("clip-italian/clip-italian" , logit_scale_init_value=1.0) __lowerCAmelCase : str = VisionTextDualEncoderProcessor.from_pretrained("clip-italian/clip-italian") __lowerCAmelCase : Tuple = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png") __lowerCAmelCase : Optional[int] = processor( text=["una foto di un gatto", "una foto di un cane"] , images=_SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE , return_tensors="np") __lowerCAmelCase : List[str] = model(**_SCREAMING_SNAKE_CASE) # verify the logits self.assertEqual(outputs.logits_per_image.shape , (inputs.pixel_values.shape[0], inputs.input_ids.shape[0])) self.assertEqual( outputs.logits_per_text.shape , (inputs.input_ids.shape[0], inputs.pixel_values.shape[0]) , ) __lowerCAmelCase : List[str] = np.array([[1.228_4727, 0.310_4122]]) self.assertTrue(np.allclose(outputs.logits_per_image , _SCREAMING_SNAKE_CASE , atol=1e-3))
269
1
"""simple docstring""" class A__ : '''simple docstring''' def __init__( self: Any , _SCREAMING_SNAKE_CASE: int) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : Optional[Any] = n __lowerCAmelCase : List[Any] = [None] * self.n __lowerCAmelCase : int = 0 # index of the first element __lowerCAmelCase : Tuple = 0 __lowerCAmelCase : Any = 0 def __len__( self: Dict) -> int: """simple docstring""" return self.size def _SCREAMING_SNAKE_CASE ( self: List[str]) -> bool: """simple docstring""" return self.size == 0 def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> List[str]: """simple docstring""" return False if self.is_empty() else self.array[self.front] def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: List[Any]) -> Tuple: """simple docstring""" if self.size >= self.n: raise Exception("QUEUE IS FULL") __lowerCAmelCase : Optional[Any] = data __lowerCAmelCase : Tuple = (self.rear + 1) % self.n self.size += 1 return self def _SCREAMING_SNAKE_CASE ( self: Any) -> int: """simple docstring""" if self.size == 0: raise Exception("UNDERFLOW") __lowerCAmelCase : Any = self.array[self.front] __lowerCAmelCase : int = None __lowerCAmelCase : Union[str, Any] = (self.front + 1) % self.n self.size -= 1 return temp
269
"""simple docstring""" from __future__ import annotations def _lowercase ( __snake_case ,__snake_case ,__snake_case ,__snake_case ) -> list: __lowerCAmelCase : Dict = [] __lowerCAmelCase , __lowerCAmelCase : Any = input_list[low:mid], input_list[mid : high + 1] while left and right: result.append((left if left[0] <= right[0] else right).pop(0 ) ) __lowerCAmelCase : int = result + left + right return input_list def _lowercase ( __snake_case ) -> list: if len(__snake_case ) <= 1: return input_list __lowerCAmelCase : int = list(__snake_case ) # iteration for two-way merging __lowerCAmelCase : Optional[int] = 2 while p <= len(__snake_case ): # getting low, high and middle value for merge-sort of single list for i in range(0 ,len(__snake_case ) ,__snake_case ): __lowerCAmelCase : Union[str, Any] = i __lowerCAmelCase : Tuple = i + p - 1 __lowerCAmelCase : Optional[Any] = (low + high + 1) // 2 __lowerCAmelCase : Any = merge(__snake_case ,__snake_case ,__snake_case ,__snake_case ) # final merge of last two parts if p * 2 >= len(__snake_case ): __lowerCAmelCase : Optional[Any] = i __lowerCAmelCase : Union[str, Any] = merge(__snake_case ,0 ,__snake_case ,len(__snake_case ) - 1 ) break p *= 2 return input_list if __name__ == "__main__": __snake_case : Union[str, Any] = input('Enter numbers separated by a comma:\n').strip() if user_input == "": __snake_case : Optional[int] = [] else: __snake_case : int = [int(item.strip()) for item in user_input.split(',')] print(iter_merge_sort(unsorted))
269
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __snake_case : int = { 'configuration_xlm_roberta_xl': [ 'XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XLMRobertaXLConfig', 'XLMRobertaXLOnnxConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : Optional[int] = [ 'XLM_ROBERTA_XL_PRETRAINED_MODEL_ARCHIVE_LIST', 'XLMRobertaXLForCausalLM', 'XLMRobertaXLForMaskedLM', 'XLMRobertaXLForMultipleChoice', 'XLMRobertaXLForQuestionAnswering', 'XLMRobertaXLForSequenceClassification', 'XLMRobertaXLForTokenClassification', 'XLMRobertaXLModel', 'XLMRobertaXLPreTrainedModel', ] if TYPE_CHECKING: from .configuration_xlm_roberta_xl import ( XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMRobertaXLConfig, XLMRobertaXLOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm_roberta_xl import ( XLM_ROBERTA_XL_PRETRAINED_MODEL_ARCHIVE_LIST, XLMRobertaXLForCausalLM, XLMRobertaXLForMaskedLM, XLMRobertaXLForMultipleChoice, XLMRobertaXLForQuestionAnswering, XLMRobertaXLForSequenceClassification, XLMRobertaXLForTokenClassification, XLMRobertaXLModel, XLMRobertaXLPreTrainedModel, ) else: import sys __snake_case : Optional[int] = _LazyModule(__name__, globals()['__file__'], _import_structure)
269
"""simple docstring""" import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class A__ ( unittest.TestCase ): '''simple docstring''' @slow def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> str: """simple docstring""" __lowerCAmelCase : Optional[Any] = FlaxMTaForConditionalGeneration.from_pretrained("google/mt5-small") __lowerCAmelCase : Optional[Any] = AutoTokenizer.from_pretrained("google/mt5-small") __lowerCAmelCase : Tuple = tokenizer("Hello there" , return_tensors="np").input_ids __lowerCAmelCase : Dict = tokenizer("Hi I am" , return_tensors="np").input_ids __lowerCAmelCase : str = shift_tokens_right(_SCREAMING_SNAKE_CASE , model.config.pad_token_id , model.config.decoder_start_token_id) __lowerCAmelCase : Optional[int] = model(_SCREAMING_SNAKE_CASE , decoder_input_ids=_SCREAMING_SNAKE_CASE).logits __lowerCAmelCase : int = optax.softmax_cross_entropy(_SCREAMING_SNAKE_CASE , onehot(_SCREAMING_SNAKE_CASE , logits.shape[-1])).mean() __lowerCAmelCase : List[str] = -(labels.shape[-1] * loss.item()) __lowerCAmelCase : str = -84.9127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE) < 1e-4)
269
1
"""simple docstring""" from random import shuffle import tensorflow as tf from numpy import array def _lowercase ( __snake_case ,__snake_case ) -> Tuple: __lowerCAmelCase : str = int(__snake_case ) assert noofclusters < len(__snake_case ) # Find out the dimensionality __lowerCAmelCase : Optional[Any] = len(vectors[0] ) # Will help select random centroids from among the available vectors __lowerCAmelCase : Optional[int] = list(range(len(__snake_case ) ) ) shuffle(__snake_case ) # GRAPH OF COMPUTATION # We initialize a new graph and set it as the default during each run # of this algorithm. This ensures that as this function is called # multiple times, the default graph doesn't keep getting crowded with # unused ops and Variables from previous function calls. __lowerCAmelCase : List[str] = tf.Graph() with graph.as_default(): # SESSION OF COMPUTATION __lowerCAmelCase : Dict = tf.Session() ##CONSTRUCTING THE ELEMENTS OF COMPUTATION ##First lets ensure we have a Variable vector for each centroid, ##initialized to one of the vectors from the available data points __lowerCAmelCase : Dict = [ tf.Variable(vectors[vector_indices[i]] ) for i in range(__snake_case ) ] ##These nodes will assign the centroid Variables the appropriate ##values __lowerCAmelCase : Union[str, Any] = tf.placeholder("float64" ,[dim] ) __lowerCAmelCase : List[Any] = [] for centroid in centroids: cent_assigns.append(tf.assign(__snake_case ,__snake_case ) ) ##Variables for cluster assignments of individual vectors(initialized ##to 0 at first) __lowerCAmelCase : Optional[Any] = [tf.Variable(0 ) for i in range(len(__snake_case ) )] ##These nodes will assign an assignment Variable the appropriate ##value __lowerCAmelCase : Union[str, Any] = tf.placeholder("int32" ) __lowerCAmelCase : Union[str, Any] = [] for assignment in assignments: cluster_assigns.append(tf.assign(__snake_case ,__snake_case ) ) ##Now lets construct the node that will compute the mean # The placeholder for the input __lowerCAmelCase : Union[str, Any] = tf.placeholder("float" ,[None, dim] ) # The Node/op takes the input and computes a mean along the 0th # dimension, i.e. the list of input vectors __lowerCAmelCase : Optional[int] = tf.reduce_mean(__snake_case ,0 ) ##Node for computing Euclidean distances # Placeholders for input __lowerCAmelCase : Union[str, Any] = tf.placeholder("float" ,[dim] ) __lowerCAmelCase : List[Any] = tf.placeholder("float" ,[dim] ) __lowerCAmelCase : Tuple = tf.sqrt(tf.reduce_sum(tf.pow(tf.sub(__snake_case ,__snake_case ) ,2 ) ) ) ##This node will figure out which cluster to assign a vector to, ##based on Euclidean distances of the vector from the centroids. # Placeholder for input __lowerCAmelCase : Union[str, Any] = tf.placeholder("float" ,[noofclusters] ) __lowerCAmelCase : List[Any] = tf.argmin(__snake_case ,0 ) ##INITIALIZING STATE VARIABLES ##This will help initialization of all Variables defined with respect ##to the graph. The Variable-initializer should be defined after ##all the Variables have been constructed, so that each of them ##will be included in the initialization. __lowerCAmelCase : Union[str, Any] = tf.initialize_all_variables() # Initialize all variables sess.run(__snake_case ) ##CLUSTERING ITERATIONS # Now perform the Expectation-Maximization steps of K-Means clustering # iterations. To keep things simple, we will only do a set number of # iterations, instead of using a Stopping Criterion. __lowerCAmelCase : Any = 100 for _ in range(__snake_case ): ##EXPECTATION STEP ##Based on the centroid locations till last iteration, compute ##the _expected_ centroid assignments. # Iterate over each vector for vector_n in range(len(__snake_case ) ): __lowerCAmelCase : Union[str, Any] = vectors[vector_n] # Compute Euclidean distance between this vector and each # centroid. Remember that this list cannot be named #'centroid_distances', since that is the input to the # cluster assignment node. __lowerCAmelCase : Any = [ sess.run(__snake_case ,feed_dict={va: vect, va: sess.run(__snake_case )} ) for centroid in centroids ] # Now use the cluster assignment node, with the distances # as the input __lowerCAmelCase : Union[str, Any] = sess.run( __snake_case ,feed_dict={centroid_distances: distances} ) # Now assign the value to the appropriate state variable sess.run( cluster_assigns[vector_n] ,feed_dict={assignment_value: assignment} ) ##MAXIMIZATION STEP # Based on the expected state computed from the Expectation Step, # compute the locations of the centroids so as to maximize the # overall objective of minimizing within-cluster Sum-of-Squares for cluster_n in range(__snake_case ): # Collect all the vectors assigned to this cluster __lowerCAmelCase : Optional[Any] = [ vectors[i] for i in range(len(__snake_case ) ) if sess.run(assignments[i] ) == cluster_n ] # Compute new centroid location __lowerCAmelCase : Tuple = sess.run( __snake_case ,feed_dict={mean_input: array(__snake_case )} ) # Assign value to appropriate variable sess.run( cent_assigns[cluster_n] ,feed_dict={centroid_value: new_location} ) # Return centroids and assignments __lowerCAmelCase : Tuple = sess.run(__snake_case ) __lowerCAmelCase : Any = sess.run(__snake_case ) return centroids, assignments
269
"""simple docstring""" import re def _lowercase ( __snake_case ) -> str: if len(re.findall("[ATCG]" ,__snake_case ) ) != len(__snake_case ): raise ValueError("Invalid Strand" ) return dna.translate(dna.maketrans("ATCG" ,"TAGC" ) ) if __name__ == "__main__": import doctest doctest.testmod()
269
1
"""simple docstring""" def _lowercase ( __snake_case = 10**12 ) -> int: __lowerCAmelCase : Tuple = 1 __lowerCAmelCase : List[Any] = 0 __lowerCAmelCase : List[Any] = 1 __lowerCAmelCase : Union[str, Any] = 1 while numerator <= 2 * min_total - 1: prev_numerator += 2 * numerator numerator += 2 * prev_numerator prev_denominator += 2 * denominator denominator += 2 * prev_denominator return (denominator + 1) // 2 if __name__ == "__main__": print(F"""{solution() = }""")
269
"""simple docstring""" import os import unittest from transformers import FunnelTokenizer, FunnelTokenizerFast from transformers.models.funnel.tokenization_funnel import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = FunnelTokenizer SCREAMING_SNAKE_CASE = FunnelTokenizerFast SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = True def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Optional[int]: """simple docstring""" super().setUp() __lowerCAmelCase : str = [ "<unk>", "<cls>", "<sep>", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] __lowerCAmelCase : int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"]) with open(self.vocab_file , "w" , encoding="utf-8") as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens])) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , **_SCREAMING_SNAKE_CASE: Union[str, Any]) -> int: """simple docstring""" return FunnelTokenizer.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Any , **_SCREAMING_SNAKE_CASE: Any) -> str: """simple docstring""" return FunnelTokenizerFast.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: str) -> Any: """simple docstring""" __lowerCAmelCase : Union[str, Any] = "UNwant\u00E9d,running" __lowerCAmelCase : str = "unwanted, running" return input_text, output_text def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> List[str]: """simple docstring""" __lowerCAmelCase : Any = self.tokenizer_class(self.vocab_file) __lowerCAmelCase : Any = tokenizer.tokenize("UNwant\u00E9d,running") self.assertListEqual(_SCREAMING_SNAKE_CASE , ["un", "##want", "##ed", ",", "runn", "##ing"]) self.assertListEqual(tokenizer.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE) , [7, 4, 5, 10, 8, 9]) def _SCREAMING_SNAKE_CASE ( self: List[str]) -> List[str]: """simple docstring""" __lowerCAmelCase : Optional[Any] = self.get_tokenizers(do_lower_case=_SCREAMING_SNAKE_CASE) for tokenizer in tokenizers: __lowerCAmelCase : List[str] = tokenizer("UNwant\u00E9d,running") __lowerCAmelCase : Optional[int] = len(inputs["input_ids"]) - 1 self.assertListEqual(inputs["token_type_ids"] , [2] + [0] * sentence_len) __lowerCAmelCase : List[str] = tokenizer("UNwant\u00E9d,running" , "UNwant\u00E9d,running") self.assertListEqual(inputs["token_type_ids"] , [2] + [0] * sentence_len + [1] * sentence_len)
269
1
"""simple docstring""" import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration __snake_case : Optional[int] = 50_000 __snake_case : Dict = 5_000 __snake_case , __snake_case : Union[str, Any] = os.path.split(__file__) __snake_case : Any = os.path.join(RESULTS_BASEPATH, 'results', RESULTS_FILENAME.replace('.py', '.json')) @get_duration def _lowercase ( __snake_case ,__snake_case ) -> Dict: for i in range(__snake_case ): __lowerCAmelCase : Union[str, Any] = dataset[i] @get_duration def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> Dict: for i in range(0 ,len(__snake_case ) ,__snake_case ): __lowerCAmelCase : List[str] = dataset[i : i + batch_size] @get_duration def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> Dict: with dataset.formatted_as(type=__snake_case ): for i in range(__snake_case ): __lowerCAmelCase : Union[str, Any] = dataset[i] @get_duration def _lowercase ( __snake_case ,__snake_case ,__snake_case ,__snake_case ) -> str: with dataset.formatted_as(type=__snake_case ): for i in range(0 ,__snake_case ,__snake_case ): __lowerCAmelCase : Optional[int] = dataset[i : i + batch_size] def _lowercase ( ) -> Union[str, Any]: __lowerCAmelCase : Optional[int] = {"num examples": SPEED_TEST_N_EXAMPLES} __lowerCAmelCase : Optional[int] = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted, {"type": "pandas", "length": SMALL_TEST}), (read_formatted, {"type": "torch", "length": SMALL_TEST}), (read_formatted, {"type": "tensorflow", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] __lowerCAmelCase : Any = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print("generating dataset" ) __lowerCAmelCase : int = datasets.Features( {"list": datasets.Sequence(datasets.Value("float32" ) ), "numbers": datasets.Value("float32" )} ) __lowerCAmelCase : str = generate_example_dataset( os.path.join(__snake_case ,"dataset.arrow" ) ,__snake_case ,num_examples=__snake_case ,seq_shapes={"list": (100,)} ,) print("first set of iterations" ) for func, kwargs in functions: print(func.__name__ ,str(__snake_case ) ) __lowerCAmelCase : str = func(__snake_case ,**__snake_case ) print("shuffling dataset" ) __lowerCAmelCase : Optional[int] = dataset.shuffle() print("Second set of iterations (after shuffling" ) for func, kwargs in functions_shuffled: print("shuffled " ,func.__name__ ,str(__snake_case ) ) __lowerCAmelCase : List[str] = func( __snake_case ,**__snake_case ) with open(__snake_case ,"wb" ) as f: f.write(json.dumps(__snake_case ).encode("utf-8" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
269
"""simple docstring""" from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def _lowercase ( __snake_case = "laptop" ) -> DataFrame: __lowerCAmelCase : str = F"""https://www.amazon.in/laptop/s?k={product}""" __lowerCAmelCase : Union[str, Any] = { "User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36\n (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36", "Accept-Language": "en-US, en;q=0.5", } __lowerCAmelCase : List[str] = BeautifulSoup(requests.get(__snake_case ,headers=__snake_case ).text ) # Initialize a Pandas dataframe with the column titles __lowerCAmelCase : Dict = DataFrame( columns=[ "Product Title", "Product Link", "Current Price of the product", "Product Rating", "MRP of the product", "Discount", ] ) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( "div" ,attrs={"class": "s-result-item", "data-component-type": "s-search-result"} ,) ,soup.find_all("div" ,attrs={"class": "a-row a-size-base a-color-base"} ) ,): try: __lowerCAmelCase : Any = item.ha.text __lowerCAmelCase : Union[str, Any] = "https://www.amazon.in/" + item.ha.a["href"] __lowerCAmelCase : Any = item.find("span" ,attrs={"class": "a-offscreen"} ).text try: __lowerCAmelCase : Union[str, Any] = item.find("span" ,attrs={"class": "a-icon-alt"} ).text except AttributeError: __lowerCAmelCase : Optional[Any] = "Not available" try: __lowerCAmelCase : Union[str, Any] = ( "₹" + item.find( "span" ,attrs={"class": "a-price a-text-price"} ).text.split("₹" )[1] ) except AttributeError: __lowerCAmelCase : Dict = "" try: __lowerCAmelCase : str = float( ( ( float(product_mrp.strip("₹" ).replace("," ,"" ) ) - float(product_price.strip("₹" ).replace("," ,"" ) ) ) / float(product_mrp.strip("₹" ).replace("," ,"" ) ) ) * 100 ) except ValueError: __lowerCAmelCase : List[str] = float("nan" ) except AttributeError: pass __lowerCAmelCase : int = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] __lowerCAmelCase : Union[str, Any] = " " __lowerCAmelCase : Union[str, Any] = " " data_frame.index += 1 return data_frame if __name__ == "__main__": __snake_case : Any = 'headphones' get_amazon_product_data(product).to_csv(F"""Amazon Product Data for {product}.csv""")
269
1
"""simple docstring""" import numpy as np from scipy.spatial.distance import cdist from sklearn.metrics import fa_score import datasets __snake_case : Optional[int] = '\\n @inproceedings{kakwani2020indicnlpsuite,\n title={{IndicNLPSuite: Monolingual Corpora, Evaluation Benchmarks and Pre-trained Multilingual Language Models for Indian Languages}},\n author={Divyanshu Kakwani and Anoop Kunchukuttan and Satish Golla and Gokul N.C. and Avik Bhattacharyya and Mitesh M. Khapra and Pratyush Kumar},\n year={2020},\n booktitle={Findings of EMNLP},\n}\n' __snake_case : str = '\\n IndicGLUE is a natural language understanding benchmark for Indian languages. It contains a wide\n variety of tasks and covers 11 major Indian languages - as, bn, gu, hi, kn, ml, mr, or, pa, ta, te.\n' __snake_case : str = '\nCompute IndicGLUE evaluation metric associated to each IndicGLUE dataset.\nArgs:\n predictions: list of predictions to score (as int64),\n except for \'cvit-mkb-clsr\' where each prediction is a vector (of float32).\n references: list of ground truth labels corresponding to the predictions (as int64),\n except for \'cvit-mkb-clsr\' where each reference is a vector (of float32).\nReturns: depending on the IndicGLUE subset, one or several of:\n "accuracy": Accuracy\n "f1": F1 score\n "precision": Precision@10\nExamples:\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wnli\') # \'wnli\' or any of ["copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md"]\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0}\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wiki-ner\')\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0, \'f1\': 1.0}\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'cvit-mkb-clsr\')\n >>> references = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]]\n >>> predictions = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'precision@10\': 1.0}\n\n' def _lowercase ( __snake_case ,__snake_case ) -> Union[str, Any]: return float((preds == labels).mean() ) def _lowercase ( __snake_case ,__snake_case ) -> str: __lowerCAmelCase : str = simple_accuracy(__snake_case ,__snake_case ) __lowerCAmelCase : Any = float(fa_score(y_true=__snake_case ,y_pred=__snake_case ) ) return { "accuracy": acc, "f1": fa, } def _lowercase ( __snake_case ,__snake_case ) -> int: __lowerCAmelCase : Union[str, Any] = np.array(__snake_case ) __lowerCAmelCase : Tuple = np.array(__snake_case ) __lowerCAmelCase : List[Any] = en_sentvecs.shape[0] # mean centering __lowerCAmelCase : Union[str, Any] = en_sentvecs - np.mean(__snake_case ,axis=0 ) __lowerCAmelCase : int = in_sentvecs - np.mean(__snake_case ,axis=0 ) __lowerCAmelCase : Optional[Any] = cdist(__snake_case ,__snake_case ,"cosine" ) __lowerCAmelCase : int = np.array(range(__snake_case ) ) __lowerCAmelCase : int = sim.argsort(axis=1 )[:, :10] __lowerCAmelCase : Optional[Any] = np.any(preds == actual[:, None] ,axis=1 ) return float(matches.mean() ) @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A__ ( datasets.Metric ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: int) -> str: """simple docstring""" if self.config_name not in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", "wiki-ner", ]: raise KeyError( "You should supply a configuration name selected in " "[\"wnli\", \"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", " "\"cvit-mkb-clsr\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\", " "\"wiki-ner\"]") return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("int64") if self.config_name != "cvit-mkb-clsr" else datasets.Sequence(datasets.Value("float32")), "references": datasets.Value("int64") if self.config_name != "cvit-mkb-clsr" else datasets.Sequence(datasets.Value("float32")), }) , codebase_urls=[] , reference_urls=[] , format="numpy" if self.config_name != "cvit-mkb-clsr" else None , ) def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Optional[Any]) -> int: """simple docstring""" if self.config_name == "cvit-mkb-clsr": return {"precision@10": precision_at_aa(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE)} elif self.config_name in ["wiki-ner"]: return acc_and_fa(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) elif self.config_name in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md", ]: return {"accuracy": simple_accuracy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE)} else: raise KeyError( "You should supply a configuration name selected in " "[\"wnli\", \"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", " "\"cvit-mkb-clsr\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\", " "\"wiki-ner\"]")
269
"""simple docstring""" import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('TEST_SAGEMAKER' , 'False' ) ) is not True , reason='Skipping test because should only be run when releasing minor transformers version' , ) @pytest.mark.usefixtures('sm_env' ) @parameterized_class( [ { 'framework': 'pytorch', 'script': 'run_glue.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_5_0, 'eval_accuracy': 0.6, 'eval_loss': 0.9}, }, { 'framework': 'tensorflow', 'script': 'run_tf.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_0_0, 'eval_accuracy': 0.3, 'eval_loss': 0.9}, }, ] ) class A__ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: int) -> Tuple: """simple docstring""" if self.framework == "pytorch": subprocess.run( F"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() , encoding="utf-8" , check=_SCREAMING_SNAKE_CASE , ) assert hasattr(self , "env") def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Any=1) -> Dict: """simple docstring""" return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=F"""{self.env.base_job_name}-single""" , instance_count=_SCREAMING_SNAKE_CASE , instance_type=self.instance_type , debugger_hook_config=_SCREAMING_SNAKE_CASE , hyperparameters={**self.env.hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version="py36" , ) def _SCREAMING_SNAKE_CASE ( self: str , _SCREAMING_SNAKE_CASE: List[Any]) -> Optional[Any]: """simple docstring""" TrainingJobAnalytics(_SCREAMING_SNAKE_CASE).export_csv(F"""{self.env.test_path}/{job_name}_metrics.csv""") def _SCREAMING_SNAKE_CASE ( self: str) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Tuple = self.create_estimator() # run training estimator.fit() # result dataframe __lowerCAmelCase : Tuple = TrainingJobAnalytics(estimator.latest_training_job.name).dataframe() # extract kpis __lowerCAmelCase : Optional[Any] = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"]) __lowerCAmelCase : Union[str, Any] = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"]) # get train time from SageMaker job, this includes starting, preprocessing, stopping __lowerCAmelCase : Tuple = ( Session().describe_training_job(estimator.latest_training_job.name).get("TrainingTimeInSeconds" , 99_9999) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy) assert all(t <= self.results["eval_loss"] for t in eval_loss) # dump tests result into json file to share in PR with open(F"""{estimator.latest_training_job.name}.json""" , "w") as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , _SCREAMING_SNAKE_CASE)
269
1
"""simple docstring""" # # This a `torch.distributed` diagnostics script that checks that all GPUs in the cluster (one or # many nodes) can talk to each other via nccl and allocate gpu memory. # # To run first adjust the number of processes and nodes: # # python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # You may need to add --master_addr $MASTER_ADDR --master_port $MASTER_PORT if using a custom addr:port # # You can also use the rdzv API: --rdzv_endpoint $MASTER_ADDR:$MASTER_PORT --rdzv_backend c10d # # use torch.distributed.launch instead of torch.distributed.run for torch < 1.9 # # If you get a hanging in `barrier` calls you have some network issues, you may try to debug this with: # # NCCL_DEBUG=INFO python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # which should tell you what's going on behind the scenes. # # # This script can be run via `srun` in the SLURM environment as well. Here is a SLURM script that # runs on 2 nodes of 4 gpus per node: # # #SBATCH --job-name=test-nodes # name # #SBATCH --nodes=2 # nodes # #SBATCH --ntasks-per-node=1 # crucial - only 1 task per dist per node! # #SBATCH --cpus-per-task=10 # number of cores per tasks # #SBATCH --gres=gpu:4 # number of gpus # #SBATCH --time 0:05:00 # maximum execution time (HH:MM:SS) # #SBATCH --output=%x-%j.out # output file name # # GPUS_PER_NODE=4 # MASTER_ADDR=$(scontrol show hostnames $SLURM_JOB_NODELIST | head -n 1) # MASTER_PORT=6000 # # srun --jobid $SLURM_JOBID bash -c 'python -m torch.distributed.run \ # --nproc_per_node $GPUS_PER_NODE --nnodes $SLURM_NNODES --node_rank $SLURM_PROCID \ # --master_addr $MASTER_ADDR --master_port $MASTER_PORT \ # torch-distributed-gpu-test.py' # import fcntl import os import socket import torch import torch.distributed as dist def _lowercase ( *__snake_case ) -> Optional[Any]: with open(__snake_case ,"r" ) as fh: fcntl.flock(__snake_case ,fcntl.LOCK_EX ) try: print(*__snake_case ) finally: fcntl.flock(__snake_case ,fcntl.LOCK_UN ) __snake_case : List[Any] = int(os.environ['LOCAL_RANK']) torch.cuda.set_device(local_rank) __snake_case : List[str] = torch.device('cuda', local_rank) __snake_case : Optional[Any] = socket.gethostname() __snake_case : str = F"""[{hostname}-{local_rank}]""" try: # test distributed dist.init_process_group('nccl') dist.all_reduce(torch.ones(1).to(device), op=dist.ReduceOp.SUM) dist.barrier() # test cuda is available and can allocate memory torch.cuda.is_available() torch.ones(1).cuda(local_rank) # global rank __snake_case : Tuple = dist.get_rank() __snake_case : Any = dist.get_world_size() printflock(F"""{gpu} is OK (global rank: {rank}/{world_size})""") dist.barrier() if rank == 0: printflock(F"""pt={torch.__version__}, cuda={torch.version.cuda}, nccl={torch.cuda.nccl.version()}""") except Exception: printflock(F"""{gpu} is broken""") raise
269
"""simple docstring""" from typing import TYPE_CHECKING from ..utils import _LazyModule __snake_case : Optional[int] = { 'config': [ 'EXTERNAL_DATA_FORMAT_SIZE_LIMIT', 'OnnxConfig', 'OnnxConfigWithPast', 'OnnxSeq2SeqConfigWithPast', 'PatchingSpec', ], 'convert': ['export', 'validate_model_outputs'], 'features': ['FeaturesManager'], 'utils': ['ParameterFormat', 'compute_serialized_parameters_size'], } if TYPE_CHECKING: from .config import ( EXTERNAL_DATA_FORMAT_SIZE_LIMIT, OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast, PatchingSpec, ) from .convert import export, validate_model_outputs from .features import FeaturesManager from .utils import ParameterFormat, compute_serialized_parameters_size else: import sys __snake_case : Any = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
269
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case : List[str] = logging.get_logger(__name__) __snake_case : Dict = {} class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'llama' SCREAMING_SNAKE_CASE = ['past_key_values'] def __init__( self: int , _SCREAMING_SNAKE_CASE: Optional[Any]=3_2000 , _SCREAMING_SNAKE_CASE: str=4096 , _SCREAMING_SNAKE_CASE: Tuple=1_1008 , _SCREAMING_SNAKE_CASE: Dict=32 , _SCREAMING_SNAKE_CASE: List[Any]=32 , _SCREAMING_SNAKE_CASE: Union[str, Any]=None , _SCREAMING_SNAKE_CASE: Optional[Any]="silu" , _SCREAMING_SNAKE_CASE: Optional[Any]=2048 , _SCREAMING_SNAKE_CASE: Dict=0.02 , _SCREAMING_SNAKE_CASE: int=1e-6 , _SCREAMING_SNAKE_CASE: List[Any]=True , _SCREAMING_SNAKE_CASE: Tuple=0 , _SCREAMING_SNAKE_CASE: Dict=1 , _SCREAMING_SNAKE_CASE: List[Any]=2 , _SCREAMING_SNAKE_CASE: str=1 , _SCREAMING_SNAKE_CASE: Optional[Any]=False , _SCREAMING_SNAKE_CASE: Optional[Any]=None , **_SCREAMING_SNAKE_CASE: int , ) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : int = vocab_size __lowerCAmelCase : List[str] = max_position_embeddings __lowerCAmelCase : Dict = hidden_size __lowerCAmelCase : Union[str, Any] = intermediate_size __lowerCAmelCase : str = num_hidden_layers __lowerCAmelCase : List[Any] = num_attention_heads # for backward compatibility if num_key_value_heads is None: __lowerCAmelCase : Optional[int] = num_attention_heads __lowerCAmelCase : Union[str, Any] = num_key_value_heads __lowerCAmelCase : str = hidden_act __lowerCAmelCase : Dict = initializer_range __lowerCAmelCase : str = rms_norm_eps __lowerCAmelCase : str = pretraining_tp __lowerCAmelCase : int = use_cache __lowerCAmelCase : int = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=_SCREAMING_SNAKE_CASE , bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE , tie_word_embeddings=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) def _SCREAMING_SNAKE_CASE ( self: int) -> List[Any]: """simple docstring""" if self.rope_scaling is None: return if not isinstance(self.rope_scaling , _SCREAMING_SNAKE_CASE) or len(self.rope_scaling) != 2: raise ValueError( "`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, " F"""got {self.rope_scaling}""") __lowerCAmelCase : List[Any] = self.rope_scaling.get("type" , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = self.rope_scaling.get("factor" , _SCREAMING_SNAKE_CASE) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( F"""`rope_scaling`'s name field must be one of ['linear', 'dynamic'], got {rope_scaling_type}""") if rope_scaling_factor is None or not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) or rope_scaling_factor <= 1.0: raise ValueError(F"""`rope_scaling`'s factor field must be an float > 1, got {rope_scaling_factor}""")
269
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices __snake_case : Optional[int] = logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'maskformer-swin' SCREAMING_SNAKE_CASE = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self: Optional[int] , _SCREAMING_SNAKE_CASE: int=224 , _SCREAMING_SNAKE_CASE: Tuple=4 , _SCREAMING_SNAKE_CASE: int=3 , _SCREAMING_SNAKE_CASE: List[Any]=96 , _SCREAMING_SNAKE_CASE: Union[str, Any]=[2, 2, 6, 2] , _SCREAMING_SNAKE_CASE: Any=[3, 6, 12, 24] , _SCREAMING_SNAKE_CASE: List[str]=7 , _SCREAMING_SNAKE_CASE: List[str]=4.0 , _SCREAMING_SNAKE_CASE: Optional[int]=True , _SCREAMING_SNAKE_CASE: Tuple=0.0 , _SCREAMING_SNAKE_CASE: Any=0.0 , _SCREAMING_SNAKE_CASE: Any=0.1 , _SCREAMING_SNAKE_CASE: str="gelu" , _SCREAMING_SNAKE_CASE: Tuple=False , _SCREAMING_SNAKE_CASE: Union[str, Any]=0.02 , _SCREAMING_SNAKE_CASE: str=1e-5 , _SCREAMING_SNAKE_CASE: Optional[int]=None , _SCREAMING_SNAKE_CASE: str=None , **_SCREAMING_SNAKE_CASE: Union[str, Any] , ) -> List[str]: """simple docstring""" super().__init__(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = image_size __lowerCAmelCase : Any = patch_size __lowerCAmelCase : Tuple = num_channels __lowerCAmelCase : Any = embed_dim __lowerCAmelCase : Any = depths __lowerCAmelCase : Dict = len(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = num_heads __lowerCAmelCase : Tuple = window_size __lowerCAmelCase : Dict = mlp_ratio __lowerCAmelCase : Any = qkv_bias __lowerCAmelCase : Union[str, Any] = hidden_dropout_prob __lowerCAmelCase : int = attention_probs_dropout_prob __lowerCAmelCase : Tuple = drop_path_rate __lowerCAmelCase : int = hidden_act __lowerCAmelCase : Optional[int] = use_absolute_embeddings __lowerCAmelCase : List[str] = layer_norm_eps __lowerCAmelCase : Any = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model __lowerCAmelCase : Optional[Any] = int(embed_dim * 2 ** (len(_SCREAMING_SNAKE_CASE) - 1)) __lowerCAmelCase : Any = ["stem"] + [F"""stage{idx}""" for idx in range(1 , len(_SCREAMING_SNAKE_CASE) + 1)] __lowerCAmelCase , __lowerCAmelCase : List[str] = get_aligned_output_features_output_indices( out_features=_SCREAMING_SNAKE_CASE , out_indices=_SCREAMING_SNAKE_CASE , stage_names=self.stage_names)
269
1
"""simple docstring""" import itertools import json import linecache import os import pickle import re import socket import string from collections import Counter from logging import getLogger from pathlib import Path from typing import Callable, Dict, Iterable, List import git import torch from torch.utils.data import Dataset from transformers import BartTokenizer, RagTokenizer, TaTokenizer def _lowercase ( __snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case=True ,__snake_case="pt" ) -> str: __lowerCAmelCase : List[Any] = {"add_prefix_space": True} if isinstance(__snake_case ,__snake_case ) and not line.startswith(" " ) else {} __lowerCAmelCase : Tuple = padding_side return tokenizer( [line] ,max_length=__snake_case ,padding="max_length" if pad_to_max_length else None ,truncation=__snake_case ,return_tensors=__snake_case ,add_special_tokens=__snake_case ,**__snake_case ,) def _lowercase ( __snake_case ,__snake_case ,__snake_case=None ,) -> Dict: __lowerCAmelCase : List[Any] = input_ids.ne(__snake_case ).any(dim=0 ) if attention_mask is None: return input_ids[:, keep_column_mask] else: return (input_ids[:, keep_column_mask], attention_mask[:, keep_column_mask]) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self: int , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Any="train" , _SCREAMING_SNAKE_CASE: int=None , _SCREAMING_SNAKE_CASE: List[Any]=None , _SCREAMING_SNAKE_CASE: Tuple=None , _SCREAMING_SNAKE_CASE: Optional[Any]="" , ) -> List[str]: """simple docstring""" super().__init__() __lowerCAmelCase : List[str] = Path(_SCREAMING_SNAKE_CASE).joinpath(type_path + ".source") __lowerCAmelCase : Tuple = Path(_SCREAMING_SNAKE_CASE).joinpath(type_path + ".target") __lowerCAmelCase : List[str] = self.get_char_lens(self.src_file) __lowerCAmelCase : List[Any] = max_source_length __lowerCAmelCase : List[Any] = max_target_length assert min(self.src_lens) > 0, F"""found empty line in {self.src_file}""" __lowerCAmelCase : Optional[int] = tokenizer __lowerCAmelCase : Optional[int] = prefix if n_obs is not None: __lowerCAmelCase : Dict = self.src_lens[:n_obs] __lowerCAmelCase : int = src_lang __lowerCAmelCase : List[str] = tgt_lang def __len__( self: List[str]) -> int: """simple docstring""" return len(self.src_lens) def __getitem__( self: str , _SCREAMING_SNAKE_CASE: Optional[Any]) -> Dict[str, torch.Tensor]: """simple docstring""" __lowerCAmelCase : Dict = index + 1 # linecache starts at 1 __lowerCAmelCase : int = self.prefix + linecache.getline(str(self.src_file) , _SCREAMING_SNAKE_CASE).rstrip("\n") __lowerCAmelCase : int = linecache.getline(str(self.tgt_file) , _SCREAMING_SNAKE_CASE).rstrip("\n") assert source_line, F"""empty source line for index {index}""" assert tgt_line, F"""empty tgt line for index {index}""" # Need to add eos token manually for T5 if isinstance(self.tokenizer , _SCREAMING_SNAKE_CASE): source_line += self.tokenizer.eos_token tgt_line += self.tokenizer.eos_token # Pad source and target to the right __lowerCAmelCase : Any = ( self.tokenizer.question_encoder if isinstance(self.tokenizer , _SCREAMING_SNAKE_CASE) else self.tokenizer ) __lowerCAmelCase : Tuple = self.tokenizer.generator if isinstance(self.tokenizer , _SCREAMING_SNAKE_CASE) else self.tokenizer __lowerCAmelCase : Optional[Any] = encode_line(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , self.max_source_length , "right") __lowerCAmelCase : int = encode_line(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , self.max_target_length , "right") __lowerCAmelCase : Tuple = source_inputs["input_ids"].squeeze() __lowerCAmelCase : Dict = target_inputs["input_ids"].squeeze() __lowerCAmelCase : Optional[Any] = source_inputs["attention_mask"].squeeze() return { "input_ids": source_ids, "attention_mask": src_mask, "decoder_input_ids": target_ids, } @staticmethod def _SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE: List[Any]) -> List[str]: """simple docstring""" return [len(_SCREAMING_SNAKE_CASE) for x in Path(_SCREAMING_SNAKE_CASE).open().readlines()] def _SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: str) -> Dict[str, torch.Tensor]: """simple docstring""" __lowerCAmelCase : Optional[int] = torch.stack([x["input_ids"] for x in batch]) __lowerCAmelCase : Dict = torch.stack([x["attention_mask"] for x in batch]) __lowerCAmelCase : List[str] = torch.stack([x["decoder_input_ids"] for x in batch]) __lowerCAmelCase : Tuple = ( self.tokenizer.generator.pad_token_id if isinstance(self.tokenizer , _SCREAMING_SNAKE_CASE) else self.tokenizer.pad_token_id ) __lowerCAmelCase : Tuple = ( self.tokenizer.question_encoder.pad_token_id if isinstance(self.tokenizer , _SCREAMING_SNAKE_CASE) else self.tokenizer.pad_token_id ) __lowerCAmelCase : Union[str, Any] = trim_batch(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase , __lowerCAmelCase : Tuple = trim_batch(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = { "input_ids": source_ids, "attention_mask": source_mask, "decoder_input_ids": y, } return batch __snake_case : Tuple = getLogger(__name__) def _lowercase ( __snake_case ) -> Tuple: return list(itertools.chain.from_iterable(__snake_case ) ) def _lowercase ( __snake_case ) -> None: __lowerCAmelCase : Any = get_git_info() save_json(__snake_case ,os.path.join(__snake_case ,"git_log.json" ) ) def _lowercase ( __snake_case ,__snake_case ,__snake_case=4 ,**__snake_case ) -> Union[str, Any]: with open(__snake_case ,"w" ) as f: json.dump(__snake_case ,__snake_case ,indent=__snake_case ,**__snake_case ) def _lowercase ( __snake_case ) -> List[Any]: with open(__snake_case ) as f: return json.load(__snake_case ) def _lowercase ( ) -> Optional[Any]: __lowerCAmelCase : Dict = git.Repo(search_parent_directories=__snake_case ) __lowerCAmelCase : Union[str, Any] = { "repo_id": str(__snake_case ), "repo_sha": str(repo.head.object.hexsha ), "repo_branch": str(repo.active_branch ), "hostname": str(socket.gethostname() ), } return repo_infos def _lowercase ( __snake_case ,__snake_case ) -> List: return list(map(__snake_case ,__snake_case ) ) def _lowercase ( __snake_case ,__snake_case ) -> Optional[Any]: with open(__snake_case ,"wb" ) as f: return pickle.dump(__snake_case ,__snake_case ) def _lowercase ( __snake_case ) -> Union[str, Any]: def remove_articles(__snake_case ): return re.sub(r"\b(a|an|the)\b" ," " ,__snake_case ) def white_space_fix(__snake_case ): return " ".join(text.split() ) def remove_punc(__snake_case ): __lowerCAmelCase : Dict = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(__snake_case ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(__snake_case ) ) ) ) def _lowercase ( __snake_case ,__snake_case ) -> str: __lowerCAmelCase : Tuple = normalize_answer(__snake_case ).split() __lowerCAmelCase : List[str] = normalize_answer(__snake_case ).split() __lowerCAmelCase : Union[str, Any] = Counter(__snake_case ) & Counter(__snake_case ) __lowerCAmelCase : Union[str, Any] = sum(common.values() ) if num_same == 0: return 0 __lowerCAmelCase : str = 1.0 * num_same / len(__snake_case ) __lowerCAmelCase : List[Any] = 1.0 * num_same / len(__snake_case ) __lowerCAmelCase : Optional[Any] = (2 * precision * recall) / (precision + recall) return fa def _lowercase ( __snake_case ,__snake_case ) -> Dict: return normalize_answer(__snake_case ) == normalize_answer(__snake_case ) def _lowercase ( __snake_case ,__snake_case ) -> Dict: assert len(__snake_case ) == len(__snake_case ) __lowerCAmelCase : Optional[int] = 0 for hypo, pred in zip(__snake_case ,__snake_case ): em += exact_match_score(__snake_case ,__snake_case ) if len(__snake_case ) > 0: em /= len(__snake_case ) return {"em": em} def _lowercase ( __snake_case ) -> Tuple: return model_prefix.startswith("rag" ) def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> Optional[Any]: __lowerCAmelCase : List[Any] = {p: p for p in extra_params} # T5 models don't have `dropout` param, they have `dropout_rate` instead __lowerCAmelCase : List[str] = "dropout_rate" for p in extra_params: if getattr(__snake_case ,__snake_case ,__snake_case ): if not hasattr(__snake_case ,__snake_case ) and not hasattr(__snake_case ,equivalent_param[p] ): logger.info("config doesn't have a `{}` attribute".format(__snake_case ) ) delattr(__snake_case ,__snake_case ) continue __lowerCAmelCase : List[Any] = p if hasattr(__snake_case ,__snake_case ) else equivalent_param[p] setattr(__snake_case ,__snake_case ,getattr(__snake_case ,__snake_case ) ) delattr(__snake_case ,__snake_case ) return hparams, config
269
"""simple docstring""" import gc import threading import time import psutil import torch class A__ : '''simple docstring''' def __init__( self: str) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Optional[Any] = psutil.Process() __lowerCAmelCase : str = False def _SCREAMING_SNAKE_CASE ( self: int) -> Optional[int]: """simple docstring""" __lowerCAmelCase : Optional[Any] = -1 while True: __lowerCAmelCase : str = max(self.process.memory_info().rss , self.cpu_memory_peak) # can't sleep or will not catch the peak right (this comment is here on purpose) if not self.peak_monitoring: break def _SCREAMING_SNAKE_CASE ( self: Tuple) -> int: """simple docstring""" __lowerCAmelCase : List[str] = True __lowerCAmelCase : str = threading.Thread(target=self.peak_monitor) __lowerCAmelCase : Tuple = True self.thread.start() def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> List[str]: """simple docstring""" __lowerCAmelCase : Tuple = False self.thread.join() return self.cpu_memory_peak __snake_case : Tuple = PeakCPUMemory() def _lowercase ( ) -> str: # Time __lowerCAmelCase : str = {"time": time.time()} gc.collect() torch.cuda.empty_cache() # CPU mem __lowerCAmelCase : Optional[Any] = psutil.Process().memory_info().rss cpu_peak_tracker.start() # GPU mem for i in range(torch.cuda.device_count() ): __lowerCAmelCase : Union[str, Any] = torch.cuda.memory_allocated(__snake_case ) torch.cuda.reset_peak_memory_stats() return measures def _lowercase ( __snake_case ) -> Optional[Any]: # Time __lowerCAmelCase : str = {"time": time.time() - start_measures["time"]} gc.collect() torch.cuda.empty_cache() # CPU mem __lowerCAmelCase : str = (psutil.Process().memory_info().rss - start_measures["cpu"]) / 2**20 __lowerCAmelCase : List[str] = (cpu_peak_tracker.stop() - start_measures["cpu"]) / 2**20 # GPU mem for i in range(torch.cuda.device_count() ): __lowerCAmelCase : Union[str, Any] = (torch.cuda.memory_allocated(__snake_case ) - start_measures[str(__snake_case )]) / 2**20 __lowerCAmelCase : Any = (torch.cuda.max_memory_allocated(__snake_case ) - start_measures[str(__snake_case )]) / 2**20 return measures def _lowercase ( __snake_case ,__snake_case ) -> Dict: print(F"""{description}:""" ) print(F"""- Time: {measures['time']:.2f}s""" ) for i in range(torch.cuda.device_count() ): print(F"""- GPU {i} allocated: {measures[str(__snake_case )]:.2f}MiB""" ) __lowerCAmelCase : Optional[Any] = measures[F"""{i}-peak"""] print(F"""- GPU {i} peak: {peak:.2f}MiB""" ) print(F"""- CPU RAM allocated: {measures['cpu']:.2f}MiB""" ) print(F"""- CPU RAM peak: {measures['cpu-peak']:.2f}MiB""" )
269
1
"""simple docstring""" from queue import PriorityQueue from typing import Any import numpy as np def _lowercase ( __snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ,) -> float | int: for nxt, d in graph[v]: if nxt in visited_forward: continue __lowerCAmelCase : Optional[Any] = cst_fwd.get(__snake_case ,np.inf ) __lowerCAmelCase : str = cst_fwd[v] + d if new_cost_f < old_cost_f: queue.put((new_cost_f, nxt) ) __lowerCAmelCase : List[str] = new_cost_f __lowerCAmelCase : Optional[Any] = v if nxt in visited_backward: if cst_fwd[v] + d + cst_bwd[nxt] < shortest_distance: __lowerCAmelCase : str = cst_fwd[v] + d + cst_bwd[nxt] return shortest_distance def _lowercase ( __snake_case ,__snake_case ,__snake_case ,__snake_case ) -> int: __lowerCAmelCase : Any = -1 __lowerCAmelCase : str = set() __lowerCAmelCase : Optional[Any] = set() __lowerCAmelCase : int = {source: 0} __lowerCAmelCase : Tuple = {destination: 0} __lowerCAmelCase : Union[str, Any] = {source: None} __lowerCAmelCase : Dict = {destination: None} __lowerCAmelCase : PriorityQueue[Any] = PriorityQueue() __lowerCAmelCase : PriorityQueue[Any] = PriorityQueue() __lowerCAmelCase : Tuple = np.inf queue_forward.put((0, source) ) queue_backward.put((0, destination) ) if source == destination: return 0 while not queue_forward.empty() and not queue_backward.empty(): __lowerCAmelCase , __lowerCAmelCase : List[Any] = queue_forward.get() visited_forward.add(__snake_case ) __lowerCAmelCase , __lowerCAmelCase : str = queue_backward.get() visited_backward.add(__snake_case ) __lowerCAmelCase : Optional[Any] = pass_and_relaxation( __snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ,) __lowerCAmelCase : Optional[Any] = pass_and_relaxation( __snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ,) if cst_fwd[v_fwd] + cst_bwd[v_bwd] >= shortest_distance: break if shortest_distance != np.inf: __lowerCAmelCase : Any = shortest_distance return shortest_path_distance __snake_case : int = { 'B': [['C', 1]], 'C': [['D', 1]], 'D': [['F', 1]], 'E': [['B', 1], ['G', 2]], 'F': [], 'G': [['F', 1]], } __snake_case : List[Any] = { 'B': [['E', 1]], 'C': [['B', 1]], 'D': [['C', 1]], 'F': [['D', 1], ['G', 1]], 'E': [[None, np.inf]], 'G': [['E', 2]], } if __name__ == "__main__": import doctest doctest.testmod()
269
"""simple docstring""" from ....configuration_utils import PretrainedConfig from ....utils import logging __snake_case : List[Any] = logging.get_logger(__name__) __snake_case : Any = { 'speechbrain/m-ctc-t-large': 'https://huggingface.co/speechbrain/m-ctc-t-large/resolve/main/config.json', # See all M-CTC-T models at https://huggingface.co/models?filter=mctct } class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'mctct' def __init__( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: str=8065 , _SCREAMING_SNAKE_CASE: str=1536 , _SCREAMING_SNAKE_CASE: str=36 , _SCREAMING_SNAKE_CASE: Optional[Any]=6144 , _SCREAMING_SNAKE_CASE: Optional[Any]=4 , _SCREAMING_SNAKE_CASE: Union[str, Any]=384 , _SCREAMING_SNAKE_CASE: Optional[Any]=920 , _SCREAMING_SNAKE_CASE: Union[str, Any]=1e-5 , _SCREAMING_SNAKE_CASE: List[Any]=0.3 , _SCREAMING_SNAKE_CASE: Optional[Any]="relu" , _SCREAMING_SNAKE_CASE: Optional[int]=0.02 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.3 , _SCREAMING_SNAKE_CASE: Dict=0.3 , _SCREAMING_SNAKE_CASE: List[Any]=1 , _SCREAMING_SNAKE_CASE: Optional[Any]=0 , _SCREAMING_SNAKE_CASE: List[str]=2 , _SCREAMING_SNAKE_CASE: Union[str, Any]=1 , _SCREAMING_SNAKE_CASE: Tuple=0.3 , _SCREAMING_SNAKE_CASE: Dict=1 , _SCREAMING_SNAKE_CASE: int=(7,) , _SCREAMING_SNAKE_CASE: str=(3,) , _SCREAMING_SNAKE_CASE: Union[str, Any]=80 , _SCREAMING_SNAKE_CASE: Tuple=1 , _SCREAMING_SNAKE_CASE: Dict=None , _SCREAMING_SNAKE_CASE: Tuple="sum" , _SCREAMING_SNAKE_CASE: List[str]=False , **_SCREAMING_SNAKE_CASE: Tuple , ) -> Tuple: """simple docstring""" super().__init__(**_SCREAMING_SNAKE_CASE , pad_token_id=_SCREAMING_SNAKE_CASE , bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = vocab_size __lowerCAmelCase : str = hidden_size __lowerCAmelCase : str = num_hidden_layers __lowerCAmelCase : str = intermediate_size __lowerCAmelCase : List[Any] = num_attention_heads __lowerCAmelCase : Dict = attention_head_dim __lowerCAmelCase : Optional[int] = max_position_embeddings __lowerCAmelCase : str = layer_norm_eps __lowerCAmelCase : Tuple = layerdrop __lowerCAmelCase : str = hidden_act __lowerCAmelCase : List[Any] = initializer_range __lowerCAmelCase : int = hidden_dropout_prob __lowerCAmelCase : Optional[Any] = attention_probs_dropout_prob __lowerCAmelCase : str = pad_token_id __lowerCAmelCase : Optional[int] = bos_token_id __lowerCAmelCase : Union[str, Any] = eos_token_id __lowerCAmelCase : Any = conv_glu_dim __lowerCAmelCase : Optional[int] = conv_dropout __lowerCAmelCase : Union[str, Any] = num_conv_layers __lowerCAmelCase : Optional[int] = input_feat_per_channel __lowerCAmelCase : Union[str, Any] = input_channels __lowerCAmelCase : Optional[Any] = conv_channels __lowerCAmelCase : Dict = ctc_loss_reduction __lowerCAmelCase : int = ctc_zero_infinity # prevents config testing fail with exporting to json __lowerCAmelCase : List[str] = list(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = list(_SCREAMING_SNAKE_CASE) if len(self.conv_kernel) != self.num_conv_layers: raise ValueError( "Configuration for convolutional module is incorrect. " "It is required that `len(config.conv_kernel)` == `config.num_conv_layers` " F"""but is `len(config.conv_kernel) = {len(self.conv_kernel)}`, """ F"""`config.num_conv_layers = {self.num_conv_layers}`.""")
269
1
"""simple docstring""" from collections.abc import Iterator, MutableMapping from dataclasses import dataclass from typing import Generic, TypeVar __snake_case : Any = TypeVar('KEY') __snake_case : int = TypeVar('VAL') @dataclass(frozen=__SCREAMING_SNAKE_CASE , slots=__SCREAMING_SNAKE_CASE ) class A__ ( Generic[KEY, VAL] ): '''simple docstring''' SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = 42 class A__ ( _Item ): '''simple docstring''' def __init__( self: Any) -> None: """simple docstring""" super().__init__(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) def __bool__( self: Any) -> bool: """simple docstring""" return False __snake_case : List[str] = _DeletedItem() class A__ ( MutableMapping[KEY, VAL] ): '''simple docstring''' def __init__( self: Any , _SCREAMING_SNAKE_CASE: int = 8 , _SCREAMING_SNAKE_CASE: float = 0.75) -> None: """simple docstring""" __lowerCAmelCase : List[str] = initial_block_size __lowerCAmelCase : list[_Item | None] = [None] * initial_block_size assert 0.0 < capacity_factor < 1.0 __lowerCAmelCase : int = capacity_factor __lowerCAmelCase : Any = 0 def _SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: KEY) -> int: """simple docstring""" return hash(_SCREAMING_SNAKE_CASE) % len(self._buckets) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: int) -> int: """simple docstring""" return (ind + 1) % len(self._buckets) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: KEY , _SCREAMING_SNAKE_CASE: VAL) -> bool: """simple docstring""" __lowerCAmelCase : Tuple = self._buckets[ind] if not stored: __lowerCAmelCase : List[Any] = _Item(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) self._len += 1 return True elif stored.key == key: __lowerCAmelCase : List[str] = _Item(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) return True else: return False def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> bool: """simple docstring""" __lowerCAmelCase : str = len(self._buckets) * self._capacity_factor return len(self) >= int(_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Tuple) -> bool: """simple docstring""" if len(self._buckets) <= self._initial_block_size: return False __lowerCAmelCase : List[Any] = len(self._buckets) * self._capacity_factor / 2 return len(self) < limit def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: int) -> None: """simple docstring""" __lowerCAmelCase : Any = self._buckets __lowerCAmelCase : List[str] = [None] * new_size __lowerCAmelCase : int = 0 for item in old_buckets: if item: self._add_item(item.key , item.val) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> None: """simple docstring""" self._resize(len(self._buckets) * 2) def _SCREAMING_SNAKE_CASE ( self: List[str]) -> None: """simple docstring""" self._resize(len(self._buckets) // 2) def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: KEY) -> Iterator[int]: """simple docstring""" __lowerCAmelCase : Optional[int] = self._get_bucket_index(_SCREAMING_SNAKE_CASE) for _ in range(len(self._buckets)): yield ind __lowerCAmelCase : Tuple = self._get_next_ind(_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: str , _SCREAMING_SNAKE_CASE: KEY , _SCREAMING_SNAKE_CASE: VAL) -> None: """simple docstring""" for ind in self._iterate_buckets(_SCREAMING_SNAKE_CASE): if self._try_set(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE): break def __setitem__( self: str , _SCREAMING_SNAKE_CASE: KEY , _SCREAMING_SNAKE_CASE: VAL) -> None: """simple docstring""" if self._is_full(): self._size_up() self._add_item(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) def __delitem__( self: Optional[int] , _SCREAMING_SNAKE_CASE: KEY) -> None: """simple docstring""" for ind in self._iterate_buckets(_SCREAMING_SNAKE_CASE): __lowerCAmelCase : List[Any] = self._buckets[ind] if item is None: raise KeyError(_SCREAMING_SNAKE_CASE) if item is _deleted: continue if item.key == key: __lowerCAmelCase : Dict = _deleted self._len -= 1 break if self._is_sparse(): self._size_down() def __getitem__( self: List[str] , _SCREAMING_SNAKE_CASE: KEY) -> VAL: """simple docstring""" for ind in self._iterate_buckets(_SCREAMING_SNAKE_CASE): __lowerCAmelCase : str = self._buckets[ind] if item is None: break if item is _deleted: continue if item.key == key: return item.val raise KeyError(_SCREAMING_SNAKE_CASE) def __len__( self: int) -> int: """simple docstring""" return self._len def __iter__( self: Optional[Any]) -> Iterator[KEY]: """simple docstring""" yield from (item.key for item in self._buckets if item) def __repr__( self: Optional[Any]) -> str: """simple docstring""" __lowerCAmelCase : Tuple = " ,".join( F"""{item.key}: {item.val}""" for item in self._buckets if item) return F"""HashMap({val_string})"""
269
"""simple docstring""" from __future__ import annotations import time import numpy as np __snake_case : Optional[Any] = [8, 5, 9, 7] __snake_case : List[Any] = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] __snake_case : Optional[int] = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class A__ : '''simple docstring''' def __init__( self: Any , _SCREAMING_SNAKE_CASE: list[int] , _SCREAMING_SNAKE_CASE: list[list[int]] , _SCREAMING_SNAKE_CASE: list[list[int]] , ) -> None: """simple docstring""" __lowerCAmelCase : Any = claim_vector __lowerCAmelCase : Tuple = allocated_resources_table __lowerCAmelCase : Tuple = maximum_claim_table def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> list[int]: """simple docstring""" return [ sum(p_item[i] for p_item in self.__allocated_resources_table) for i in range(len(self.__allocated_resources_table[0])) ] def _SCREAMING_SNAKE_CASE ( self: int) -> list[int]: """simple docstring""" return np.array(self.__claim_vector) - np.array( self.__processes_resource_summation()) def _SCREAMING_SNAKE_CASE ( self: int) -> list[list[int]]: """simple docstring""" return [ list(np.array(self.__maximum_claim_table[i]) - np.array(_SCREAMING_SNAKE_CASE)) for i, allocated_resource in enumerate(self.__allocated_resources_table) ] def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> dict[int, list[int]]: """simple docstring""" return {self.__need().index(_SCREAMING_SNAKE_CASE): i for i in self.__need()} def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , **_SCREAMING_SNAKE_CASE: List[Any]) -> None: """simple docstring""" __lowerCAmelCase : Optional[int] = self.__need() __lowerCAmelCase : int = self.__allocated_resources_table __lowerCAmelCase : Dict = self.__available_resources() __lowerCAmelCase : str = self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print("_" * 50 + "\n") while need_list: __lowerCAmelCase : int = False for each_need in need_list: __lowerCAmelCase : Dict = True for index, need in enumerate(_SCREAMING_SNAKE_CASE): if need > available_resources[index]: __lowerCAmelCase : Dict = False break if execution: __lowerCAmelCase : Any = True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: __lowerCAmelCase : Union[str, Any] = original_need_index print(F"""Process {process_number + 1} is executing.""") # remove the process run from stack need_list.remove(_SCREAMING_SNAKE_CASE) # update available/freed resources stack __lowerCAmelCase : Dict = np.array(_SCREAMING_SNAKE_CASE) + np.array( alloc_resources_table[process_number]) print( "Updated available resource stack for processes: " + " ".join([str(_SCREAMING_SNAKE_CASE) for x in available_resources])) break if safe: print("The process is in a safe state.\n") else: print("System in unsafe state. Aborting...\n") break def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> List[Any]: """simple docstring""" print(" " * 9 + "Allocated Resource Table") for item in self.__allocated_resources_table: print( F"""P{self.__allocated_resources_table.index(_SCREAMING_SNAKE_CASE) + 1}""" + " ".join(F"""{it:>8}""" for it in item) + "\n") print(" " * 9 + "System Resource Table") for item in self.__maximum_claim_table: print( F"""P{self.__maximum_claim_table.index(_SCREAMING_SNAKE_CASE) + 1}""" + " ".join(F"""{it:>8}""" for it in item) + "\n") print( "Current Usage by Active Processes: " + " ".join(str(_SCREAMING_SNAKE_CASE) for x in self.__claim_vector)) print( "Initial Available Resources: " + " ".join(str(_SCREAMING_SNAKE_CASE) for x in self.__available_resources())) time.sleep(1) if __name__ == "__main__": import doctest doctest.testmod()
269
1
"""simple docstring""" from __future__ import annotations import unittest from transformers import is_tf_available, is_torch_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, is_pt_tf_cross_test, slow if is_tf_available(): from transformers import ( AutoConfig, BertConfig, GPTaConfig, TaConfig, TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST if is_torch_available(): from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelWithLMHead, BertForMaskedLM, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertModel, GPTaLMHeadModel, RobertaForMaskedLM, TaForConditionalGeneration, ) @is_pt_tf_cross_test class A__ ( unittest.TestCase ): '''simple docstring''' @slow def _SCREAMING_SNAKE_CASE ( self: str) -> Dict: """simple docstring""" for model_name in ["bert-base-uncased"]: __lowerCAmelCase : Optional[int] = AutoConfig.from_pretrained(_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = TFAutoModel.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = AutoModel.from_pretrained(_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) @slow def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> Tuple: """simple docstring""" for model_name in ["bert-base-uncased"]: __lowerCAmelCase : Union[str, Any] = AutoConfig.from_pretrained(_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = TFAutoModelForPreTraining.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = AutoModelForPreTraining.from_pretrained(_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) @slow def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> Dict: """simple docstring""" for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase : Union[str, Any] = AutoConfig.from_pretrained(_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = TFAutoModelForCausalLM.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE) __lowerCAmelCase , __lowerCAmelCase : str = TFAutoModelForCausalLM.from_pretrained( _SCREAMING_SNAKE_CASE , output_loading_info=_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = AutoModelForCausalLM.from_pretrained(_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE) __lowerCAmelCase , __lowerCAmelCase : int = AutoModelForCausalLM.from_pretrained( _SCREAMING_SNAKE_CASE , output_loading_info=_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) @slow def _SCREAMING_SNAKE_CASE ( self: Any) -> Optional[Any]: """simple docstring""" for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase : str = AutoConfig.from_pretrained(_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = TFAutoModelWithLMHead.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = AutoModelWithLMHead.from_pretrained(_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) @slow def _SCREAMING_SNAKE_CASE ( self: str) -> Any: """simple docstring""" for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase : Dict = AutoConfig.from_pretrained(_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = TFAutoModelForMaskedLM.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE) __lowerCAmelCase , __lowerCAmelCase : Dict = TFAutoModelForMaskedLM.from_pretrained( _SCREAMING_SNAKE_CASE , output_loading_info=_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = AutoModelForMaskedLM.from_pretrained(_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE) __lowerCAmelCase , __lowerCAmelCase : Any = AutoModelForMaskedLM.from_pretrained( _SCREAMING_SNAKE_CASE , output_loading_info=_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) @slow def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> List[Any]: """simple docstring""" for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase : Union[str, Any] = AutoConfig.from_pretrained(_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = TFAutoModelForSeqaSeqLM.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE) __lowerCAmelCase , __lowerCAmelCase : Optional[Any] = TFAutoModelForSeqaSeqLM.from_pretrained( _SCREAMING_SNAKE_CASE , output_loading_info=_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = AutoModelForSeqaSeqLM.from_pretrained(_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE) __lowerCAmelCase , __lowerCAmelCase : str = AutoModelForSeqaSeqLM.from_pretrained( _SCREAMING_SNAKE_CASE , output_loading_info=_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) @slow def _SCREAMING_SNAKE_CASE ( self: int) -> int: """simple docstring""" for model_name in ["bert-base-uncased"]: __lowerCAmelCase : Tuple = AutoConfig.from_pretrained(_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = TFAutoModelForSequenceClassification.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = AutoModelForSequenceClassification.from_pretrained(_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) @slow def _SCREAMING_SNAKE_CASE ( self: str) -> int: """simple docstring""" for model_name in ["bert-base-uncased"]: __lowerCAmelCase : Optional[int] = AutoConfig.from_pretrained(_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = TFAutoModelForQuestionAnswering.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = AutoModelForQuestionAnswering.from_pretrained(_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Dict) -> List[str]: """simple docstring""" __lowerCAmelCase : Dict = TFAutoModelWithLMHead.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) self.assertEqual(model.num_parameters() , 1_4410) self.assertEqual(model.num_parameters(only_trainable=_SCREAMING_SNAKE_CASE) , 1_4410) __lowerCAmelCase : Union[str, Any] = AutoModelWithLMHead.from_pretrained(_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) self.assertEqual(model.num_parameters() , 1_4410) self.assertEqual(model.num_parameters(only_trainable=_SCREAMING_SNAKE_CASE) , 1_4410) def _SCREAMING_SNAKE_CASE ( self: str) -> Dict: """simple docstring""" __lowerCAmelCase : Any = TFAutoModelWithLMHead.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) self.assertEqual(model.num_parameters() , 1_4410) self.assertEqual(model.num_parameters(only_trainable=_SCREAMING_SNAKE_CASE) , 1_4410) __lowerCAmelCase : Optional[Any] = AutoModelWithLMHead.from_pretrained(_SCREAMING_SNAKE_CASE , from_tf=_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) self.assertEqual(model.num_parameters() , 1_4410) self.assertEqual(model.num_parameters(only_trainable=_SCREAMING_SNAKE_CASE) , 1_4410)
269
"""simple docstring""" # # This a `torch.distributed` diagnostics script that checks that all GPUs in the cluster (one or # many nodes) can talk to each other via nccl and allocate gpu memory. # # To run first adjust the number of processes and nodes: # # python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # You may need to add --master_addr $MASTER_ADDR --master_port $MASTER_PORT if using a custom addr:port # # You can also use the rdzv API: --rdzv_endpoint $MASTER_ADDR:$MASTER_PORT --rdzv_backend c10d # # use torch.distributed.launch instead of torch.distributed.run for torch < 1.9 # # If you get a hanging in `barrier` calls you have some network issues, you may try to debug this with: # # NCCL_DEBUG=INFO python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # which should tell you what's going on behind the scenes. # # # This script can be run via `srun` in the SLURM environment as well. Here is a SLURM script that # runs on 2 nodes of 4 gpus per node: # # #SBATCH --job-name=test-nodes # name # #SBATCH --nodes=2 # nodes # #SBATCH --ntasks-per-node=1 # crucial - only 1 task per dist per node! # #SBATCH --cpus-per-task=10 # number of cores per tasks # #SBATCH --gres=gpu:4 # number of gpus # #SBATCH --time 0:05:00 # maximum execution time (HH:MM:SS) # #SBATCH --output=%x-%j.out # output file name # # GPUS_PER_NODE=4 # MASTER_ADDR=$(scontrol show hostnames $SLURM_JOB_NODELIST | head -n 1) # MASTER_PORT=6000 # # srun --jobid $SLURM_JOBID bash -c 'python -m torch.distributed.run \ # --nproc_per_node $GPUS_PER_NODE --nnodes $SLURM_NNODES --node_rank $SLURM_PROCID \ # --master_addr $MASTER_ADDR --master_port $MASTER_PORT \ # torch-distributed-gpu-test.py' # import fcntl import os import socket import torch import torch.distributed as dist def _lowercase ( *__snake_case ) -> Optional[Any]: with open(__snake_case ,"r" ) as fh: fcntl.flock(__snake_case ,fcntl.LOCK_EX ) try: print(*__snake_case ) finally: fcntl.flock(__snake_case ,fcntl.LOCK_UN ) __snake_case : List[Any] = int(os.environ['LOCAL_RANK']) torch.cuda.set_device(local_rank) __snake_case : List[str] = torch.device('cuda', local_rank) __snake_case : Optional[Any] = socket.gethostname() __snake_case : str = F"""[{hostname}-{local_rank}]""" try: # test distributed dist.init_process_group('nccl') dist.all_reduce(torch.ones(1).to(device), op=dist.ReduceOp.SUM) dist.barrier() # test cuda is available and can allocate memory torch.cuda.is_available() torch.ones(1).cuda(local_rank) # global rank __snake_case : Tuple = dist.get_rank() __snake_case : Any = dist.get_world_size() printflock(F"""{gpu} is OK (global rank: {rank}/{world_size})""") dist.barrier() if rank == 0: printflock(F"""pt={torch.__version__}, cuda={torch.version.cuda}, nccl={torch.cuda.nccl.version()}""") except Exception: printflock(F"""{gpu} is broken""") raise
269
1
"""simple docstring""" import numpy as np import datasets __snake_case : str = '\nCompute the Mahalanobis Distance\n\nMahalonobis distance is the distance between a point and a distribution.\nAnd not between two distinct points. It is effectively a multivariate equivalent of the Euclidean distance.\nIt was introduced by Prof. P. C. Mahalanobis in 1936\nand has been used in various statistical applications ever since\n[source: https://www.machinelearningplus.com/statistics/mahalanobis-distance/]\n' __snake_case : Optional[Any] = '\\n@article{de2000mahalanobis,\n title={The mahalanobis distance},\n author={De Maesschalck, Roy and Jouan-Rimbaud, Delphine and Massart, D{\'e}sir{\'e} L},\n journal={Chemometrics and intelligent laboratory systems},\n volume={50},\n number={1},\n pages={1--18},\n year={2000},\n publisher={Elsevier}\n}\n' __snake_case : List[Any] = '\nArgs:\n X: List of datapoints to be compared with the `reference_distribution`.\n reference_distribution: List of datapoints from the reference distribution we want to compare to.\nReturns:\n mahalanobis: The Mahalonobis distance for each datapoint in `X`.\nExamples:\n\n >>> mahalanobis_metric = datasets.load_metric("mahalanobis")\n >>> results = mahalanobis_metric.compute(reference_distribution=[[0, 1], [1, 0]], X=[[0, 1]])\n >>> print(results)\n {\'mahalanobis\': array([0.5])}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A__ ( datasets.Metric ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Any) -> Any: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "X": datasets.Sequence(datasets.Value("float" , id="sequence") , id="X"), }) , ) def _SCREAMING_SNAKE_CASE ( self: Optional[int] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Any) -> Dict: """simple docstring""" __lowerCAmelCase : Any = np.array(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = np.array(_SCREAMING_SNAKE_CASE) # Assert that arrays are 2D if len(X.shape) != 2: raise ValueError("Expected `X` to be a 2D vector") if len(reference_distribution.shape) != 2: raise ValueError("Expected `reference_distribution` to be a 2D vector") if reference_distribution.shape[0] < 2: raise ValueError( "Expected `reference_distribution` to be a 2D vector with more than one element in the first dimension") # Get mahalanobis distance for each prediction __lowerCAmelCase : Any = X - np.mean(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = np.cov(reference_distribution.T) try: __lowerCAmelCase : Union[str, Any] = np.linalg.inv(_SCREAMING_SNAKE_CASE) except np.linalg.LinAlgError: __lowerCAmelCase : Optional[int] = np.linalg.pinv(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = np.dot(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = np.dot(_SCREAMING_SNAKE_CASE , X_minus_mu.T).diagonal() return {"mahalanobis": mahal_dist}
269
"""simple docstring""" import numpy as np from scipy.spatial.distance import cdist from sklearn.metrics import fa_score import datasets __snake_case : Optional[int] = '\\n @inproceedings{kakwani2020indicnlpsuite,\n title={{IndicNLPSuite: Monolingual Corpora, Evaluation Benchmarks and Pre-trained Multilingual Language Models for Indian Languages}},\n author={Divyanshu Kakwani and Anoop Kunchukuttan and Satish Golla and Gokul N.C. and Avik Bhattacharyya and Mitesh M. Khapra and Pratyush Kumar},\n year={2020},\n booktitle={Findings of EMNLP},\n}\n' __snake_case : str = '\\n IndicGLUE is a natural language understanding benchmark for Indian languages. It contains a wide\n variety of tasks and covers 11 major Indian languages - as, bn, gu, hi, kn, ml, mr, or, pa, ta, te.\n' __snake_case : str = '\nCompute IndicGLUE evaluation metric associated to each IndicGLUE dataset.\nArgs:\n predictions: list of predictions to score (as int64),\n except for \'cvit-mkb-clsr\' where each prediction is a vector (of float32).\n references: list of ground truth labels corresponding to the predictions (as int64),\n except for \'cvit-mkb-clsr\' where each reference is a vector (of float32).\nReturns: depending on the IndicGLUE subset, one or several of:\n "accuracy": Accuracy\n "f1": F1 score\n "precision": Precision@10\nExamples:\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wnli\') # \'wnli\' or any of ["copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md"]\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0}\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wiki-ner\')\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0, \'f1\': 1.0}\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'cvit-mkb-clsr\')\n >>> references = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]]\n >>> predictions = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'precision@10\': 1.0}\n\n' def _lowercase ( __snake_case ,__snake_case ) -> Union[str, Any]: return float((preds == labels).mean() ) def _lowercase ( __snake_case ,__snake_case ) -> str: __lowerCAmelCase : str = simple_accuracy(__snake_case ,__snake_case ) __lowerCAmelCase : Any = float(fa_score(y_true=__snake_case ,y_pred=__snake_case ) ) return { "accuracy": acc, "f1": fa, } def _lowercase ( __snake_case ,__snake_case ) -> int: __lowerCAmelCase : Union[str, Any] = np.array(__snake_case ) __lowerCAmelCase : Tuple = np.array(__snake_case ) __lowerCAmelCase : List[Any] = en_sentvecs.shape[0] # mean centering __lowerCAmelCase : Union[str, Any] = en_sentvecs - np.mean(__snake_case ,axis=0 ) __lowerCAmelCase : int = in_sentvecs - np.mean(__snake_case ,axis=0 ) __lowerCAmelCase : Optional[Any] = cdist(__snake_case ,__snake_case ,"cosine" ) __lowerCAmelCase : int = np.array(range(__snake_case ) ) __lowerCAmelCase : int = sim.argsort(axis=1 )[:, :10] __lowerCAmelCase : Optional[Any] = np.any(preds == actual[:, None] ,axis=1 ) return float(matches.mean() ) @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A__ ( datasets.Metric ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: int) -> str: """simple docstring""" if self.config_name not in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", "wiki-ner", ]: raise KeyError( "You should supply a configuration name selected in " "[\"wnli\", \"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", " "\"cvit-mkb-clsr\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\", " "\"wiki-ner\"]") return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("int64") if self.config_name != "cvit-mkb-clsr" else datasets.Sequence(datasets.Value("float32")), "references": datasets.Value("int64") if self.config_name != "cvit-mkb-clsr" else datasets.Sequence(datasets.Value("float32")), }) , codebase_urls=[] , reference_urls=[] , format="numpy" if self.config_name != "cvit-mkb-clsr" else None , ) def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Optional[Any]) -> int: """simple docstring""" if self.config_name == "cvit-mkb-clsr": return {"precision@10": precision_at_aa(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE)} elif self.config_name in ["wiki-ner"]: return acc_and_fa(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) elif self.config_name in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md", ]: return {"accuracy": simple_accuracy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE)} else: raise KeyError( "You should supply a configuration name selected in " "[\"wnli\", \"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", " "\"cvit-mkb-clsr\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\", " "\"wiki-ner\"]")
269
1
"""simple docstring""" import os import unittest from huggingface_hub.utils import are_progress_bars_disabled import transformers.models.bart.tokenization_bart from transformers import logging from transformers.testing_utils import CaptureLogger, mockenv, mockenv_context from transformers.utils.logging import disable_progress_bar, enable_progress_bar class A__ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: int) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : Dict = logging.get_logger() # the current default level is logging.WARNING __lowerCAmelCase : Any = logging.get_verbosity() logging.set_verbosity_error() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity()) logging.set_verbosity_warning() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity()) logging.set_verbosity_info() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity()) logging.set_verbosity_debug() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity()) # restore to the original level logging.set_verbosity(_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: int) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : Dict = logging.get_verbosity() __lowerCAmelCase : Union[str, Any] = logging.get_logger("transformers.models.bart.tokenization_bart") __lowerCAmelCase : Tuple = "Testing 1, 2, 3" # should be able to log warnings (if default settings weren't overridden by `pytest --log-level-all`) if level_origin <= logging.WARNING: with CaptureLogger(_SCREAMING_SNAKE_CASE) as cl: logger.warning(_SCREAMING_SNAKE_CASE) self.assertEqual(cl.out , msg + "\n") # this is setting the level for all of `transformers.*` loggers logging.set_verbosity_error() # should not be able to log warnings with CaptureLogger(_SCREAMING_SNAKE_CASE) as cl: logger.warning(_SCREAMING_SNAKE_CASE) self.assertEqual(cl.out , "") # should be able to log warnings again logging.set_verbosity_warning() with CaptureLogger(_SCREAMING_SNAKE_CASE) as cl: logger.warning(_SCREAMING_SNAKE_CASE) self.assertEqual(cl.out , msg + "\n") # restore to the original level logging.set_verbosity(_SCREAMING_SNAKE_CASE) @mockenv(TRANSFORMERS_VERBOSITY="error") def _SCREAMING_SNAKE_CASE ( self: str) -> List[Any]: """simple docstring""" transformers.utils.logging._reset_library_root_logger() # this action activates the env var __lowerCAmelCase : List[str] = logging.get_logger("transformers.models.bart.tokenization_bart") __lowerCAmelCase : List[str] = os.getenv("TRANSFORMERS_VERBOSITY" , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = logging.log_levels[env_level_str] __lowerCAmelCase : Optional[Any] = logging.get_verbosity() self.assertEqual( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , F"""TRANSFORMERS_VERBOSITY={env_level_str}/{env_level}, but internal verbosity is {current_level}""" , ) # restore to the original level __lowerCAmelCase : Union[str, Any] = "" transformers.utils.logging._reset_library_root_logger() @mockenv(TRANSFORMERS_VERBOSITY="super-error") def _SCREAMING_SNAKE_CASE ( self: int) -> List[Any]: """simple docstring""" transformers.utils.logging._reset_library_root_logger() __lowerCAmelCase : int = logging.logging.getLogger() with CaptureLogger(_SCREAMING_SNAKE_CASE) as cl: # this action activates the env var logging.get_logger("transformers.models.bart.tokenization_bart") self.assertIn("Unknown option TRANSFORMERS_VERBOSITY=super-error" , cl.out) # no need to restore as nothing was changed def _SCREAMING_SNAKE_CASE ( self: List[str]) -> Tuple: """simple docstring""" transformers.utils.logging._reset_library_root_logger() __lowerCAmelCase : Dict = logging.get_logger("transformers.models.bart.tokenization_bart") __lowerCAmelCase : str = "Testing 1, 2, 3" with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS="1"): # nothing should be logged as env var disables this method with CaptureLogger(_SCREAMING_SNAKE_CASE) as cl: logger.warning_advice(_SCREAMING_SNAKE_CASE) self.assertEqual(cl.out , "") with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS=""): # should log normally as TRANSFORMERS_NO_ADVISORY_WARNINGS is unset with CaptureLogger(_SCREAMING_SNAKE_CASE) as cl: logger.warning_advice(_SCREAMING_SNAKE_CASE) self.assertEqual(cl.out , msg + "\n") def _lowercase ( ) -> List[str]: disable_progress_bar() assert are_progress_bars_disabled() enable_progress_bar() assert not are_progress_bars_disabled()
269
"""simple docstring""" def _lowercase ( __snake_case ,__snake_case ) -> float: if digit_amount > 0: return round(number - int(__snake_case ) ,__snake_case ) return number - int(__snake_case ) if __name__ == "__main__": print(decimal_isolate(1.53, 0)) print(decimal_isolate(35.345, 1)) print(decimal_isolate(35.345, 2)) print(decimal_isolate(35.345, 3)) print(decimal_isolate(-14.789, 3)) print(decimal_isolate(0, 2)) print(decimal_isolate(-14.123, 1)) print(decimal_isolate(-14.123, 2)) print(decimal_isolate(-14.123, 3))
269
1
"""simple docstring""" from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = None class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 2 @register_to_config def __init__( self: str , _SCREAMING_SNAKE_CASE: float = 0.02 , _SCREAMING_SNAKE_CASE: float = 100 , _SCREAMING_SNAKE_CASE: float = 1.007 , _SCREAMING_SNAKE_CASE: float = 80 , _SCREAMING_SNAKE_CASE: float = 0.05 , _SCREAMING_SNAKE_CASE: float = 50 , ) -> Any: """simple docstring""" __lowerCAmelCase : Tuple = sigma_max # setable values __lowerCAmelCase : int = None __lowerCAmelCase : np.IntTensor = None __lowerCAmelCase : torch.FloatTensor = None # sigma(t_i) def _SCREAMING_SNAKE_CASE ( self: int , _SCREAMING_SNAKE_CASE: torch.FloatTensor , _SCREAMING_SNAKE_CASE: Optional[int] = None) -> torch.FloatTensor: """simple docstring""" return sample def _SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Union[str, torch.device] = None) -> List[Any]: """simple docstring""" __lowerCAmelCase : str = num_inference_steps __lowerCAmelCase : Tuple = np.arange(0 , self.num_inference_steps)[::-1].copy() __lowerCAmelCase : Union[str, Any] = torch.from_numpy(_SCREAMING_SNAKE_CASE).to(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] __lowerCAmelCase : Dict = torch.tensor(_SCREAMING_SNAKE_CASE , dtype=torch.floataa , device=_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: int , _SCREAMING_SNAKE_CASE: torch.FloatTensor , _SCREAMING_SNAKE_CASE: float , _SCREAMING_SNAKE_CASE: Optional[torch.Generator] = None) -> Tuple[torch.FloatTensor, float]: """simple docstring""" if self.config.s_min <= sigma <= self.config.s_max: __lowerCAmelCase : Dict = min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1) else: __lowerCAmelCase : int = 0 # sample eps ~ N(0, S_noise^2 * I) __lowerCAmelCase : Tuple = self.config.s_noise * randn_tensor(sample.shape , generator=_SCREAMING_SNAKE_CASE).to(sample.device) __lowerCAmelCase : Optional[int] = sigma + gamma * sigma __lowerCAmelCase : Tuple = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def _SCREAMING_SNAKE_CASE ( self: str , _SCREAMING_SNAKE_CASE: torch.FloatTensor , _SCREAMING_SNAKE_CASE: float , _SCREAMING_SNAKE_CASE: float , _SCREAMING_SNAKE_CASE: torch.FloatTensor , _SCREAMING_SNAKE_CASE: bool = True , ) -> Union[KarrasVeOutput, Tuple]: """simple docstring""" __lowerCAmelCase : List[Any] = sample_hat + sigma_hat * model_output __lowerCAmelCase : Any = (sample_hat - pred_original_sample) / sigma_hat __lowerCAmelCase : Dict = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=_SCREAMING_SNAKE_CASE , derivative=_SCREAMING_SNAKE_CASE , pred_original_sample=_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: torch.FloatTensor , _SCREAMING_SNAKE_CASE: float , _SCREAMING_SNAKE_CASE: float , _SCREAMING_SNAKE_CASE: torch.FloatTensor , _SCREAMING_SNAKE_CASE: torch.FloatTensor , _SCREAMING_SNAKE_CASE: torch.FloatTensor , _SCREAMING_SNAKE_CASE: bool = True , ) -> Union[KarrasVeOutput, Tuple]: """simple docstring""" __lowerCAmelCase : List[Any] = sample_prev + sigma_prev * model_output __lowerCAmelCase : Any = (sample_prev - pred_original_sample) / sigma_prev __lowerCAmelCase : Any = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=_SCREAMING_SNAKE_CASE , derivative=_SCREAMING_SNAKE_CASE , pred_original_sample=_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Any) -> List[str]: """simple docstring""" raise NotImplementedError()
269
"""simple docstring""" from ..utils import DummyObject, requires_backends class A__ ( metaclass=__SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = ['torch', 'torchsde'] def __init__( self: int , *_SCREAMING_SNAKE_CASE: Optional[Any] , **_SCREAMING_SNAKE_CASE: str) -> Dict: """simple docstring""" requires_backends(self , ["torch", "torchsde"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls: Optional[Any] , *_SCREAMING_SNAKE_CASE: Optional[int] , **_SCREAMING_SNAKE_CASE: Optional[int]) -> Optional[Any]: """simple docstring""" requires_backends(cls , ["torch", "torchsde"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls: Dict , *_SCREAMING_SNAKE_CASE: List[Any] , **_SCREAMING_SNAKE_CASE: Any) -> Optional[int]: """simple docstring""" requires_backends(cls , ["torch", "torchsde"])
269
1
"""simple docstring""" import gc import unittest import numpy as np import torch import torch.nn.functional as F from transformers import ( ClapTextConfig, ClapTextModelWithProjection, RobertaTokenizer, SpeechTaHifiGan, SpeechTaHifiGanConfig, ) from diffusers import ( AudioLDMPipeline, AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_AUDIO_BATCH_PARAMS, TEXT_TO_AUDIO_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = AudioLDMPipeline SCREAMING_SNAKE_CASE = TEXT_TO_AUDIO_PARAMS SCREAMING_SNAKE_CASE = TEXT_TO_AUDIO_BATCH_PARAMS SCREAMING_SNAKE_CASE = frozenset( [ 'num_inference_steps', 'num_waveforms_per_prompt', 'generator', 'latents', 'output_type', 'return_dict', 'callback', 'callback_steps', ] ) def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> Tuple: """simple docstring""" torch.manual_seed(0) __lowerCAmelCase : Optional[int] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=(32, 64) , class_embed_type="simple_projection" , projection_class_embeddings_input_dim=32 , class_embeddings_concat=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : List[Any] = DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=_SCREAMING_SNAKE_CASE , set_alpha_to_one=_SCREAMING_SNAKE_CASE , ) torch.manual_seed(0) __lowerCAmelCase : int = AutoencoderKL( block_out_channels=[32, 64] , in_channels=1 , out_channels=1 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) torch.manual_seed(0) __lowerCAmelCase : List[str] = ClapTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , projection_dim=32 , ) __lowerCAmelCase : List[Any] = ClapTextModelWithProjection(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = RobertaTokenizer.from_pretrained("hf-internal-testing/tiny-random-roberta" , model_max_length=77) __lowerCAmelCase : Union[str, Any] = SpeechTaHifiGanConfig( model_in_dim=8 , sampling_rate=1_6000 , upsample_initial_channel=16 , upsample_rates=[2, 2] , upsample_kernel_sizes=[4, 4] , resblock_kernel_sizes=[3, 7] , resblock_dilation_sizes=[[1, 3, 5], [1, 3, 5]] , normalize_before=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Dict = SpeechTaHifiGan(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "vocoder": vocoder, } return components def _SCREAMING_SNAKE_CASE ( self: Optional[int] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Union[str, Any]=0) -> Optional[int]: """simple docstring""" if str(_SCREAMING_SNAKE_CASE).startswith("mps"): __lowerCAmelCase : Optional[int] = torch.manual_seed(_SCREAMING_SNAKE_CASE) else: __lowerCAmelCase : Dict = torch.Generator(device=_SCREAMING_SNAKE_CASE).manual_seed(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = { "prompt": "A hammer hitting a wooden surface", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, } return inputs def _SCREAMING_SNAKE_CASE ( self: int) -> Any: """simple docstring""" __lowerCAmelCase : Dict = "cpu" # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : Tuple = self.get_dummy_components() __lowerCAmelCase : Optional[int] = AudioLDMPipeline(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = audioldm_pipe.to(_SCREAMING_SNAKE_CASE) audioldm_pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = self.get_dummy_inputs(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = audioldm_pipe(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = output.audios[0] assert audio.ndim == 1 assert len(_SCREAMING_SNAKE_CASE) == 256 __lowerCAmelCase : List[Any] = audio[:10] __lowerCAmelCase : Union[str, Any] = np.array( [-0.0050, 0.0050, -0.0060, 0.0033, -0.0026, 0.0033, -0.0027, 0.0033, -0.0028, 0.0033]) assert np.abs(audio_slice - expected_slice).max() < 1e-2 def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : List[Any] = self.get_dummy_components() __lowerCAmelCase : str = AudioLDMPipeline(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = audioldm_pipe.to(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = audioldm_pipe.to(_SCREAMING_SNAKE_CASE) audioldm_pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = self.get_dummy_inputs(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = 3 * [inputs["prompt"]] # forward __lowerCAmelCase : str = audioldm_pipe(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = output.audios[0] __lowerCAmelCase : List[Any] = self.get_dummy_inputs(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = 3 * [inputs.pop("prompt")] __lowerCAmelCase : List[Any] = audioldm_pipe.tokenizer( _SCREAMING_SNAKE_CASE , padding="max_length" , max_length=audioldm_pipe.tokenizer.model_max_length , truncation=_SCREAMING_SNAKE_CASE , return_tensors="pt" , ) __lowerCAmelCase : Optional[int] = text_inputs["input_ids"].to(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = audioldm_pipe.text_encoder( _SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Union[str, Any] = prompt_embeds.text_embeds # additional L_2 normalization over each hidden-state __lowerCAmelCase : Dict = F.normalize(_SCREAMING_SNAKE_CASE , dim=-1) __lowerCAmelCase : Union[str, Any] = prompt_embeds # forward __lowerCAmelCase : Union[str, Any] = audioldm_pipe(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = output.audios[0] assert np.abs(audio_a - audio_a).max() < 1e-2 def _SCREAMING_SNAKE_CASE ( self: Any) -> int: """simple docstring""" __lowerCAmelCase : Any = self.get_dummy_components() __lowerCAmelCase : Optional[int] = AudioLDMPipeline(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = audioldm_pipe.to(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = audioldm_pipe.to(_SCREAMING_SNAKE_CASE) audioldm_pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = self.get_dummy_inputs(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = 3 * ["this is a negative prompt"] __lowerCAmelCase : int = negative_prompt __lowerCAmelCase : Optional[Any] = 3 * [inputs["prompt"]] # forward __lowerCAmelCase : Any = audioldm_pipe(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = output.audios[0] __lowerCAmelCase : str = self.get_dummy_inputs(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = 3 * [inputs.pop("prompt")] __lowerCAmelCase : Any = [] for p in [prompt, negative_prompt]: __lowerCAmelCase : Optional[int] = audioldm_pipe.tokenizer( _SCREAMING_SNAKE_CASE , padding="max_length" , max_length=audioldm_pipe.tokenizer.model_max_length , truncation=_SCREAMING_SNAKE_CASE , return_tensors="pt" , ) __lowerCAmelCase : Optional[int] = text_inputs["input_ids"].to(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = audioldm_pipe.text_encoder( _SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : List[Any] = text_embeds.text_embeds # additional L_2 normalization over each hidden-state __lowerCAmelCase : Tuple = F.normalize(_SCREAMING_SNAKE_CASE , dim=-1) embeds.append(_SCREAMING_SNAKE_CASE) __lowerCAmelCase , __lowerCAmelCase : int = embeds # forward __lowerCAmelCase : int = audioldm_pipe(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = output.audios[0] assert np.abs(audio_a - audio_a).max() < 1e-2 def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> Tuple: """simple docstring""" __lowerCAmelCase : int = "cpu" # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : Optional[Any] = self.get_dummy_components() __lowerCAmelCase : Union[str, Any] = PNDMScheduler(skip_prk_steps=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = AudioLDMPipeline(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = audioldm_pipe.to(_SCREAMING_SNAKE_CASE) audioldm_pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = self.get_dummy_inputs(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = "egg cracking" __lowerCAmelCase : List[Any] = audioldm_pipe(**_SCREAMING_SNAKE_CASE , negative_prompt=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = output.audios[0] assert audio.ndim == 1 assert len(_SCREAMING_SNAKE_CASE) == 256 __lowerCAmelCase : Any = audio[:10] __lowerCAmelCase : Tuple = np.array( [-0.0051, 0.0050, -0.0060, 0.0034, -0.0026, 0.0033, -0.0027, 0.0033, -0.0028, 0.0032]) assert np.abs(audio_slice - expected_slice).max() < 1e-2 def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> str: """simple docstring""" __lowerCAmelCase : Union[str, Any] = "cpu" # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : Any = self.get_dummy_components() __lowerCAmelCase : Optional[int] = PNDMScheduler(skip_prk_steps=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = AudioLDMPipeline(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = audioldm_pipe.to(_SCREAMING_SNAKE_CASE) audioldm_pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = "A hammer hitting a wooden surface" # test num_waveforms_per_prompt=1 (default) __lowerCAmelCase : Any = audioldm_pipe(_SCREAMING_SNAKE_CASE , num_inference_steps=2).audios assert audios.shape == (1, 256) # test num_waveforms_per_prompt=1 (default) for batch of prompts __lowerCAmelCase : Optional[Any] = 2 __lowerCAmelCase : Dict = audioldm_pipe([prompt] * batch_size , num_inference_steps=2).audios assert audios.shape == (batch_size, 256) # test num_waveforms_per_prompt for single prompt __lowerCAmelCase : List[str] = 2 __lowerCAmelCase : str = audioldm_pipe(_SCREAMING_SNAKE_CASE , num_inference_steps=2 , num_waveforms_per_prompt=_SCREAMING_SNAKE_CASE).audios assert audios.shape == (num_waveforms_per_prompt, 256) # test num_waveforms_per_prompt for batch of prompts __lowerCAmelCase : int = 2 __lowerCAmelCase : Optional[int] = audioldm_pipe( [prompt] * batch_size , num_inference_steps=2 , num_waveforms_per_prompt=_SCREAMING_SNAKE_CASE).audios assert audios.shape == (batch_size * num_waveforms_per_prompt, 256) def _SCREAMING_SNAKE_CASE ( self: List[str]) -> Dict: """simple docstring""" __lowerCAmelCase : Union[str, Any] = "cpu" # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : Union[str, Any] = self.get_dummy_components() __lowerCAmelCase : str = AudioLDMPipeline(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = audioldm_pipe.to(_SCREAMING_SNAKE_CASE) audioldm_pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = audioldm_pipe.vocoder.config.sampling_rate __lowerCAmelCase : Optional[int] = self.get_dummy_inputs(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = audioldm_pipe(audio_length_in_s=0.016 , **_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = output.audios[0] assert audio.ndim == 1 assert len(_SCREAMING_SNAKE_CASE) / vocoder_sampling_rate == 0.016 __lowerCAmelCase : int = audioldm_pipe(audio_length_in_s=0.032 , **_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = output.audios[0] assert audio.ndim == 1 assert len(_SCREAMING_SNAKE_CASE) / vocoder_sampling_rate == 0.032 def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> int: """simple docstring""" __lowerCAmelCase : Optional[Any] = self.get_dummy_components() __lowerCAmelCase : Optional[Any] = AudioLDMPipeline(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = audioldm_pipe.to(_SCREAMING_SNAKE_CASE) audioldm_pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = ["hey"] __lowerCAmelCase : List[Any] = audioldm_pipe(_SCREAMING_SNAKE_CASE , num_inference_steps=1) __lowerCAmelCase : List[Any] = output.audios.shape assert audio_shape == (1, 256) __lowerCAmelCase : Tuple = audioldm_pipe.vocoder.config config.model_in_dim *= 2 __lowerCAmelCase : int = SpeechTaHifiGan(_SCREAMING_SNAKE_CASE).to(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = audioldm_pipe(_SCREAMING_SNAKE_CASE , num_inference_steps=1) __lowerCAmelCase : Optional[Any] = output.audios.shape # waveform shape is unchanged, we just have 2x the number of mel channels in the spectrogram assert audio_shape == (1, 256) def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> List[Any]: """simple docstring""" self._test_attention_slicing_forward_pass(test_mean_pixel_difference=_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Dict) -> List[Any]: """simple docstring""" self._test_inference_batch_single_identical(test_mean_pixel_difference=_SCREAMING_SNAKE_CASE) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def _SCREAMING_SNAKE_CASE ( self: str) -> Optional[int]: """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=_SCREAMING_SNAKE_CASE) @slow class A__ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Any) -> int: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Any="cpu" , _SCREAMING_SNAKE_CASE: Optional[int]=torch.floataa , _SCREAMING_SNAKE_CASE: Optional[Any]=0) -> int: """simple docstring""" __lowerCAmelCase : Optional[Any] = torch.Generator(device=_SCREAMING_SNAKE_CASE).manual_seed(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = np.random.RandomState(_SCREAMING_SNAKE_CASE).standard_normal((1, 8, 128, 16)) __lowerCAmelCase : Tuple = torch.from_numpy(_SCREAMING_SNAKE_CASE).to(device=_SCREAMING_SNAKE_CASE , dtype=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = { "prompt": "A hammer hitting a wooden surface", "latents": latents, "generator": generator, "num_inference_steps": 3, "guidance_scale": 2.5, } return inputs def _SCREAMING_SNAKE_CASE ( self: Dict) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : List[str] = AudioLDMPipeline.from_pretrained("cvssp/audioldm") __lowerCAmelCase : Dict = audioldm_pipe.to(_SCREAMING_SNAKE_CASE) audioldm_pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = self.get_inputs(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = 25 __lowerCAmelCase : Any = audioldm_pipe(**_SCREAMING_SNAKE_CASE).audios[0] assert audio.ndim == 1 assert len(_SCREAMING_SNAKE_CASE) == 8_1920 __lowerCAmelCase : Dict = audio[7_7230:7_7240] __lowerCAmelCase : List[str] = np.array( [-0.4884, -0.4607, 0.0023, 0.5007, 0.5896, 0.5151, 0.3813, -0.0208, -0.3687, -0.4315]) __lowerCAmelCase : List[Any] = np.abs(expected_slice - audio_slice).max() assert max_diff < 1e-2 def _SCREAMING_SNAKE_CASE ( self: int) -> int: """simple docstring""" __lowerCAmelCase : Optional[int] = AudioLDMPipeline.from_pretrained("cvssp/audioldm") __lowerCAmelCase : List[str] = LMSDiscreteScheduler.from_config(audioldm_pipe.scheduler.config) __lowerCAmelCase : Dict = audioldm_pipe.to(_SCREAMING_SNAKE_CASE) audioldm_pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = self.get_inputs(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = audioldm_pipe(**_SCREAMING_SNAKE_CASE).audios[0] assert audio.ndim == 1 assert len(_SCREAMING_SNAKE_CASE) == 8_1920 __lowerCAmelCase : int = audio[2_7780:2_7790] __lowerCAmelCase : Optional[int] = np.array([-0.2131, -0.0873, -0.0124, -0.0189, 0.0569, 0.1373, 0.1883, 0.2886, 0.3297, 0.2212]) __lowerCAmelCase : Dict = np.abs(expected_slice - audio_slice).max() assert max_diff < 3e-2
269
"""simple docstring""" def _lowercase ( ) -> int: return 1 def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else two_pence(x - 2 ) + one_pence() def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else five_pence(x - 5 ) + two_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else ten_pence(x - 10 ) + five_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else twenty_pence(x - 20 ) + ten_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else fifty_pence(x - 50 ) + twenty_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else one_pound(x - 100 ) + fifty_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else two_pound(x - 200 ) + one_pound(__snake_case ) def _lowercase ( __snake_case = 200 ) -> int: return two_pound(__snake_case ) if __name__ == "__main__": print(solution(int(input().strip())))
269
1
"""simple docstring""" import importlib import inspect import json import os import re import shutil import sys from pathlib import Path from typing import Dict, Optional, Union from urllib import request from huggingface_hub import HfFolder, cached_download, hf_hub_download, model_info from packaging import version from .. import __version__ from . import DIFFUSERS_DYNAMIC_MODULE_NAME, HF_MODULES_CACHE, logging __snake_case : Union[str, Any] = ( 'https://raw.githubusercontent.com/huggingface/diffusers/{revision}/examples/community/{pipeline}.py' ) __snake_case : Optional[int] = logging.get_logger(__name__) # pylint: disable=invalid-name def _lowercase ( ) -> int: __lowerCAmelCase : str = "https://pypi.org/pypi/diffusers/json" __lowerCAmelCase : List[str] = json.loads(request.urlopen(__snake_case ).read() )["releases"].keys() return sorted(__snake_case ,key=lambda __snake_case : version.Version(__snake_case ) ) def _lowercase ( ) -> Optional[int]: # This function has already been executed if HF_MODULES_CACHE already is in the Python path. if HF_MODULES_CACHE in sys.path: return sys.path.append(__snake_case ) os.makedirs(__snake_case ,exist_ok=__snake_case ) __lowerCAmelCase : List[str] = Path(__snake_case ) / "__init__.py" if not init_path.exists(): init_path.touch() def _lowercase ( __snake_case ) -> List[str]: init_hf_modules() __lowerCAmelCase : int = Path(__snake_case ) / name # If the parent module does not exist yet, recursively create it. if not dynamic_module_path.parent.exists(): create_dynamic_module(dynamic_module_path.parent ) os.makedirs(__snake_case ,exist_ok=__snake_case ) __lowerCAmelCase : str = dynamic_module_path / "__init__.py" if not init_path.exists(): init_path.touch() def _lowercase ( __snake_case ) -> Optional[Any]: with open(__snake_case ,"r" ,encoding="utf-8" ) as f: __lowerCAmelCase : Union[str, Any] = f.read() # Imports of the form `import .xxx` __lowerCAmelCase : Union[str, Any] = re.findall("^\s*import\s+\.(\S+)\s*$" ,__snake_case ,flags=re.MULTILINE ) # Imports of the form `from .xxx import yyy` relative_imports += re.findall("^\s*from\s+\.(\S+)\s+import" ,__snake_case ,flags=re.MULTILINE ) # Unique-ify return list(set(__snake_case ) ) def _lowercase ( __snake_case ) -> Dict: __lowerCAmelCase : Optional[Any] = False __lowerCAmelCase : Optional[Any] = [module_file] __lowerCAmelCase : Tuple = [] # Let's recurse through all relative imports while not no_change: __lowerCAmelCase : Dict = [] for f in files_to_check: new_imports.extend(get_relative_imports(__snake_case ) ) __lowerCAmelCase : int = Path(__snake_case ).parent __lowerCAmelCase : Union[str, Any] = [str(module_path / m ) for m in new_imports] __lowerCAmelCase : str = [f for f in new_import_files if f not in all_relative_imports] __lowerCAmelCase : Tuple = [F"""{f}.py""" for f in new_import_files] __lowerCAmelCase : Any = len(__snake_case ) == 0 all_relative_imports.extend(__snake_case ) return all_relative_imports def _lowercase ( __snake_case ) -> Optional[int]: with open(__snake_case ,"r" ,encoding="utf-8" ) as f: __lowerCAmelCase : Optional[int] = f.read() # Imports of the form `import xxx` __lowerCAmelCase : List[str] = re.findall("^\s*import\s+(\S+)\s*$" ,__snake_case ,flags=re.MULTILINE ) # Imports of the form `from xxx import yyy` imports += re.findall("^\s*from\s+(\S+)\s+import" ,__snake_case ,flags=re.MULTILINE ) # Only keep the top-level module __lowerCAmelCase : Any = [imp.split("." )[0] for imp in imports if not imp.startswith("." )] # Unique-ify and test we got them all __lowerCAmelCase : Tuple = list(set(__snake_case ) ) __lowerCAmelCase : Optional[int] = [] for imp in imports: try: importlib.import_module(__snake_case ) except ImportError: missing_packages.append(__snake_case ) if len(__snake_case ) > 0: raise ImportError( "This modeling file requires the following packages that were not found in your environment: " F"""{', '.join(__snake_case )}. Run `pip install {' '.join(__snake_case )}`""" ) return get_relative_imports(__snake_case ) def _lowercase ( __snake_case ,__snake_case ) -> str: __lowerCAmelCase : Tuple = module_path.replace(os.path.sep ,"." ) __lowerCAmelCase : Optional[int] = importlib.import_module(__snake_case ) if class_name is None: return find_pipeline_class(__snake_case ) return getattr(__snake_case ,__snake_case ) def _lowercase ( __snake_case ) -> Optional[int]: from ..pipelines import DiffusionPipeline __lowerCAmelCase : Any = dict(inspect.getmembers(__snake_case ,inspect.isclass ) ) __lowerCAmelCase : str = None for cls_name, cls in cls_members.items(): if ( cls_name != DiffusionPipeline.__name__ and issubclass(cls ,__snake_case ) and cls.__module__.split("." )[0] != "diffusers" ): if pipeline_class is not None: raise ValueError( F"""Multiple classes that inherit from {DiffusionPipeline.__name__} have been found:""" F""" {pipeline_class.__name__}, and {cls_name}. Please make sure to define only one in""" F""" {loaded_module}.""" ) __lowerCAmelCase : str = cls return pipeline_class def _lowercase ( __snake_case ,__snake_case ,__snake_case = None ,__snake_case = False ,__snake_case = False ,__snake_case = None ,__snake_case = None ,__snake_case = None ,__snake_case = False ,) -> Optional[int]: __lowerCAmelCase : Optional[int] = str(__snake_case ) __lowerCAmelCase : List[str] = os.path.join(__snake_case ,__snake_case ) if os.path.isfile(__snake_case ): __lowerCAmelCase : int = module_file_or_url __lowerCAmelCase : int = "local" elif pretrained_model_name_or_path.count("/" ) == 0: __lowerCAmelCase : Optional[int] = get_diffusers_versions() # cut ".dev0" __lowerCAmelCase : Any = "v" + ".".join(__version__.split("." )[:3] ) # retrieve github version that matches if revision is None: __lowerCAmelCase : str = latest_version if latest_version[1:] in available_versions else "main" logger.info(F"""Defaulting to latest_version: {revision}.""" ) elif revision in available_versions: __lowerCAmelCase : List[str] = F"""v{revision}""" elif revision == "main": __lowerCAmelCase : str = revision else: raise ValueError( F"""`custom_revision`: {revision} does not exist. Please make sure to choose one of""" F""" {', '.join(available_versions + ['main'] )}.""" ) # community pipeline on GitHub __lowerCAmelCase : List[Any] = COMMUNITY_PIPELINES_URL.format(revision=__snake_case ,pipeline=__snake_case ) try: __lowerCAmelCase : str = cached_download( __snake_case ,cache_dir=__snake_case ,force_download=__snake_case ,proxies=__snake_case ,resume_download=__snake_case ,local_files_only=__snake_case ,use_auth_token=__snake_case ,) __lowerCAmelCase : Any = "git" __lowerCAmelCase : List[str] = pretrained_model_name_or_path + ".py" except EnvironmentError: logger.error(F"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise else: try: # Load from URL or cache if already cached __lowerCAmelCase : Optional[Any] = hf_hub_download( __snake_case ,__snake_case ,cache_dir=__snake_case ,force_download=__snake_case ,proxies=__snake_case ,resume_download=__snake_case ,local_files_only=__snake_case ,use_auth_token=__snake_case ,) __lowerCAmelCase : Any = os.path.join("local" ,"--".join(pretrained_model_name_or_path.split("/" ) ) ) except EnvironmentError: logger.error(F"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise # Check we have all the requirements in our environment __lowerCAmelCase : Optional[int] = check_imports(__snake_case ) # Now we move the module inside our cached dynamic modules. __lowerCAmelCase : Tuple = DIFFUSERS_DYNAMIC_MODULE_NAME + os.path.sep + submodule create_dynamic_module(__snake_case ) __lowerCAmelCase : Dict = Path(__snake_case ) / full_submodule if submodule == "local" or submodule == "git": # We always copy local files (we could hash the file to see if there was a change, and give them the name of # that hash, to only copy when there is a modification but it seems overkill for now). # The only reason we do the copy is to avoid putting too many folders in sys.path. shutil.copy(__snake_case ,submodule_path / module_file ) for module_needed in modules_needed: __lowerCAmelCase : List[Any] = F"""{module_needed}.py""" shutil.copy(os.path.join(__snake_case ,__snake_case ) ,submodule_path / module_needed ) else: # Get the commit hash # TODO: we will get this info in the etag soon, so retrieve it from there and not here. if isinstance(__snake_case ,__snake_case ): __lowerCAmelCase : List[str] = use_auth_token elif use_auth_token is True: __lowerCAmelCase : Union[str, Any] = HfFolder.get_token() else: __lowerCAmelCase : Union[str, Any] = None __lowerCAmelCase : Tuple = model_info(__snake_case ,revision=__snake_case ,token=__snake_case ).sha # The module file will end up being placed in a subfolder with the git hash of the repo. This way we get the # benefit of versioning. __lowerCAmelCase : Union[str, Any] = submodule_path / commit_hash __lowerCAmelCase : Tuple = full_submodule + os.path.sep + commit_hash create_dynamic_module(__snake_case ) if not (submodule_path / module_file).exists(): shutil.copy(__snake_case ,submodule_path / module_file ) # Make sure we also have every file with relative for module_needed in modules_needed: if not (submodule_path / module_needed).exists(): get_cached_module_file( __snake_case ,F"""{module_needed}.py""" ,cache_dir=__snake_case ,force_download=__snake_case ,resume_download=__snake_case ,proxies=__snake_case ,use_auth_token=__snake_case ,revision=__snake_case ,local_files_only=__snake_case ,) return os.path.join(__snake_case ,__snake_case ) def _lowercase ( __snake_case ,__snake_case ,__snake_case = None ,__snake_case = None ,__snake_case = False ,__snake_case = False ,__snake_case = None ,__snake_case = None ,__snake_case = None ,__snake_case = False ,**__snake_case ,) -> List[str]: __lowerCAmelCase : Union[str, Any] = get_cached_module_file( __snake_case ,__snake_case ,cache_dir=__snake_case ,force_download=__snake_case ,resume_download=__snake_case ,proxies=__snake_case ,use_auth_token=__snake_case ,revision=__snake_case ,local_files_only=__snake_case ,) return get_class_in_module(__snake_case ,final_module.replace(".py" ,"" ) )
269
"""simple docstring""" import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Optional[int] , _SCREAMING_SNAKE_CASE: str) -> Dict: """simple docstring""" with open(_SCREAMING_SNAKE_CASE , encoding="utf-8") as input_file: __lowerCAmelCase : List[str] = re.compile(r"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)") __lowerCAmelCase : List[Any] = input_file.read() __lowerCAmelCase : Any = regexp.search(_SCREAMING_SNAKE_CASE) return match def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: str) -> Optional[Any]: """simple docstring""" with open(_SCREAMING_SNAKE_CASE , encoding="utf-8") as input_file: __lowerCAmelCase : Any = re.compile(r"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL) __lowerCAmelCase : Optional[int] = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` __lowerCAmelCase : int = regexp.finditer(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = [match for match in matches if match is not None and match.group(1) is not None] return matches[0] if matches else None def _SCREAMING_SNAKE_CASE ( self: str) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[Any] = Path("./datasets") __lowerCAmelCase : Optional[int] = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_encoding_on_file_open(str(_SCREAMING_SNAKE_CASE)): raise AssertionError(F"""open(...) must use utf-8 encoding in {dataset}""") def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> Optional[int]: """simple docstring""" __lowerCAmelCase : Dict = Path("./datasets") __lowerCAmelCase : Union[str, Any] = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_print_statements(str(_SCREAMING_SNAKE_CASE)): raise AssertionError(F"""print statement found in {dataset}. Use datasets.logger/logging instead.""")
269
1
"""simple docstring""" import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class A__ ( unittest.TestCase ): '''simple docstring''' def __init__( self: int , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: Union[str, Any]=7 , _SCREAMING_SNAKE_CASE: Tuple=3 , _SCREAMING_SNAKE_CASE: Optional[Any]=30 , _SCREAMING_SNAKE_CASE: int=400 , _SCREAMING_SNAKE_CASE: Union[str, Any]=True , _SCREAMING_SNAKE_CASE: str=None , _SCREAMING_SNAKE_CASE: Optional[Any]=True , _SCREAMING_SNAKE_CASE: List[Any]=[0.5, 0.5, 0.5] , _SCREAMING_SNAKE_CASE: List[str]=[0.5, 0.5, 0.5] , _SCREAMING_SNAKE_CASE: Any=True , _SCREAMING_SNAKE_CASE: Tuple=1 / 255 , _SCREAMING_SNAKE_CASE: str=True , ) -> Tuple: """simple docstring""" __lowerCAmelCase : str = size if size is not None else {"shortest_edge": 18, "longest_edge": 1333} __lowerCAmelCase : str = parent __lowerCAmelCase : Optional[int] = batch_size __lowerCAmelCase : Optional[int] = num_channels __lowerCAmelCase : str = min_resolution __lowerCAmelCase : List[Any] = max_resolution __lowerCAmelCase : Tuple = do_resize __lowerCAmelCase : List[str] = size __lowerCAmelCase : Union[str, Any] = do_normalize __lowerCAmelCase : str = image_mean __lowerCAmelCase : List[Any] = image_std __lowerCAmelCase : Optional[int] = do_rescale __lowerCAmelCase : Any = rescale_factor __lowerCAmelCase : Optional[Any] = do_pad def _SCREAMING_SNAKE_CASE ( self: str) -> str: """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: str=False) -> Dict: """simple docstring""" if not batched: __lowerCAmelCase : Optional[Any] = image_inputs[0] if isinstance(_SCREAMING_SNAKE_CASE , Image.Image): __lowerCAmelCase , __lowerCAmelCase : Any = image.size else: __lowerCAmelCase , __lowerCAmelCase : Optional[Any] = image.shape[1], image.shape[2] if w < h: __lowerCAmelCase : Any = int(self.size["shortest_edge"] * h / w) __lowerCAmelCase : List[Any] = self.size["shortest_edge"] elif w > h: __lowerCAmelCase : str = self.size["shortest_edge"] __lowerCAmelCase : Optional[int] = int(self.size["shortest_edge"] * w / h) else: __lowerCAmelCase : List[str] = self.size["shortest_edge"] __lowerCAmelCase : str = self.size["shortest_edge"] else: __lowerCAmelCase : Optional[int] = [] for image in image_inputs: __lowerCAmelCase , __lowerCAmelCase : Dict = self.get_expected_values([image]) expected_values.append((expected_height, expected_width)) __lowerCAmelCase : int = max(_SCREAMING_SNAKE_CASE , key=lambda _SCREAMING_SNAKE_CASE: item[0])[0] __lowerCAmelCase : Optional[Any] = max(_SCREAMING_SNAKE_CASE , key=lambda _SCREAMING_SNAKE_CASE: item[1])[1] return expected_height, expected_width @require_torch @require_vision class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = ConditionalDetrImageProcessor if is_vision_available() else None def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> Tuple: """simple docstring""" __lowerCAmelCase : str = ConditionalDetrImageProcessingTester(self) @property def _SCREAMING_SNAKE_CASE ( self: str) -> int: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def _SCREAMING_SNAKE_CASE ( self: int) -> Optional[int]: """simple docstring""" __lowerCAmelCase : Tuple = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , "image_mean")) self.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , "image_std")) self.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , "do_normalize")) self.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , "do_resize")) self.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , "size")) def _SCREAMING_SNAKE_CASE ( self: Any) -> Dict: """simple docstring""" __lowerCAmelCase : Tuple = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 1333}) self.assertEqual(image_processor.do_pad , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=_SCREAMING_SNAKE_CASE) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84}) self.assertEqual(image_processor.do_pad , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: int) -> Optional[int]: """simple docstring""" pass def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Dict: """simple docstring""" __lowerCAmelCase : Union[str, Any] = self.image_processing_class(**self.image_processor_dict) # create random PIL images __lowerCAmelCase : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=_SCREAMING_SNAKE_CASE) for image in image_inputs: self.assertIsInstance(_SCREAMING_SNAKE_CASE , Image.Image) # Test not batched input __lowerCAmelCase : List[str] = image_processing(image_inputs[0] , return_tensors="pt").pixel_values __lowerCAmelCase , __lowerCAmelCase : Optional[int] = self.image_processor_tester.get_expected_values(_SCREAMING_SNAKE_CASE) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCAmelCase , __lowerCAmelCase : List[Any] = self.image_processor_tester.get_expected_values(_SCREAMING_SNAKE_CASE , batched=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = image_processing(_SCREAMING_SNAKE_CASE , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _SCREAMING_SNAKE_CASE ( self: int) -> int: """simple docstring""" __lowerCAmelCase : Dict = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors __lowerCAmelCase : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_SCREAMING_SNAKE_CASE , numpify=_SCREAMING_SNAKE_CASE) for image in image_inputs: self.assertIsInstance(_SCREAMING_SNAKE_CASE , np.ndarray) # Test not batched input __lowerCAmelCase : Any = image_processing(image_inputs[0] , return_tensors="pt").pixel_values __lowerCAmelCase , __lowerCAmelCase : List[str] = self.image_processor_tester.get_expected_values(_SCREAMING_SNAKE_CASE) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCAmelCase : Any = image_processing(_SCREAMING_SNAKE_CASE , return_tensors="pt").pixel_values __lowerCAmelCase , __lowerCAmelCase : List[Any] = self.image_processor_tester.get_expected_values(_SCREAMING_SNAKE_CASE , batched=_SCREAMING_SNAKE_CASE) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> int: """simple docstring""" __lowerCAmelCase : Union[str, Any] = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors __lowerCAmelCase : List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_SCREAMING_SNAKE_CASE , torchify=_SCREAMING_SNAKE_CASE) for image in image_inputs: self.assertIsInstance(_SCREAMING_SNAKE_CASE , torch.Tensor) # Test not batched input __lowerCAmelCase : Tuple = image_processing(image_inputs[0] , return_tensors="pt").pixel_values __lowerCAmelCase , __lowerCAmelCase : Optional[Any] = self.image_processor_tester.get_expected_values(_SCREAMING_SNAKE_CASE) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCAmelCase : Optional[Any] = image_processing(_SCREAMING_SNAKE_CASE , return_tensors="pt").pixel_values __lowerCAmelCase , __lowerCAmelCase : Optional[int] = self.image_processor_tester.get_expected_values(_SCREAMING_SNAKE_CASE , batched=_SCREAMING_SNAKE_CASE) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def _SCREAMING_SNAKE_CASE ( self: Dict) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : List[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png") with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r") as f: __lowerCAmelCase : List[Any] = json.loads(f.read()) __lowerCAmelCase : Tuple = {"image_id": 3_9769, "annotations": target} # encode them __lowerCAmelCase : Optional[Any] = ConditionalDetrImageProcessor.from_pretrained("microsoft/conditional-detr-resnet-50") __lowerCAmelCase : int = image_processing(images=_SCREAMING_SNAKE_CASE , annotations=_SCREAMING_SNAKE_CASE , return_tensors="pt") # verify pixel values __lowerCAmelCase : str = torch.Size([1, 3, 800, 1066]) self.assertEqual(encoding["pixel_values"].shape , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = torch.tensor([0.2796, 0.3138, 0.3481]) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , _SCREAMING_SNAKE_CASE , atol=1e-4)) # verify area __lowerCAmelCase : int = torch.tensor([5887.9600, 1_1250.2061, 48_9353.8438, 83_7122.7500, 14_7967.5156, 16_5732.3438]) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , _SCREAMING_SNAKE_CASE)) # verify boxes __lowerCAmelCase : Optional[Any] = torch.Size([6, 4]) self.assertEqual(encoding["labels"][0]["boxes"].shape , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215]) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , _SCREAMING_SNAKE_CASE , atol=1e-3)) # verify image_id __lowerCAmelCase : List[Any] = torch.tensor([3_9769]) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , _SCREAMING_SNAKE_CASE)) # verify is_crowd __lowerCAmelCase : List[str] = torch.tensor([0, 0, 0, 0, 0, 0]) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , _SCREAMING_SNAKE_CASE)) # verify class_labels __lowerCAmelCase : Tuple = torch.tensor([75, 75, 63, 65, 17, 17]) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , _SCREAMING_SNAKE_CASE)) # verify orig_size __lowerCAmelCase : List[Any] = torch.tensor([480, 640]) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , _SCREAMING_SNAKE_CASE)) # verify size __lowerCAmelCase : int = torch.tensor([800, 1066]) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , _SCREAMING_SNAKE_CASE)) @slow def _SCREAMING_SNAKE_CASE ( self: Dict) -> int: """simple docstring""" __lowerCAmelCase : Optional[int] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png") with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r") as f: __lowerCAmelCase : Optional[int] = json.loads(f.read()) __lowerCAmelCase : Optional[Any] = {"file_name": "000000039769.png", "image_id": 3_9769, "segments_info": target} __lowerCAmelCase : int = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic") # encode them __lowerCAmelCase : List[Any] = ConditionalDetrImageProcessor(format="coco_panoptic") __lowerCAmelCase : List[Any] = image_processing(images=_SCREAMING_SNAKE_CASE , annotations=_SCREAMING_SNAKE_CASE , masks_path=_SCREAMING_SNAKE_CASE , return_tensors="pt") # verify pixel values __lowerCAmelCase : str = torch.Size([1, 3, 800, 1066]) self.assertEqual(encoding["pixel_values"].shape , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = torch.tensor([0.2796, 0.3138, 0.3481]) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , _SCREAMING_SNAKE_CASE , atol=1e-4)) # verify area __lowerCAmelCase : Dict = torch.tensor([14_7979.6875, 16_5527.0469, 48_4638.5938, 1_1292.9375, 5879.6562, 7634.1147]) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , _SCREAMING_SNAKE_CASE)) # verify boxes __lowerCAmelCase : List[Any] = torch.Size([6, 4]) self.assertEqual(encoding["labels"][0]["boxes"].shape , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625]) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , _SCREAMING_SNAKE_CASE , atol=1e-3)) # verify image_id __lowerCAmelCase : int = torch.tensor([3_9769]) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , _SCREAMING_SNAKE_CASE)) # verify is_crowd __lowerCAmelCase : Optional[int] = torch.tensor([0, 0, 0, 0, 0, 0]) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , _SCREAMING_SNAKE_CASE)) # verify class_labels __lowerCAmelCase : Optional[Any] = torch.tensor([17, 17, 63, 75, 75, 93]) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , _SCREAMING_SNAKE_CASE)) # verify masks __lowerCAmelCase : Any = 82_2873 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , _SCREAMING_SNAKE_CASE) # verify orig_size __lowerCAmelCase : Tuple = torch.tensor([480, 640]) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , _SCREAMING_SNAKE_CASE)) # verify size __lowerCAmelCase : Dict = torch.tensor([800, 1066]) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , _SCREAMING_SNAKE_CASE))
269
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxSeqaSeqConfigWithPast from ...utils import logging if TYPE_CHECKING: from ...feature_extraction_utils import FeatureExtractionMixin from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType __snake_case : Optional[Any] = logging.get_logger(__name__) __snake_case : Optional[Any] = { 'openai/whisper-base': 'https://huggingface.co/openai/whisper-base/resolve/main/config.json', } # fmt: off __snake_case : Any = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 357, 366, 438, 532, 685, 705, 796, 930, 1_058, 1_220, 1_267, 1_279, 1_303, 1_343, 1_377, 1_391, 1_635, 1_782, 1_875, 2_162, 2_361, 2_488, 3_467, 4_008, 4_211, 4_600, 4_808, 5_299, 5_855, 6_329, 7_203, 9_609, 9_959, 10_563, 10_786, 11_420, 11_709, 11_907, 13_163, 13_697, 13_700, 14_808, 15_306, 16_410, 16_791, 17_992, 19_203, 19_510, 20_724, 22_305, 22_935, 27_007, 30_109, 30_420, 33_409, 34_949, 40_283, 40_493, 40_549, 47_282, 49_146, 50_257, 50_359, 50_360, 50_361 ] __snake_case : str = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 359, 503, 522, 542, 873, 893, 902, 918, 922, 931, 1_350, 1_853, 1_982, 2_460, 2_627, 3_246, 3_253, 3_268, 3_536, 3_846, 3_961, 4_183, 4_667, 6_585, 6_647, 7_273, 9_061, 9_383, 10_428, 10_929, 11_938, 12_033, 12_331, 12_562, 13_793, 14_157, 14_635, 15_265, 15_618, 16_553, 16_604, 18_362, 18_956, 20_075, 21_675, 22_520, 26_130, 26_161, 26_435, 28_279, 29_464, 31_650, 32_302, 32_470, 36_865, 42_863, 47_425, 49_870, 50_254, 50_258, 50_360, 50_361, 50_362 ] class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'whisper' SCREAMING_SNAKE_CASE = ['past_key_values'] SCREAMING_SNAKE_CASE = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self: Optional[int] , _SCREAMING_SNAKE_CASE: Any=5_1865 , _SCREAMING_SNAKE_CASE: Optional[Any]=80 , _SCREAMING_SNAKE_CASE: Optional[int]=6 , _SCREAMING_SNAKE_CASE: Any=4 , _SCREAMING_SNAKE_CASE: Dict=6 , _SCREAMING_SNAKE_CASE: Dict=4 , _SCREAMING_SNAKE_CASE: Optional[Any]=1536 , _SCREAMING_SNAKE_CASE: List[str]=1536 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.0 , _SCREAMING_SNAKE_CASE: str=0.0 , _SCREAMING_SNAKE_CASE: List[str]=5_0257 , _SCREAMING_SNAKE_CASE: Optional[Any]=True , _SCREAMING_SNAKE_CASE: List[str]=True , _SCREAMING_SNAKE_CASE: Optional[int]="gelu" , _SCREAMING_SNAKE_CASE: Tuple=256 , _SCREAMING_SNAKE_CASE: str=0.0 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.0 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.0 , _SCREAMING_SNAKE_CASE: List[Any]=0.02 , _SCREAMING_SNAKE_CASE: Tuple=False , _SCREAMING_SNAKE_CASE: Union[str, Any]=1500 , _SCREAMING_SNAKE_CASE: str=448 , _SCREAMING_SNAKE_CASE: Any=5_0256 , _SCREAMING_SNAKE_CASE: Any=5_0256 , _SCREAMING_SNAKE_CASE: List[str]=5_0256 , _SCREAMING_SNAKE_CASE: Dict=None , _SCREAMING_SNAKE_CASE: List[Any]=[220, 5_0256] , _SCREAMING_SNAKE_CASE: Dict=False , _SCREAMING_SNAKE_CASE: str=256 , _SCREAMING_SNAKE_CASE: List[str]=False , _SCREAMING_SNAKE_CASE: Tuple=0.05 , _SCREAMING_SNAKE_CASE: List[str]=10 , _SCREAMING_SNAKE_CASE: str=2 , _SCREAMING_SNAKE_CASE: Any=0.0 , _SCREAMING_SNAKE_CASE: Optional[int]=10 , _SCREAMING_SNAKE_CASE: int=0 , _SCREAMING_SNAKE_CASE: Any=7 , **_SCREAMING_SNAKE_CASE: List[str] , ) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[int] = vocab_size __lowerCAmelCase : Optional[Any] = num_mel_bins __lowerCAmelCase : int = d_model __lowerCAmelCase : List[Any] = encoder_layers __lowerCAmelCase : List[Any] = encoder_attention_heads __lowerCAmelCase : List[str] = decoder_layers __lowerCAmelCase : Tuple = decoder_attention_heads __lowerCAmelCase : Any = decoder_ffn_dim __lowerCAmelCase : Tuple = encoder_ffn_dim __lowerCAmelCase : List[str] = dropout __lowerCAmelCase : Union[str, Any] = attention_dropout __lowerCAmelCase : Union[str, Any] = activation_dropout __lowerCAmelCase : Dict = activation_function __lowerCAmelCase : Tuple = init_std __lowerCAmelCase : str = encoder_layerdrop __lowerCAmelCase : int = decoder_layerdrop __lowerCAmelCase : Optional[int] = use_cache __lowerCAmelCase : Union[str, Any] = encoder_layers __lowerCAmelCase : List[str] = scale_embedding # scale factor will be sqrt(d_model) if True __lowerCAmelCase : int = max_source_positions __lowerCAmelCase : Any = max_target_positions # Audio Classification-specific parameters. Feel free to ignore for other classes. __lowerCAmelCase : Dict = classifier_proj_size __lowerCAmelCase : Dict = use_weighted_layer_sum # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __lowerCAmelCase : int = apply_spec_augment __lowerCAmelCase : Union[str, Any] = mask_time_prob __lowerCAmelCase : str = mask_time_length __lowerCAmelCase : int = mask_time_min_masks __lowerCAmelCase : List[Any] = mask_feature_prob __lowerCAmelCase : Tuple = mask_feature_length __lowerCAmelCase : Any = mask_feature_min_masks __lowerCAmelCase : Union[str, Any] = median_filter_width super().__init__( pad_token_id=_SCREAMING_SNAKE_CASE , bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE , is_encoder_decoder=_SCREAMING_SNAKE_CASE , decoder_start_token_id=_SCREAMING_SNAKE_CASE , suppress_tokens=_SCREAMING_SNAKE_CASE , begin_suppress_tokens=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' @property def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> Mapping[str, Mapping[int, str]]: """simple docstring""" __lowerCAmelCase : List[str] = OrderedDict( [ ("input_features", {0: "batch", 1: "feature_size", 2: "encoder_sequence"}), ]) if self.use_past: __lowerCAmelCase : Tuple = {0: "batch"} else: __lowerCAmelCase : Union[str, Any] = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(_SCREAMING_SNAKE_CASE , direction="inputs") return common_inputs def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: Union["PreTrainedTokenizerBase", "FeatureExtractionMixin"] , _SCREAMING_SNAKE_CASE: int = -1 , _SCREAMING_SNAKE_CASE: int = -1 , _SCREAMING_SNAKE_CASE: bool = False , _SCREAMING_SNAKE_CASE: Optional["TensorType"] = None , _SCREAMING_SNAKE_CASE: int = 2_2050 , _SCREAMING_SNAKE_CASE: float = 5.0 , _SCREAMING_SNAKE_CASE: int = 220 , ) -> Mapping[str, Any]: """simple docstring""" __lowerCAmelCase : int = OrderedDict() __lowerCAmelCase : Optional[Any] = OnnxConfig.generate_dummy_inputs( self , preprocessor=preprocessor.feature_extractor , batch_size=_SCREAMING_SNAKE_CASE , framework=_SCREAMING_SNAKE_CASE , sampling_rate=_SCREAMING_SNAKE_CASE , time_duration=_SCREAMING_SNAKE_CASE , frequency=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : List[str] = encoder_inputs["input_features"].shape[2] __lowerCAmelCase : List[str] = encoder_sequence_length // 2 if self.use_past else seq_length __lowerCAmelCase : List[Any] = super().generate_dummy_inputs( preprocessor.tokenizer , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = encoder_inputs.pop("input_features") __lowerCAmelCase : List[Any] = decoder_inputs.pop("decoder_input_ids") if "past_key_values" in decoder_inputs: __lowerCAmelCase : int = decoder_inputs.pop("past_key_values") return dummy_inputs @property def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> float: """simple docstring""" return 1e-3
269
1
"""simple docstring""" from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
269
"""simple docstring""" import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration __snake_case : Optional[int] = 50_000 __snake_case : Dict = 5_000 __snake_case , __snake_case : Union[str, Any] = os.path.split(__file__) __snake_case : Any = os.path.join(RESULTS_BASEPATH, 'results', RESULTS_FILENAME.replace('.py', '.json')) @get_duration def _lowercase ( __snake_case ,__snake_case ) -> Dict: for i in range(__snake_case ): __lowerCAmelCase : Union[str, Any] = dataset[i] @get_duration def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> Dict: for i in range(0 ,len(__snake_case ) ,__snake_case ): __lowerCAmelCase : List[str] = dataset[i : i + batch_size] @get_duration def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> Dict: with dataset.formatted_as(type=__snake_case ): for i in range(__snake_case ): __lowerCAmelCase : Union[str, Any] = dataset[i] @get_duration def _lowercase ( __snake_case ,__snake_case ,__snake_case ,__snake_case ) -> str: with dataset.formatted_as(type=__snake_case ): for i in range(0 ,__snake_case ,__snake_case ): __lowerCAmelCase : Optional[int] = dataset[i : i + batch_size] def _lowercase ( ) -> Union[str, Any]: __lowerCAmelCase : Optional[int] = {"num examples": SPEED_TEST_N_EXAMPLES} __lowerCAmelCase : Optional[int] = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted, {"type": "pandas", "length": SMALL_TEST}), (read_formatted, {"type": "torch", "length": SMALL_TEST}), (read_formatted, {"type": "tensorflow", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] __lowerCAmelCase : Any = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print("generating dataset" ) __lowerCAmelCase : int = datasets.Features( {"list": datasets.Sequence(datasets.Value("float32" ) ), "numbers": datasets.Value("float32" )} ) __lowerCAmelCase : str = generate_example_dataset( os.path.join(__snake_case ,"dataset.arrow" ) ,__snake_case ,num_examples=__snake_case ,seq_shapes={"list": (100,)} ,) print("first set of iterations" ) for func, kwargs in functions: print(func.__name__ ,str(__snake_case ) ) __lowerCAmelCase : str = func(__snake_case ,**__snake_case ) print("shuffling dataset" ) __lowerCAmelCase : Optional[int] = dataset.shuffle() print("Second set of iterations (after shuffling" ) for func, kwargs in functions_shuffled: print("shuffled " ,func.__name__ ,str(__snake_case ) ) __lowerCAmelCase : List[str] = func( __snake_case ,**__snake_case ) with open(__snake_case ,"wb" ) as f: f.write(json.dumps(__snake_case ).encode("utf-8" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
269
1
"""simple docstring""" import json import os from collections import Counter import torch import torchvision import torchvision.transforms as transforms from PIL import Image from torch import nn from torch.utils.data import Dataset __snake_case : Any = {1: (1, 1), 2: (2, 1), 3: (3, 1), 4: (2, 2), 5: (5, 1), 6: (3, 2), 7: (7, 1), 8: (4, 2), 9: (3, 3)} class A__ ( nn.Module ): '''simple docstring''' def __init__( self: Dict , _SCREAMING_SNAKE_CASE: int) -> List[str]: """simple docstring""" super().__init__() __lowerCAmelCase : Any = torchvision.models.resnetaaa(pretrained=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = list(model.children())[:-2] __lowerCAmelCase : str = nn.Sequential(*_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = nn.AdaptiveAvgPoolad(POOLING_BREAKDOWN[args.num_image_embeds]) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: Any) -> Dict: """simple docstring""" __lowerCAmelCase : List[str] = self.pool(self.model(_SCREAMING_SNAKE_CASE)) __lowerCAmelCase : Dict = torch.flatten(_SCREAMING_SNAKE_CASE , start_dim=2) __lowerCAmelCase : Optional[Any] = out.transpose(1 , 2).contiguous() return out # BxNx2048 class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self: int , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: int) -> int: """simple docstring""" __lowerCAmelCase : Any = [json.loads(_SCREAMING_SNAKE_CASE) for l in open(_SCREAMING_SNAKE_CASE)] __lowerCAmelCase : Optional[Any] = os.path.dirname(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = tokenizer __lowerCAmelCase : List[Any] = labels __lowerCAmelCase : Optional[Any] = len(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = max_seq_length __lowerCAmelCase : Tuple = transforms def __len__( self: int) -> Optional[int]: """simple docstring""" return len(self.data) def __getitem__( self: Optional[Any] , _SCREAMING_SNAKE_CASE: str) -> Any: """simple docstring""" __lowerCAmelCase : List[Any] = torch.LongTensor(self.tokenizer.encode(self.data[index]["text"] , add_special_tokens=_SCREAMING_SNAKE_CASE)) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Optional[Any] = sentence[0], sentence[1:-1], sentence[-1] __lowerCAmelCase : Optional[Any] = sentence[: self.max_seq_length] __lowerCAmelCase : Dict = torch.zeros(self.n_classes) __lowerCAmelCase : List[str] = 1 __lowerCAmelCase : Any = Image.open(os.path.join(self.data_dir , self.data[index]["img"])).convert("RGB") __lowerCAmelCase : Dict = self.transforms(_SCREAMING_SNAKE_CASE) return { "image_start_token": start_token, "image_end_token": end_token, "sentence": sentence, "image": image, "label": label, } def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> int: """simple docstring""" __lowerCAmelCase : Optional[Any] = Counter() for row in self.data: label_freqs.update(row["label"]) return label_freqs def _lowercase ( __snake_case ) -> Optional[Any]: __lowerCAmelCase : List[Any] = [len(row["sentence"] ) for row in batch] __lowerCAmelCase , __lowerCAmelCase : List[Any] = len(__snake_case ), max(__snake_case ) __lowerCAmelCase : Dict = torch.zeros(__snake_case ,__snake_case ,dtype=torch.long ) __lowerCAmelCase : Dict = torch.zeros(__snake_case ,__snake_case ,dtype=torch.long ) for i_batch, (input_row, length) in enumerate(zip(__snake_case ,__snake_case ) ): __lowerCAmelCase : List[Any] = input_row["sentence"] __lowerCAmelCase : Dict = 1 __lowerCAmelCase : str = torch.stack([row["image"] for row in batch] ) __lowerCAmelCase : Dict = torch.stack([row["label"] for row in batch] ) __lowerCAmelCase : str = torch.stack([row["image_start_token"] for row in batch] ) __lowerCAmelCase : Optional[Any] = torch.stack([row["image_end_token"] for row in batch] ) return text_tensor, mask_tensor, img_tensor, img_start_token, img_end_token, tgt_tensor def _lowercase ( ) -> List[Any]: return [ "Crime", "Drama", "Thriller", "Action", "Comedy", "Romance", "Documentary", "Short", "Mystery", "History", "Family", "Adventure", "Fantasy", "Sci-Fi", "Western", "Horror", "Sport", "War", "Music", "Musical", "Animation", "Biography", "Film-Noir", ] def _lowercase ( ) -> List[str]: return transforms.Compose( [ transforms.Resize(256 ), transforms.CenterCrop(224 ), transforms.ToTensor(), transforms.Normalize( mean=[0.46777044, 0.44531429, 0.40661017] ,std=[0.12221994, 0.12145835, 0.14380469] ,), ] )
269
"""simple docstring""" import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self: Tuple , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: List[str]=13 , _SCREAMING_SNAKE_CASE: Tuple=7 , _SCREAMING_SNAKE_CASE: int=True , _SCREAMING_SNAKE_CASE: Optional[Any]=True , _SCREAMING_SNAKE_CASE: str=False , _SCREAMING_SNAKE_CASE: Optional[Any]=True , _SCREAMING_SNAKE_CASE: int=99 , _SCREAMING_SNAKE_CASE: int=32 , _SCREAMING_SNAKE_CASE: List[str]=5 , _SCREAMING_SNAKE_CASE: Union[str, Any]=4 , _SCREAMING_SNAKE_CASE: int=64 , _SCREAMING_SNAKE_CASE: List[str]="gelu" , _SCREAMING_SNAKE_CASE: str=0.1 , _SCREAMING_SNAKE_CASE: Any=0.1 , _SCREAMING_SNAKE_CASE: Optional[int]=512 , _SCREAMING_SNAKE_CASE: Tuple=16 , _SCREAMING_SNAKE_CASE: Any=2 , _SCREAMING_SNAKE_CASE: List[str]=0.02 , _SCREAMING_SNAKE_CASE: Tuple=3 , _SCREAMING_SNAKE_CASE: Optional[Any]=4 , _SCREAMING_SNAKE_CASE: int=None , _SCREAMING_SNAKE_CASE: int=2 , _SCREAMING_SNAKE_CASE: str=2 , _SCREAMING_SNAKE_CASE: Union[str, Any]=2 , _SCREAMING_SNAKE_CASE: List[Any]=2 , _SCREAMING_SNAKE_CASE: int=4 , _SCREAMING_SNAKE_CASE: List[str]=1 , ) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : List[str] = parent __lowerCAmelCase : Optional[Any] = batch_size __lowerCAmelCase : Union[str, Any] = seq_length __lowerCAmelCase : Optional[Any] = is_training __lowerCAmelCase : Optional[int] = use_input_mask __lowerCAmelCase : Dict = use_token_type_ids __lowerCAmelCase : Dict = use_labels __lowerCAmelCase : Dict = vocab_size __lowerCAmelCase : Tuple = hidden_size __lowerCAmelCase : List[Any] = num_hidden_layers __lowerCAmelCase : Union[str, Any] = num_attention_heads __lowerCAmelCase : Tuple = intermediate_size __lowerCAmelCase : List[Any] = hidden_act __lowerCAmelCase : Optional[Any] = hidden_dropout_prob __lowerCAmelCase : Optional[Any] = attention_probs_dropout_prob __lowerCAmelCase : Optional[int] = max_position_embeddings __lowerCAmelCase : Union[str, Any] = type_vocab_size __lowerCAmelCase : Optional[int] = type_sequence_label_size __lowerCAmelCase : Dict = initializer_range __lowerCAmelCase : Tuple = num_labels __lowerCAmelCase : Optional[Any] = num_choices __lowerCAmelCase : Union[str, Any] = scope __lowerCAmelCase : Optional[Any] = q_groups __lowerCAmelCase : Optional[int] = k_groups __lowerCAmelCase : Any = v_groups __lowerCAmelCase : int = post_attention_groups __lowerCAmelCase : List[str] = intermediate_groups __lowerCAmelCase : Optional[Any] = output_groups def _SCREAMING_SNAKE_CASE ( self: Dict) -> List[str]: """simple docstring""" __lowerCAmelCase : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) __lowerCAmelCase : Union[str, Any] = None if self.use_input_mask: __lowerCAmelCase : Tuple = random_attention_mask([self.batch_size, self.seq_length]) __lowerCAmelCase : Optional[int] = None __lowerCAmelCase : List[Any] = None __lowerCAmelCase : str = None if self.use_labels: __lowerCAmelCase : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size) __lowerCAmelCase : str = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) __lowerCAmelCase : str = ids_tensor([self.batch_size] , self.num_choices) __lowerCAmelCase : Any = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _SCREAMING_SNAKE_CASE ( self: Tuple) -> int: """simple docstring""" return SqueezeBertConfig( embedding_size=self.hidden_size , vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , attention_probs_dropout_prob=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , q_groups=self.q_groups , k_groups=self.k_groups , v_groups=self.v_groups , post_attention_groups=self.post_attention_groups , intermediate_groups=self.intermediate_groups , output_groups=self.output_groups , ) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Tuple , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Tuple) -> Tuple: """simple docstring""" __lowerCAmelCase : List[Any] = SqueezeBertModel(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : List[Any] = model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = model(_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Tuple) -> Dict: """simple docstring""" __lowerCAmelCase : int = SqueezeBertForMaskedLM(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : Dict = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def _SCREAMING_SNAKE_CASE ( self: Optional[int] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Union[str, Any]) -> int: """simple docstring""" __lowerCAmelCase : str = SqueezeBertForQuestionAnswering(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : Union[str, Any] = model( _SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , start_positions=_SCREAMING_SNAKE_CASE , end_positions=_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: Tuple , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: Tuple) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : List[Any] = self.num_labels __lowerCAmelCase : Union[str, Any] = SqueezeBertForSequenceClassification(_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : int = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[int]) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : Dict = self.num_labels __lowerCAmelCase : Optional[int] = SqueezeBertForTokenClassification(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : List[str] = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: int) -> Tuple: """simple docstring""" __lowerCAmelCase : List[str] = self.num_choices __lowerCAmelCase : str = SqueezeBertForMultipleChoice(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : int = input_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() __lowerCAmelCase : Union[str, Any] = input_mask.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() __lowerCAmelCase : str = model( _SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices)) def _SCREAMING_SNAKE_CASE ( self: str) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[int] = self.prepare_config_and_inputs() ((__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase)) : Union[str, Any] = config_and_inputs __lowerCAmelCase : int = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) SCREAMING_SNAKE_CASE = ( { 'feature-extraction': SqueezeBertModel, 'fill-mask': SqueezeBertForMaskedLM, 'question-answering': SqueezeBertForQuestionAnswering, 'text-classification': SqueezeBertForSequenceClassification, 'token-classification': SqueezeBertForTokenClassification, 'zero-shot': SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = False def _SCREAMING_SNAKE_CASE ( self: Dict) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Any = SqueezeBertModelTester(self) __lowerCAmelCase : Optional[int] = ConfigTester(self , config_class=_SCREAMING_SNAKE_CASE , dim=37) def _SCREAMING_SNAKE_CASE ( self: Any) -> Tuple: """simple docstring""" self.config_tester.run_common_tests() def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> Any: """simple docstring""" __lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> List[Any]: """simple docstring""" __lowerCAmelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> int: """simple docstring""" __lowerCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Any) -> int: """simple docstring""" __lowerCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> str: """simple docstring""" __lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*_SCREAMING_SNAKE_CASE) @slow def _SCREAMING_SNAKE_CASE ( self: Any) -> Dict: """simple docstring""" for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase : Optional[Any] = SqueezeBertModel.from_pretrained(_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) @require_sentencepiece @require_tokenizers @require_torch class A__ ( unittest.TestCase ): '''simple docstring''' @slow def _SCREAMING_SNAKE_CASE ( self: int) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[Any] = SqueezeBertForSequenceClassification.from_pretrained("squeezebert/squeezebert-mnli") __lowerCAmelCase : List[Any] = torch.tensor([[1, 2_9414, 232, 328, 740, 1140, 1_2695, 69, 13, 1588, 2]]) __lowerCAmelCase : List[Any] = model(_SCREAMING_SNAKE_CASE)[0] __lowerCAmelCase : Any = torch.Size((1, 3)) self.assertEqual(output.shape , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = torch.tensor([[0.6401, -0.0349, -0.6041]]) self.assertTrue(torch.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=1e-4))
269
1
"""simple docstring""" from __future__ import annotations from collections.abc import Generator def _lowercase ( ) -> Generator[int, None, None]: __lowerCAmelCase : dict[int, int] = {} __lowerCAmelCase : List[Any] = 2 while True: __lowerCAmelCase : int = factor_map.pop(__snake_case ,__snake_case ) if factor: __lowerCAmelCase : Union[str, Any] = factor + prime while x in factor_map: x += factor __lowerCAmelCase : str = factor else: __lowerCAmelCase : List[str] = prime yield prime prime += 1 def _lowercase ( __snake_case = 1e1_0 ) -> int: __lowerCAmelCase : List[Any] = sieve() __lowerCAmelCase : List[Any] = 1 while True: __lowerCAmelCase : Union[str, Any] = next(__snake_case ) if (2 * prime * n) > limit: return n # Ignore the next prime as the reminder will be 2. next(__snake_case ) n += 2 if __name__ == "__main__": print(solution())
269
"""simple docstring""" import os from math import logaa def _lowercase ( __snake_case = "base_exp.txt" ) -> int: __lowerCAmelCase : float = 0 __lowerCAmelCase : Any = 0 for i, line in enumerate(open(os.path.join(os.path.dirname(__snake_case ) ,__snake_case ) ) ): __lowerCAmelCase , __lowerCAmelCase : List[str] = list(map(__snake_case ,line.split("," ) ) ) if x * logaa(__snake_case ) > largest: __lowerCAmelCase : Tuple = x * logaa(__snake_case ) __lowerCAmelCase : Optional[Any] = i + 1 return result if __name__ == "__main__": print(solution())
269
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_owlvit import OwlViTImageProcessor __snake_case : Union[str, Any] = logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self: Tuple , *_SCREAMING_SNAKE_CASE: Optional[Any] , **_SCREAMING_SNAKE_CASE: Union[str, Any]) -> None: """simple docstring""" warnings.warn( "The class OwlViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use OwlViTImageProcessor instead." , _SCREAMING_SNAKE_CASE , ) super().__init__(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE)
269
"""simple docstring""" import collections import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_flax_cross_test, require_flax, require_torch, require_vision, slow, torch_device, ) from transformers.utils import is_flax_available, is_torch_available, is_vision_available from ...test_modeling_flax_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_flax_bert import FlaxBertModelTester from ..clip.test_modeling_flax_clip import FlaxCLIPVisionModelTester from ..vit.test_modeling_flax_vit import FlaxViTModelTester if is_flax_available(): from transformers import ( FlaxBertModel, FlaxCLIPVisionModel, FlaxVisionTextDualEncoderModel, FlaxViTModel, VisionTextDualEncoderConfig, VisionTextDualEncoderProcessor, ) from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) if is_torch_available(): import torch from transformers import VisionTextDualEncoderModel if is_vision_available(): from PIL import Image def _lowercase ( __snake_case ) -> List[str]: if isinstance(__snake_case ,collections.abc.Iterable ): return x return (x, x) @require_flax class A__ : '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Optional[Any]) -> str: """simple docstring""" pass def _SCREAMING_SNAKE_CASE ( self: str) -> int: """simple docstring""" pass def _SCREAMING_SNAKE_CASE ( self: List[str]) -> Tuple: """simple docstring""" pass def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: np.ndarray , _SCREAMING_SNAKE_CASE: np.ndarray , _SCREAMING_SNAKE_CASE: float) -> List[Any]: """simple docstring""" __lowerCAmelCase : str = np.abs((a - b)).max() self.assertLessEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , F"""Difference between torch and flax is {diff} (>= {tol}).""") def _SCREAMING_SNAKE_CASE ( self: str , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Any=None , **_SCREAMING_SNAKE_CASE: Tuple) -> List[str]: """simple docstring""" __lowerCAmelCase : Tuple = VisionTextDualEncoderConfig.from_vision_text_configs(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = FlaxVisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = model(input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], config.projection_dim)) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], config.projection_dim)) def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Optional[Any]=None , **_SCREAMING_SNAKE_CASE: List[str]) -> Dict: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : List[Any] = self.get_vision_text_model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = {"vision_model": vision_model, "text_model": text_model} __lowerCAmelCase : Optional[Any] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = model(input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], model.config.projection_dim)) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], model.config.projection_dim)) def _SCREAMING_SNAKE_CASE ( self: Dict , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: List[Any]=None , **_SCREAMING_SNAKE_CASE: Tuple) -> str: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : Optional[Any] = self.get_vision_text_model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = {"vision_model": vision_model, "text_model": text_model} __lowerCAmelCase : Any = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = model(input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = output[0] with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = FlaxVisionTextDualEncoderModel.from_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = model(input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = after_output[0] __lowerCAmelCase : Any = np.amax(np.abs(out_a - out_a)) self.assertLessEqual(_SCREAMING_SNAKE_CASE , 1e-3) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Any=None , **_SCREAMING_SNAKE_CASE: List[str]) -> Optional[int]: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : Any = self.get_vision_text_model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = {"vision_model": vision_model, "text_model": text_model} __lowerCAmelCase : Optional[Any] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = model( input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , output_attentions=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = output.vision_model_output.attentions self.assertEqual(len(_SCREAMING_SNAKE_CASE) , vision_config.num_hidden_layers) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) __lowerCAmelCase : List[str] = to_atuple(vision_model.config.image_size) __lowerCAmelCase : Any = to_atuple(vision_model.config.patch_size) __lowerCAmelCase : Union[str, Any] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) __lowerCAmelCase : Tuple = num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len)) __lowerCAmelCase : Union[str, Any] = output.text_model_output.attentions self.assertEqual(len(_SCREAMING_SNAKE_CASE) , text_config.num_hidden_layers) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: int) -> str: """simple docstring""" pt_model.to(_SCREAMING_SNAKE_CASE) pt_model.eval() # prepare inputs __lowerCAmelCase : Union[str, Any] = inputs_dict __lowerCAmelCase : Union[str, Any] = {k: torch.tensor(v.tolist()) for k, v in flax_inputs.items()} with torch.no_grad(): __lowerCAmelCase : Any = pt_model(**_SCREAMING_SNAKE_CASE).to_tuple() __lowerCAmelCase : List[Any] = fx_model(**_SCREAMING_SNAKE_CASE).to_tuple() self.assertEqual(len(_SCREAMING_SNAKE_CASE) , len(_SCREAMING_SNAKE_CASE) , "Output lengths differ between Flax and PyTorch") for fx_output, pt_output in zip(fx_outputs[:4] , pt_outputs[:4]): self.assert_almost_equals(_SCREAMING_SNAKE_CASE , pt_output.numpy() , 4e-2) # PT -> Flax with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = FlaxVisionTextDualEncoderModel.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = fx_model_loaded(**_SCREAMING_SNAKE_CASE).to_tuple() self.assertEqual(len(_SCREAMING_SNAKE_CASE) , len(_SCREAMING_SNAKE_CASE) , "Output lengths differ between Flax and PyTorch") for fx_output_loaded, pt_output in zip(fx_outputs_loaded[:4] , pt_outputs[:4]): self.assert_almost_equals(_SCREAMING_SNAKE_CASE , pt_output.numpy() , 4e-2) # Flax -> PT with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = VisionTextDualEncoderModel.from_pretrained(_SCREAMING_SNAKE_CASE , from_flax=_SCREAMING_SNAKE_CASE) pt_model_loaded.to(_SCREAMING_SNAKE_CASE) pt_model_loaded.eval() with torch.no_grad(): __lowerCAmelCase : Optional[Any] = pt_model_loaded(**_SCREAMING_SNAKE_CASE).to_tuple() self.assertEqual(len(_SCREAMING_SNAKE_CASE) , len(_SCREAMING_SNAKE_CASE) , "Output lengths differ between Flax and PyTorch") for fx_output, pt_output_loaded in zip(fx_outputs[:4] , pt_outputs_loaded[:4]): self.assert_almost_equals(_SCREAMING_SNAKE_CASE , pt_output_loaded.numpy() , 4e-2) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Dict) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Union[str, Any] = VisionTextDualEncoderConfig.from_vision_text_configs(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = VisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = FlaxVisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = fx_state self.check_pt_flax_equivalence(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Tuple , _SCREAMING_SNAKE_CASE: Dict) -> str: """simple docstring""" __lowerCAmelCase : Optional[int] = VisionTextDualEncoderConfig.from_vision_text_configs(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = VisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = FlaxVisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = load_flax_weights_in_pytorch_model(_SCREAMING_SNAKE_CASE , fx_model.params) self.check_pt_flax_equivalence(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> str: """simple docstring""" __lowerCAmelCase : List[str] = self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Dict) -> int: """simple docstring""" __lowerCAmelCase : List[Any] = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> List[Any]: """simple docstring""" __lowerCAmelCase : Dict = self.prepare_config_and_inputs() self.check_save_load(**_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: int) -> Dict: """simple docstring""" __lowerCAmelCase : List[str] = self.prepare_config_and_inputs() self.check_vision_text_output_attention(**_SCREAMING_SNAKE_CASE) @is_pt_flax_cross_test def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> Any: """simple docstring""" __lowerCAmelCase : Dict = self.prepare_config_and_inputs() __lowerCAmelCase : List[Any] = config_inputs_dict.pop("vision_config") __lowerCAmelCase : str = config_inputs_dict.pop("text_config") __lowerCAmelCase : Union[str, Any] = config_inputs_dict self.check_equivalence_pt_to_flax(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) self.check_equivalence_flax_to_pt(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) @slow def _SCREAMING_SNAKE_CASE ( self: str) -> Dict: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : Dict = self.get_pretrained_model_and_inputs() __lowerCAmelCase : Union[str, Any] = model_a(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = outputs[0] with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = FlaxVisionTextDualEncoderModel.from_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = model_a(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = after_outputs[0] __lowerCAmelCase : List[Any] = np.amax(np.abs(out_a - out_a)) self.assertLessEqual(_SCREAMING_SNAKE_CASE , 1e-5) @require_flax class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: List[str]) -> List[Any]: """simple docstring""" __lowerCAmelCase : Any = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( "hf-internal-testing/tiny-random-vit" , "hf-internal-testing/tiny-bert" , vision_from_pt=_SCREAMING_SNAKE_CASE , text_from_pt=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Union[str, Any] = 13 __lowerCAmelCase : Optional[int] = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ]) __lowerCAmelCase : List[Any] = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size) __lowerCAmelCase : List[Any] = random_attention_mask([batch_size, 4]) __lowerCAmelCase : str = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def _SCREAMING_SNAKE_CASE ( self: int , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Union[str, Any]) -> Optional[int]: """simple docstring""" __lowerCAmelCase : List[str] = FlaxViTModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = FlaxBertModel(_SCREAMING_SNAKE_CASE) return vision_model, text_model def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> int: """simple docstring""" __lowerCAmelCase : List[Any] = FlaxViTModelTester(self) __lowerCAmelCase : Optional[Any] = FlaxBertModelTester(self) __lowerCAmelCase : int = vit_model_tester.prepare_config_and_inputs() __lowerCAmelCase : List[str] = bert_model_tester.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase : Tuple = vision_config_and_inputs __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Union[str, Any] = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_torch class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Dict) -> List[str]: """simple docstring""" __lowerCAmelCase : Optional[int] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( "hf-internal-testing/tiny-random-clip" , "hf-internal-testing/tiny-bert" , vision_from_pt=_SCREAMING_SNAKE_CASE , text_from_pt=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Optional[int] = 13 __lowerCAmelCase : List[str] = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ]) __lowerCAmelCase : Any = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size) __lowerCAmelCase : str = random_attention_mask([batch_size, 4]) __lowerCAmelCase : Optional[int] = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Union[str, Any]) -> int: """simple docstring""" __lowerCAmelCase : int = FlaxCLIPVisionModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = FlaxBertModel(_SCREAMING_SNAKE_CASE) return vision_model, text_model def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : List[Any] = FlaxCLIPVisionModelTester(self) __lowerCAmelCase : str = FlaxBertModelTester(self) __lowerCAmelCase : Optional[Any] = clip_model_tester.prepare_config_and_inputs() __lowerCAmelCase : Dict = bert_model_tester.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase : Any = vision_config_and_inputs __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : List[Any] = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_flax @require_vision class A__ ( unittest.TestCase ): '''simple docstring''' @slow def _SCREAMING_SNAKE_CASE ( self: List[str]) -> List[str]: """simple docstring""" __lowerCAmelCase : Dict = FlaxVisionTextDualEncoderModel.from_pretrained("clip-italian/clip-italian" , logit_scale_init_value=1.0) __lowerCAmelCase : str = VisionTextDualEncoderProcessor.from_pretrained("clip-italian/clip-italian") __lowerCAmelCase : Tuple = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png") __lowerCAmelCase : Optional[int] = processor( text=["una foto di un gatto", "una foto di un cane"] , images=_SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE , return_tensors="np") __lowerCAmelCase : List[str] = model(**_SCREAMING_SNAKE_CASE) # verify the logits self.assertEqual(outputs.logits_per_image.shape , (inputs.pixel_values.shape[0], inputs.input_ids.shape[0])) self.assertEqual( outputs.logits_per_text.shape , (inputs.input_ids.shape[0], inputs.pixel_values.shape[0]) , ) __lowerCAmelCase : List[str] = np.array([[1.228_4727, 0.310_4122]]) self.assertTrue(np.allclose(outputs.logits_per_image , _SCREAMING_SNAKE_CASE , atol=1e-3))
269
1
"""simple docstring""" from __future__ import annotations from math import pi def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> dict[str, float]: if (inductance, frequency, reactance).count(0 ) != 1: raise ValueError("One and only one argument must be 0" ) if inductance < 0: raise ValueError("Inductance cannot be negative" ) if frequency < 0: raise ValueError("Frequency cannot be negative" ) if reactance < 0: raise ValueError("Inductive reactance cannot be negative" ) if inductance == 0: return {"inductance": reactance / (2 * pi * frequency)} elif frequency == 0: return {"frequency": reactance / (2 * pi * inductance)} elif reactance == 0: return {"reactance": 2 * pi * frequency * inductance} else: raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
269
"""simple docstring""" from __future__ import annotations def _lowercase ( __snake_case ,__snake_case ,__snake_case ,__snake_case ) -> list: __lowerCAmelCase : Dict = [] __lowerCAmelCase , __lowerCAmelCase : Any = input_list[low:mid], input_list[mid : high + 1] while left and right: result.append((left if left[0] <= right[0] else right).pop(0 ) ) __lowerCAmelCase : int = result + left + right return input_list def _lowercase ( __snake_case ) -> list: if len(__snake_case ) <= 1: return input_list __lowerCAmelCase : int = list(__snake_case ) # iteration for two-way merging __lowerCAmelCase : Optional[int] = 2 while p <= len(__snake_case ): # getting low, high and middle value for merge-sort of single list for i in range(0 ,len(__snake_case ) ,__snake_case ): __lowerCAmelCase : Union[str, Any] = i __lowerCAmelCase : Tuple = i + p - 1 __lowerCAmelCase : Optional[Any] = (low + high + 1) // 2 __lowerCAmelCase : Any = merge(__snake_case ,__snake_case ,__snake_case ,__snake_case ) # final merge of last two parts if p * 2 >= len(__snake_case ): __lowerCAmelCase : Optional[Any] = i __lowerCAmelCase : Union[str, Any] = merge(__snake_case ,0 ,__snake_case ,len(__snake_case ) - 1 ) break p *= 2 return input_list if __name__ == "__main__": __snake_case : Union[str, Any] = input('Enter numbers separated by a comma:\n').strip() if user_input == "": __snake_case : Optional[int] = [] else: __snake_case : int = [int(item.strip()) for item in user_input.split(',')] print(iter_merge_sort(unsorted))
269
1
"""simple docstring""" import json import sys def _lowercase ( __snake_case ,__snake_case ) -> str: with open(__snake_case ,encoding="utf-8" ) as f: __lowerCAmelCase : Optional[int] = json.load(__snake_case ) __lowerCAmelCase : str = ["<details>", "<summary>Show updated benchmarks!</summary>", " "] for benchmark_name in sorted(__snake_case ): __lowerCAmelCase : Union[str, Any] = results[benchmark_name] __lowerCAmelCase : Optional[int] = benchmark_name.split("/" )[-1] output_md.append(F"""### Benchmark: {benchmark_file_name}""" ) __lowerCAmelCase : str = "| metric |" __lowerCAmelCase : str = "|--------|" __lowerCAmelCase : Union[str, Any] = "| new / old (diff) |" for metric_name in sorted(__snake_case ): __lowerCAmelCase : Dict = benchmark_res[metric_name] __lowerCAmelCase : Any = metric_vals["new"] __lowerCAmelCase : List[Any] = metric_vals.get("old" ,__snake_case ) __lowerCAmelCase : Tuple = metric_vals.get("diff" ,__snake_case ) __lowerCAmelCase : List[Any] = F""" {new_val:f}""" if isinstance(__snake_case ,(int, float) ) else "None" if old_val is not None: val_str += F""" / {old_val:f}""" if isinstance(__snake_case ,(int, float) ) else "None" if dif_val is not None: val_str += F""" ({dif_val:f})""" if isinstance(__snake_case ,(int, float) ) else "None" title += " " + metric_name + " |" lines += "---|" value += val_str + " |" output_md += [title, lines, value, " "] output_md.append("</details>" ) with open(__snake_case ,"w" ,encoding="utf-8" ) as f: f.writelines("\n".join(__snake_case ) ) if __name__ == "__main__": __snake_case : int = sys.argv[1] __snake_case : Dict = sys.argv[2] format_json_to_md(input_json_file, output_md_file)
269
"""simple docstring""" import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class A__ ( unittest.TestCase ): '''simple docstring''' @slow def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> str: """simple docstring""" __lowerCAmelCase : Optional[Any] = FlaxMTaForConditionalGeneration.from_pretrained("google/mt5-small") __lowerCAmelCase : Optional[Any] = AutoTokenizer.from_pretrained("google/mt5-small") __lowerCAmelCase : Tuple = tokenizer("Hello there" , return_tensors="np").input_ids __lowerCAmelCase : Dict = tokenizer("Hi I am" , return_tensors="np").input_ids __lowerCAmelCase : str = shift_tokens_right(_SCREAMING_SNAKE_CASE , model.config.pad_token_id , model.config.decoder_start_token_id) __lowerCAmelCase : Optional[int] = model(_SCREAMING_SNAKE_CASE , decoder_input_ids=_SCREAMING_SNAKE_CASE).logits __lowerCAmelCase : int = optax.softmax_cross_entropy(_SCREAMING_SNAKE_CASE , onehot(_SCREAMING_SNAKE_CASE , logits.shape[-1])).mean() __lowerCAmelCase : List[str] = -(labels.shape[-1] * loss.item()) __lowerCAmelCase : str = -84.9127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE) < 1e-4)
269
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices __snake_case : Optional[int] = logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'maskformer-swin' SCREAMING_SNAKE_CASE = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self: Optional[int] , _SCREAMING_SNAKE_CASE: int=224 , _SCREAMING_SNAKE_CASE: Tuple=4 , _SCREAMING_SNAKE_CASE: int=3 , _SCREAMING_SNAKE_CASE: List[Any]=96 , _SCREAMING_SNAKE_CASE: Union[str, Any]=[2, 2, 6, 2] , _SCREAMING_SNAKE_CASE: Any=[3, 6, 12, 24] , _SCREAMING_SNAKE_CASE: List[str]=7 , _SCREAMING_SNAKE_CASE: List[str]=4.0 , _SCREAMING_SNAKE_CASE: Optional[int]=True , _SCREAMING_SNAKE_CASE: Tuple=0.0 , _SCREAMING_SNAKE_CASE: Any=0.0 , _SCREAMING_SNAKE_CASE: Any=0.1 , _SCREAMING_SNAKE_CASE: str="gelu" , _SCREAMING_SNAKE_CASE: Tuple=False , _SCREAMING_SNAKE_CASE: Union[str, Any]=0.02 , _SCREAMING_SNAKE_CASE: str=1e-5 , _SCREAMING_SNAKE_CASE: Optional[int]=None , _SCREAMING_SNAKE_CASE: str=None , **_SCREAMING_SNAKE_CASE: Union[str, Any] , ) -> List[str]: """simple docstring""" super().__init__(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = image_size __lowerCAmelCase : Any = patch_size __lowerCAmelCase : Tuple = num_channels __lowerCAmelCase : Any = embed_dim __lowerCAmelCase : Any = depths __lowerCAmelCase : Dict = len(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = num_heads __lowerCAmelCase : Tuple = window_size __lowerCAmelCase : Dict = mlp_ratio __lowerCAmelCase : Any = qkv_bias __lowerCAmelCase : Union[str, Any] = hidden_dropout_prob __lowerCAmelCase : int = attention_probs_dropout_prob __lowerCAmelCase : Tuple = drop_path_rate __lowerCAmelCase : int = hidden_act __lowerCAmelCase : Optional[int] = use_absolute_embeddings __lowerCAmelCase : List[str] = layer_norm_eps __lowerCAmelCase : Any = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model __lowerCAmelCase : Optional[Any] = int(embed_dim * 2 ** (len(_SCREAMING_SNAKE_CASE) - 1)) __lowerCAmelCase : Any = ["stem"] + [F"""stage{idx}""" for idx in range(1 , len(_SCREAMING_SNAKE_CASE) + 1)] __lowerCAmelCase , __lowerCAmelCase : List[str] = get_aligned_output_features_output_indices( out_features=_SCREAMING_SNAKE_CASE , out_indices=_SCREAMING_SNAKE_CASE , stage_names=self.stage_names)
269
"""simple docstring""" import re def _lowercase ( __snake_case ) -> str: if len(re.findall("[ATCG]" ,__snake_case ) ) != len(__snake_case ): raise ValueError("Invalid Strand" ) return dna.translate(dna.maketrans("ATCG" ,"TAGC" ) ) if __name__ == "__main__": import doctest doctest.testmod()
269
1
"""simple docstring""" import glob import os import random from string import ascii_lowercase, digits import cva __snake_case : Optional[Any] = '' __snake_case : List[str] = '' __snake_case : Dict = '' __snake_case : List[str] = 1 # (0 is vertical, 1 is horizontal) def _lowercase ( ) -> None: __lowerCAmelCase , __lowerCAmelCase : Union[str, Any] = get_dataset(__snake_case ,__snake_case ) print("Processing..." ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Optional[Any] = update_image_and_anno(__snake_case ,__snake_case ,__snake_case ) for index, image in enumerate(__snake_case ): # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' __lowerCAmelCase : Union[str, Any] = random_chars(32 ) __lowerCAmelCase : Optional[Any] = paths[index].split(os.sep )[-1].rsplit("." ,1 )[0] __lowerCAmelCase : List[Any] = F"""{OUTPUT_DIR}/{file_name}_FLIP_{letter_code}""" cva.imwrite(F"""/{file_root}.jpg""" ,__snake_case ,[cva.IMWRITE_JPEG_QUALITY, 85] ) print(F"""Success {index+1}/{len(__snake_case )} with {file_name}""" ) __lowerCAmelCase : Union[str, Any] = [] for anno in new_annos[index]: __lowerCAmelCase : List[str] = F"""{anno[0]} {anno[1]} {anno[2]} {anno[3]} {anno[4]}""" annos_list.append(__snake_case ) with open(F"""/{file_root}.txt""" ,"w" ) as outfile: outfile.write("\n".join(line for line in annos_list ) ) def _lowercase ( __snake_case ,__snake_case ) -> tuple[list, list]: __lowerCAmelCase : List[Any] = [] __lowerCAmelCase : List[str] = [] for label_file in glob.glob(os.path.join(__snake_case ,"*.txt" ) ): __lowerCAmelCase : Optional[int] = label_file.split(os.sep )[-1].rsplit("." ,1 )[0] with open(__snake_case ) as in_file: __lowerCAmelCase : Any = in_file.readlines() __lowerCAmelCase : Optional[Any] = os.path.join(__snake_case ,F"""{label_name}.jpg""" ) __lowerCAmelCase : Optional[int] = [] for obj_list in obj_lists: __lowerCAmelCase : List[Any] = obj_list.rstrip("\n" ).split(" " ) boxes.append( [ int(obj[0] ), float(obj[1] ), float(obj[2] ), float(obj[3] ), float(obj[4] ), ] ) if not boxes: continue img_paths.append(__snake_case ) labels.append(__snake_case ) return img_paths, labels def _lowercase ( __snake_case ,__snake_case ,__snake_case = 1 ) -> tuple[list, list, list]: __lowerCAmelCase : Tuple = [] __lowerCAmelCase : Union[str, Any] = [] __lowerCAmelCase : List[str] = [] for idx in range(len(__snake_case ) ): __lowerCAmelCase : Tuple = [] __lowerCAmelCase : int = img_list[idx] path_list.append(__snake_case ) __lowerCAmelCase : Union[str, Any] = anno_list[idx] __lowerCAmelCase : List[str] = cva.imread(__snake_case ) if flip_type == 1: __lowerCAmelCase : str = cva.flip(__snake_case ,__snake_case ) for bbox in img_annos: __lowerCAmelCase : List[str] = 1 - bbox[1] new_annos.append([bbox[0], x_center_new, bbox[2], bbox[3], bbox[4]] ) elif flip_type == 0: __lowerCAmelCase : str = cva.flip(__snake_case ,__snake_case ) for bbox in img_annos: __lowerCAmelCase : Any = 1 - bbox[2] new_annos.append([bbox[0], bbox[1], y_center_new, bbox[3], bbox[4]] ) new_annos_lists.append(__snake_case ) new_imgs_list.append(__snake_case ) return new_imgs_list, new_annos_lists, path_list def _lowercase ( __snake_case = 32 ) -> str: assert number_char > 1, "The number of character should greater than 1" __lowerCAmelCase : Union[str, Any] = ascii_lowercase + digits return "".join(random.choice(__snake_case ) for _ in range(__snake_case ) ) if __name__ == "__main__": main() print('DONE ✅')
269
"""simple docstring""" import os import unittest from transformers import FunnelTokenizer, FunnelTokenizerFast from transformers.models.funnel.tokenization_funnel import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = FunnelTokenizer SCREAMING_SNAKE_CASE = FunnelTokenizerFast SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = True def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Optional[int]: """simple docstring""" super().setUp() __lowerCAmelCase : str = [ "<unk>", "<cls>", "<sep>", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] __lowerCAmelCase : int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"]) with open(self.vocab_file , "w" , encoding="utf-8") as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens])) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , **_SCREAMING_SNAKE_CASE: Union[str, Any]) -> int: """simple docstring""" return FunnelTokenizer.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Any , **_SCREAMING_SNAKE_CASE: Any) -> str: """simple docstring""" return FunnelTokenizerFast.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: str) -> Any: """simple docstring""" __lowerCAmelCase : Union[str, Any] = "UNwant\u00E9d,running" __lowerCAmelCase : str = "unwanted, running" return input_text, output_text def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> List[str]: """simple docstring""" __lowerCAmelCase : Any = self.tokenizer_class(self.vocab_file) __lowerCAmelCase : Any = tokenizer.tokenize("UNwant\u00E9d,running") self.assertListEqual(_SCREAMING_SNAKE_CASE , ["un", "##want", "##ed", ",", "runn", "##ing"]) self.assertListEqual(tokenizer.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE) , [7, 4, 5, 10, 8, 9]) def _SCREAMING_SNAKE_CASE ( self: List[str]) -> List[str]: """simple docstring""" __lowerCAmelCase : Optional[Any] = self.get_tokenizers(do_lower_case=_SCREAMING_SNAKE_CASE) for tokenizer in tokenizers: __lowerCAmelCase : List[str] = tokenizer("UNwant\u00E9d,running") __lowerCAmelCase : Optional[int] = len(inputs["input_ids"]) - 1 self.assertListEqual(inputs["token_type_ids"] , [2] + [0] * sentence_len) __lowerCAmelCase : List[str] = tokenizer("UNwant\u00E9d,running" , "UNwant\u00E9d,running") self.assertListEqual(inputs["token_type_ids"] , [2] + [0] * sentence_len + [1] * sentence_len)
269
1
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import ( DiffusionPipeline, UnCLIPImageVariationPipeline, UnCLIPScheduler, UNetaDConditionModel, UNetaDModel, ) from diffusers.pipelines.unclip.text_proj import UnCLIPTextProjModel from diffusers.utils import floats_tensor, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, load_image, require_torch_gpu, skip_mps from ..pipeline_params import IMAGE_VARIATION_BATCH_PARAMS, IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = UnCLIPImageVariationPipeline SCREAMING_SNAKE_CASE = IMAGE_VARIATION_PARAMS - {'height', 'width', 'guidance_scale'} SCREAMING_SNAKE_CASE = IMAGE_VARIATION_BATCH_PARAMS SCREAMING_SNAKE_CASE = [ 'generator', 'return_dict', 'decoder_num_inference_steps', 'super_res_num_inference_steps', ] SCREAMING_SNAKE_CASE = False @property def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> Union[str, Any]: """simple docstring""" return 32 @property def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> Any: """simple docstring""" return 32 @property def _SCREAMING_SNAKE_CASE ( self: Any) -> Tuple: """simple docstring""" return self.time_input_dim @property def _SCREAMING_SNAKE_CASE ( self: str) -> int: """simple docstring""" return self.time_input_dim * 4 @property def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Tuple: """simple docstring""" return 100 @property def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> Dict: """simple docstring""" __lowerCAmelCase : str = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip") return tokenizer @property def _SCREAMING_SNAKE_CASE ( self: List[str]) -> Tuple: """simple docstring""" torch.manual_seed(0) __lowerCAmelCase : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModelWithProjection(_SCREAMING_SNAKE_CASE) @property def _SCREAMING_SNAKE_CASE ( self: Tuple) -> str: """simple docstring""" torch.manual_seed(0) __lowerCAmelCase : List[str] = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , num_hidden_layers=5 , num_attention_heads=4 , image_size=32 , intermediate_size=37 , patch_size=1 , ) return CLIPVisionModelWithProjection(_SCREAMING_SNAKE_CASE) @property def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> Union[str, Any]: """simple docstring""" torch.manual_seed(0) __lowerCAmelCase : List[Any] = { "clip_embeddings_dim": self.text_embedder_hidden_size, "time_embed_dim": self.time_embed_dim, "cross_attention_dim": self.cross_attention_dim, } __lowerCAmelCase : Tuple = UnCLIPTextProjModel(**_SCREAMING_SNAKE_CASE) return model @property def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> Dict: """simple docstring""" torch.manual_seed(0) __lowerCAmelCase : Optional[Any] = { "sample_size": 32, # RGB in channels "in_channels": 3, # Out channels is double in channels because predicts mean and variance "out_channels": 6, "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": "identity", } __lowerCAmelCase : str = UNetaDConditionModel(**_SCREAMING_SNAKE_CASE) return model @property def _SCREAMING_SNAKE_CASE ( self: str) -> Optional[int]: """simple docstring""" return { "sample_size": 64, "layers_per_block": 1, "down_block_types": ("ResnetDownsampleBlock2D", "ResnetDownsampleBlock2D"), "up_block_types": ("ResnetUpsampleBlock2D", "ResnetUpsampleBlock2D"), "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "in_channels": 6, "out_channels": 3, } @property def _SCREAMING_SNAKE_CASE ( self: Any) -> Tuple: """simple docstring""" torch.manual_seed(0) __lowerCAmelCase : Optional[Any] = UNetaDModel(**self.dummy_super_res_kwargs) return model @property def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Optional[Any]: """simple docstring""" torch.manual_seed(1) __lowerCAmelCase : List[str] = UNetaDModel(**self.dummy_super_res_kwargs) return model def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Optional[int]: """simple docstring""" __lowerCAmelCase : List[str] = self.dummy_decoder __lowerCAmelCase : str = self.dummy_text_proj __lowerCAmelCase : List[Any] = self.dummy_text_encoder __lowerCAmelCase : int = self.dummy_tokenizer __lowerCAmelCase : Union[str, Any] = self.dummy_super_res_first __lowerCAmelCase : str = self.dummy_super_res_last __lowerCAmelCase : Union[str, Any] = UnCLIPScheduler( variance_type="learned_range" , prediction_type="epsilon" , num_train_timesteps=1000 , ) __lowerCAmelCase : str = UnCLIPScheduler( variance_type="fixed_small_log" , prediction_type="epsilon" , num_train_timesteps=1000 , ) __lowerCAmelCase : List[str] = CLIPImageProcessor(crop_size=32 , size=32) __lowerCAmelCase : List[str] = self.dummy_image_encoder return { "decoder": decoder, "text_encoder": text_encoder, "tokenizer": tokenizer, "text_proj": text_proj, "feature_extractor": feature_extractor, "image_encoder": image_encoder, "super_res_first": super_res_first, "super_res_last": super_res_last, "decoder_scheduler": decoder_scheduler, "super_res_scheduler": super_res_scheduler, } def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: str=0 , _SCREAMING_SNAKE_CASE: int=True) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : Tuple = floats_tensor((1, 3, 32, 32) , rng=random.Random(_SCREAMING_SNAKE_CASE)).to(_SCREAMING_SNAKE_CASE) if str(_SCREAMING_SNAKE_CASE).startswith("mps"): __lowerCAmelCase : List[Any] = torch.manual_seed(_SCREAMING_SNAKE_CASE) else: __lowerCAmelCase : str = torch.Generator(device=_SCREAMING_SNAKE_CASE).manual_seed(_SCREAMING_SNAKE_CASE) if pil_image: __lowerCAmelCase : str = input_image * 0.5 + 0.5 __lowerCAmelCase : Optional[int] = input_image.clamp(0 , 1) __lowerCAmelCase : int = input_image.cpu().permute(0 , 2 , 3 , 1).float().numpy() __lowerCAmelCase : Optional[int] = DiffusionPipeline.numpy_to_pil(_SCREAMING_SNAKE_CASE)[0] return { "image": input_image, "generator": generator, "decoder_num_inference_steps": 2, "super_res_num_inference_steps": 2, "output_type": "np", } def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> int: """simple docstring""" __lowerCAmelCase : List[str] = "cpu" __lowerCAmelCase : int = self.get_dummy_components() __lowerCAmelCase : int = self.pipeline_class(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = pipe.to(_SCREAMING_SNAKE_CASE) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = self.get_dummy_inputs(_SCREAMING_SNAKE_CASE , pil_image=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = pipe(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = output.images __lowerCAmelCase : Tuple = self.get_dummy_inputs(_SCREAMING_SNAKE_CASE , pil_image=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = pipe( **_SCREAMING_SNAKE_CASE , return_dict=_SCREAMING_SNAKE_CASE , )[0] __lowerCAmelCase : Optional[Any] = image[0, -3:, -3:, -1] __lowerCAmelCase : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) __lowerCAmelCase : Optional[int] = np.array( [ 0.9997, 0.0002, 0.9997, 0.9997, 0.9969, 0.0023, 0.9997, 0.9969, 0.9970, ]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 def _SCREAMING_SNAKE_CASE ( self: Dict) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Optional[Any] = "cpu" __lowerCAmelCase : List[str] = self.get_dummy_components() __lowerCAmelCase : Dict = self.pipeline_class(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = pipe.to(_SCREAMING_SNAKE_CASE) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = self.get_dummy_inputs(_SCREAMING_SNAKE_CASE , pil_image=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = pipe(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = output.images __lowerCAmelCase : Union[str, Any] = self.get_dummy_inputs(_SCREAMING_SNAKE_CASE , pil_image=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = pipe( **_SCREAMING_SNAKE_CASE , return_dict=_SCREAMING_SNAKE_CASE , )[0] __lowerCAmelCase : Optional[Any] = image[0, -3:, -3:, -1] __lowerCAmelCase : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) __lowerCAmelCase : Dict = np.array([0.9997, 0.0003, 0.9997, 0.9997, 0.9970, 0.0024, 0.9997, 0.9971, 0.9971]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : Tuple = "cpu" __lowerCAmelCase : List[Any] = self.get_dummy_components() __lowerCAmelCase : List[Any] = self.pipeline_class(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = pipe.to(_SCREAMING_SNAKE_CASE) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = self.get_dummy_inputs(_SCREAMING_SNAKE_CASE , pil_image=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = [ pipeline_inputs["image"], pipeline_inputs["image"], ] __lowerCAmelCase : Optional[int] = pipe(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = output.images __lowerCAmelCase : Dict = self.get_dummy_inputs(_SCREAMING_SNAKE_CASE , pil_image=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = [ tuple_pipeline_inputs["image"], tuple_pipeline_inputs["image"], ] __lowerCAmelCase : Tuple = pipe( **_SCREAMING_SNAKE_CASE , return_dict=_SCREAMING_SNAKE_CASE , )[0] __lowerCAmelCase : Dict = image[0, -3:, -3:, -1] __lowerCAmelCase : int = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (2, 64, 64, 3) __lowerCAmelCase : List[str] = np.array( [ 0.9997, 0.9989, 0.0008, 0.0021, 0.9960, 0.0018, 0.0014, 0.0002, 0.9933, ]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 def _SCREAMING_SNAKE_CASE ( self: Any) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Tuple = torch.device("cpu") class A__ : '''simple docstring''' SCREAMING_SNAKE_CASE = 1 __lowerCAmelCase : Optional[Any] = self.get_dummy_components() __lowerCAmelCase : List[str] = self.pipeline_class(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = pipe.to(_SCREAMING_SNAKE_CASE) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = torch.Generator(device=_SCREAMING_SNAKE_CASE).manual_seed(0) __lowerCAmelCase : Any = pipe.decoder.dtype __lowerCAmelCase : List[Any] = 1 __lowerCAmelCase : int = ( batch_size, pipe.decoder.config.in_channels, pipe.decoder.config.sample_size, pipe.decoder.config.sample_size, ) __lowerCAmelCase : Union[str, Any] = pipe.prepare_latents( _SCREAMING_SNAKE_CASE , dtype=_SCREAMING_SNAKE_CASE , device=_SCREAMING_SNAKE_CASE , generator=_SCREAMING_SNAKE_CASE , latents=_SCREAMING_SNAKE_CASE , scheduler=DummyScheduler()) __lowerCAmelCase : List[str] = ( batch_size, pipe.super_res_first.config.in_channels // 2, pipe.super_res_first.config.sample_size, pipe.super_res_first.config.sample_size, ) __lowerCAmelCase : Union[str, Any] = pipe.prepare_latents( _SCREAMING_SNAKE_CASE , dtype=_SCREAMING_SNAKE_CASE , device=_SCREAMING_SNAKE_CASE , generator=_SCREAMING_SNAKE_CASE , latents=_SCREAMING_SNAKE_CASE , scheduler=DummyScheduler()) __lowerCAmelCase : Union[str, Any] = self.get_dummy_inputs(_SCREAMING_SNAKE_CASE , pil_image=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = pipe( **_SCREAMING_SNAKE_CASE , decoder_latents=_SCREAMING_SNAKE_CASE , super_res_latents=_SCREAMING_SNAKE_CASE).images __lowerCAmelCase : Dict = self.get_dummy_inputs(_SCREAMING_SNAKE_CASE , pil_image=_SCREAMING_SNAKE_CASE) # Don't pass image, instead pass embedding __lowerCAmelCase : Any = pipeline_inputs.pop("image") __lowerCAmelCase : Optional[Any] = pipe.image_encoder(_SCREAMING_SNAKE_CASE).image_embeds __lowerCAmelCase : int = pipe( **_SCREAMING_SNAKE_CASE , decoder_latents=_SCREAMING_SNAKE_CASE , super_res_latents=_SCREAMING_SNAKE_CASE , image_embeddings=_SCREAMING_SNAKE_CASE , ).images # make sure passing text embeddings manually is identical assert np.abs(img_out_a - img_out_a).max() < 1e-4 @skip_mps def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> Tuple: """simple docstring""" __lowerCAmelCase : Union[str, Any] = torch_device == "cpu" # Check is relaxed because there is not a torch 2.0 sliced attention added kv processor __lowerCAmelCase : Union[str, Any] = 1e-2 self._test_attention_slicing_forward_pass( test_max_difference=_SCREAMING_SNAKE_CASE , expected_max_diff=_SCREAMING_SNAKE_CASE) @skip_mps def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Dict: """simple docstring""" __lowerCAmelCase : Dict = torch_device == "cpu" __lowerCAmelCase : int = True __lowerCAmelCase : int = [ "decoder_num_inference_steps", "super_res_num_inference_steps", ] self._test_inference_batch_single_identical( test_max_difference=_SCREAMING_SNAKE_CASE , relax_max_difference=_SCREAMING_SNAKE_CASE , additional_params_copy_to_batched_inputs=_SCREAMING_SNAKE_CASE , ) def _SCREAMING_SNAKE_CASE ( self: Tuple) -> int: """simple docstring""" __lowerCAmelCase : int = [ "decoder_num_inference_steps", "super_res_num_inference_steps", ] if torch_device == "mps": # TODO: MPS errors with larger batch sizes __lowerCAmelCase : Dict = [2, 3] self._test_inference_batch_consistent( batch_sizes=_SCREAMING_SNAKE_CASE , additional_params_copy_to_batched_inputs=_SCREAMING_SNAKE_CASE , ) else: self._test_inference_batch_consistent( additional_params_copy_to_batched_inputs=_SCREAMING_SNAKE_CASE) @skip_mps def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> str: """simple docstring""" return super().test_dict_tuple_outputs_equivalent() @skip_mps def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> Any: """simple docstring""" return super().test_save_load_local() @skip_mps def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> Any: """simple docstring""" return super().test_save_load_optional_components() @slow @require_torch_gpu class A__ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Tuple) -> str: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _SCREAMING_SNAKE_CASE ( self: str) -> List[str]: """simple docstring""" __lowerCAmelCase : Optional[Any] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/unclip/cat.png") __lowerCAmelCase : int = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/unclip/karlo_v1_alpha_cat_variation_fp16.npy") __lowerCAmelCase : Union[str, Any] = UnCLIPImageVariationPipeline.from_pretrained( "kakaobrain/karlo-v1-alpha-image-variations" , torch_dtype=torch.floataa) __lowerCAmelCase : Optional[Any] = pipeline.to(_SCREAMING_SNAKE_CASE) pipeline.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = torch.Generator(device="cpu").manual_seed(0) __lowerCAmelCase : Any = pipeline( _SCREAMING_SNAKE_CASE , generator=_SCREAMING_SNAKE_CASE , output_type="np" , ) __lowerCAmelCase : Optional[int] = output.images[0] assert image.shape == (256, 256, 3) assert_mean_pixel_difference(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , 15)
269
"""simple docstring""" from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def _lowercase ( __snake_case = "laptop" ) -> DataFrame: __lowerCAmelCase : str = F"""https://www.amazon.in/laptop/s?k={product}""" __lowerCAmelCase : Union[str, Any] = { "User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36\n (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36", "Accept-Language": "en-US, en;q=0.5", } __lowerCAmelCase : List[str] = BeautifulSoup(requests.get(__snake_case ,headers=__snake_case ).text ) # Initialize a Pandas dataframe with the column titles __lowerCAmelCase : Dict = DataFrame( columns=[ "Product Title", "Product Link", "Current Price of the product", "Product Rating", "MRP of the product", "Discount", ] ) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( "div" ,attrs={"class": "s-result-item", "data-component-type": "s-search-result"} ,) ,soup.find_all("div" ,attrs={"class": "a-row a-size-base a-color-base"} ) ,): try: __lowerCAmelCase : Any = item.ha.text __lowerCAmelCase : Union[str, Any] = "https://www.amazon.in/" + item.ha.a["href"] __lowerCAmelCase : Any = item.find("span" ,attrs={"class": "a-offscreen"} ).text try: __lowerCAmelCase : Union[str, Any] = item.find("span" ,attrs={"class": "a-icon-alt"} ).text except AttributeError: __lowerCAmelCase : Optional[Any] = "Not available" try: __lowerCAmelCase : Union[str, Any] = ( "₹" + item.find( "span" ,attrs={"class": "a-price a-text-price"} ).text.split("₹" )[1] ) except AttributeError: __lowerCAmelCase : Dict = "" try: __lowerCAmelCase : str = float( ( ( float(product_mrp.strip("₹" ).replace("," ,"" ) ) - float(product_price.strip("₹" ).replace("," ,"" ) ) ) / float(product_mrp.strip("₹" ).replace("," ,"" ) ) ) * 100 ) except ValueError: __lowerCAmelCase : List[str] = float("nan" ) except AttributeError: pass __lowerCAmelCase : int = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] __lowerCAmelCase : Union[str, Any] = " " __lowerCAmelCase : Union[str, Any] = " " data_frame.index += 1 return data_frame if __name__ == "__main__": __snake_case : Any = 'headphones' get_amazon_product_data(product).to_csv(F"""Amazon Product Data for {product}.csv""")
269
1
"""simple docstring""" from math import factorial, radians def _lowercase ( __snake_case ,__snake_case = 18 ,__snake_case = 10 ) -> float: __lowerCAmelCase : Dict = angle_in_degrees - ((angle_in_degrees // 360.0) * 360.0) # Converting from degrees to radians __lowerCAmelCase : Union[str, Any] = radians(__snake_case ) __lowerCAmelCase : Any = angle_in_radians __lowerCAmelCase : Union[str, Any] = 3 __lowerCAmelCase : Union[str, Any] = -1 for _ in range(__snake_case ): result += (b * (angle_in_radians**a)) / factorial(__snake_case ) __lowerCAmelCase : Any = -b # One positive term and the next will be negative and so on... a += 2 # Increased by 2 for every term. return round(__snake_case ,__snake_case ) if __name__ == "__main__": __import__('doctest').testmod()
269
"""simple docstring""" import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('TEST_SAGEMAKER' , 'False' ) ) is not True , reason='Skipping test because should only be run when releasing minor transformers version' , ) @pytest.mark.usefixtures('sm_env' ) @parameterized_class( [ { 'framework': 'pytorch', 'script': 'run_glue.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_5_0, 'eval_accuracy': 0.6, 'eval_loss': 0.9}, }, { 'framework': 'tensorflow', 'script': 'run_tf.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_0_0, 'eval_accuracy': 0.3, 'eval_loss': 0.9}, }, ] ) class A__ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: int) -> Tuple: """simple docstring""" if self.framework == "pytorch": subprocess.run( F"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() , encoding="utf-8" , check=_SCREAMING_SNAKE_CASE , ) assert hasattr(self , "env") def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Any=1) -> Dict: """simple docstring""" return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=F"""{self.env.base_job_name}-single""" , instance_count=_SCREAMING_SNAKE_CASE , instance_type=self.instance_type , debugger_hook_config=_SCREAMING_SNAKE_CASE , hyperparameters={**self.env.hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version="py36" , ) def _SCREAMING_SNAKE_CASE ( self: str , _SCREAMING_SNAKE_CASE: List[Any]) -> Optional[Any]: """simple docstring""" TrainingJobAnalytics(_SCREAMING_SNAKE_CASE).export_csv(F"""{self.env.test_path}/{job_name}_metrics.csv""") def _SCREAMING_SNAKE_CASE ( self: str) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Tuple = self.create_estimator() # run training estimator.fit() # result dataframe __lowerCAmelCase : Tuple = TrainingJobAnalytics(estimator.latest_training_job.name).dataframe() # extract kpis __lowerCAmelCase : Optional[Any] = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"]) __lowerCAmelCase : Union[str, Any] = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"]) # get train time from SageMaker job, this includes starting, preprocessing, stopping __lowerCAmelCase : Tuple = ( Session().describe_training_job(estimator.latest_training_job.name).get("TrainingTimeInSeconds" , 99_9999) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy) assert all(t <= self.results["eval_loss"] for t in eval_loss) # dump tests result into json file to share in PR with open(F"""{estimator.latest_training_job.name}.json""" , "w") as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , _SCREAMING_SNAKE_CASE)
269
1
"""simple docstring""" import json import os import torch from diffusers import UNetaDModel os.makedirs('hub/hopper-medium-v2/unet/hor32', exist_ok=True) os.makedirs('hub/hopper-medium-v2/unet/hor128', exist_ok=True) os.makedirs('hub/hopper-medium-v2/value_function', exist_ok=True) def _lowercase ( __snake_case ) -> Optional[Any]: if hor == 128: __lowerCAmelCase : Dict = ("DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D") __lowerCAmelCase : Union[str, Any] = (32, 128, 256) __lowerCAmelCase : Dict = ("UpResnetBlock1D", "UpResnetBlock1D") elif hor == 32: __lowerCAmelCase : List[str] = ("DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D") __lowerCAmelCase : List[str] = (32, 64, 128, 256) __lowerCAmelCase : str = ("UpResnetBlock1D", "UpResnetBlock1D", "UpResnetBlock1D") __lowerCAmelCase : Union[str, Any] = torch.load(F"""/Users/bglickenhaus/Documents/diffuser/temporal_unet-hopper-mediumv2-hor{hor}.torch""" ) __lowerCAmelCase : Dict = model.state_dict() __lowerCAmelCase : Optional[Any] = { "down_block_types": down_block_types, "block_out_channels": block_out_channels, "up_block_types": up_block_types, "layers_per_block": 1, "use_timestep_embedding": True, "out_block_type": "OutConv1DBlock", "norm_num_groups": 8, "downsample_each_block": False, "in_channels": 14, "out_channels": 14, "extra_in_channels": 0, "time_embedding_type": "positional", "flip_sin_to_cos": False, "freq_shift": 1, "sample_size": 65_536, "mid_block_type": "MidResTemporalBlock1D", "act_fn": "mish", } __lowerCAmelCase : Union[str, Any] = UNetaDModel(**__snake_case ) print(F"""length of state dict: {len(state_dict.keys() )}""" ) print(F"""length of value function dict: {len(hf_value_function.state_dict().keys() )}""" ) __lowerCAmelCase : Any = dict(zip(model.state_dict().keys() ,hf_value_function.state_dict().keys() ) ) for k, v in mapping.items(): __lowerCAmelCase : List[str] = state_dict.pop(__snake_case ) hf_value_function.load_state_dict(__snake_case ) torch.save(hf_value_function.state_dict() ,F"""hub/hopper-medium-v2/unet/hor{hor}/diffusion_pytorch_model.bin""" ) with open(F"""hub/hopper-medium-v2/unet/hor{hor}/config.json""" ,"w" ) as f: json.dump(__snake_case ,__snake_case ) def _lowercase ( ) -> Union[str, Any]: __lowerCAmelCase : str = { "in_channels": 14, "down_block_types": ("DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D"), "up_block_types": (), "out_block_type": "ValueFunction", "mid_block_type": "ValueFunctionMidBlock1D", "block_out_channels": (32, 64, 128, 256), "layers_per_block": 1, "downsample_each_block": True, "sample_size": 65_536, "out_channels": 14, "extra_in_channels": 0, "time_embedding_type": "positional", "use_timestep_embedding": True, "flip_sin_to_cos": False, "freq_shift": 1, "norm_num_groups": 8, "act_fn": "mish", } __lowerCAmelCase : Union[str, Any] = torch.load("/Users/bglickenhaus/Documents/diffuser/value_function-hopper-mediumv2-hor32.torch" ) __lowerCAmelCase : Any = model __lowerCAmelCase : Union[str, Any] = UNetaDModel(**__snake_case ) print(F"""length of state dict: {len(state_dict.keys() )}""" ) print(F"""length of value function dict: {len(hf_value_function.state_dict().keys() )}""" ) __lowerCAmelCase : Tuple = dict(zip(state_dict.keys() ,hf_value_function.state_dict().keys() ) ) for k, v in mapping.items(): __lowerCAmelCase : int = state_dict.pop(__snake_case ) hf_value_function.load_state_dict(__snake_case ) torch.save(hf_value_function.state_dict() ,"hub/hopper-medium-v2/value_function/diffusion_pytorch_model.bin" ) with open("hub/hopper-medium-v2/value_function/config.json" ,"w" ) as f: json.dump(__snake_case ,__snake_case ) if __name__ == "__main__": unet(32) # unet(128) value_function()
269
"""simple docstring""" from typing import TYPE_CHECKING from ..utils import _LazyModule __snake_case : Optional[int] = { 'config': [ 'EXTERNAL_DATA_FORMAT_SIZE_LIMIT', 'OnnxConfig', 'OnnxConfigWithPast', 'OnnxSeq2SeqConfigWithPast', 'PatchingSpec', ], 'convert': ['export', 'validate_model_outputs'], 'features': ['FeaturesManager'], 'utils': ['ParameterFormat', 'compute_serialized_parameters_size'], } if TYPE_CHECKING: from .config import ( EXTERNAL_DATA_FORMAT_SIZE_LIMIT, OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast, PatchingSpec, ) from .convert import export, validate_model_outputs from .features import FeaturesManager from .utils import ParameterFormat, compute_serialized_parameters_size else: import sys __snake_case : Any = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
269
1
"""simple docstring""" import argparse import torch from transformers import FunnelBaseModel, FunnelConfig, FunnelModel, load_tf_weights_in_funnel from transformers.utils import logging logging.set_verbosity_info() def _lowercase ( __snake_case ,__snake_case ,__snake_case ,__snake_case ) -> Dict: # Initialise PyTorch model __lowerCAmelCase : Optional[Any] = FunnelConfig.from_json_file(__snake_case ) print(F"""Building PyTorch model from configuration: {config}""" ) __lowerCAmelCase : Union[str, Any] = FunnelBaseModel(__snake_case ) if base_model else FunnelModel(__snake_case ) # Load weights from tf checkpoint load_tf_weights_in_funnel(__snake_case ,__snake_case ,__snake_case ) # Save pytorch-model print(F"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() ,__snake_case ) if __name__ == "__main__": __snake_case : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help='The config json file corresponding to the pre-trained model. \nThis specifies the model architecture.', ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) parser.add_argument( '--base_model', action='store_true', help='Whether you want just the base model (no decoder) or not.' ) __snake_case : List[str] = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path, args.base_model )
269
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices __snake_case : Optional[int] = logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'maskformer-swin' SCREAMING_SNAKE_CASE = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self: Optional[int] , _SCREAMING_SNAKE_CASE: int=224 , _SCREAMING_SNAKE_CASE: Tuple=4 , _SCREAMING_SNAKE_CASE: int=3 , _SCREAMING_SNAKE_CASE: List[Any]=96 , _SCREAMING_SNAKE_CASE: Union[str, Any]=[2, 2, 6, 2] , _SCREAMING_SNAKE_CASE: Any=[3, 6, 12, 24] , _SCREAMING_SNAKE_CASE: List[str]=7 , _SCREAMING_SNAKE_CASE: List[str]=4.0 , _SCREAMING_SNAKE_CASE: Optional[int]=True , _SCREAMING_SNAKE_CASE: Tuple=0.0 , _SCREAMING_SNAKE_CASE: Any=0.0 , _SCREAMING_SNAKE_CASE: Any=0.1 , _SCREAMING_SNAKE_CASE: str="gelu" , _SCREAMING_SNAKE_CASE: Tuple=False , _SCREAMING_SNAKE_CASE: Union[str, Any]=0.02 , _SCREAMING_SNAKE_CASE: str=1e-5 , _SCREAMING_SNAKE_CASE: Optional[int]=None , _SCREAMING_SNAKE_CASE: str=None , **_SCREAMING_SNAKE_CASE: Union[str, Any] , ) -> List[str]: """simple docstring""" super().__init__(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = image_size __lowerCAmelCase : Any = patch_size __lowerCAmelCase : Tuple = num_channels __lowerCAmelCase : Any = embed_dim __lowerCAmelCase : Any = depths __lowerCAmelCase : Dict = len(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = num_heads __lowerCAmelCase : Tuple = window_size __lowerCAmelCase : Dict = mlp_ratio __lowerCAmelCase : Any = qkv_bias __lowerCAmelCase : Union[str, Any] = hidden_dropout_prob __lowerCAmelCase : int = attention_probs_dropout_prob __lowerCAmelCase : Tuple = drop_path_rate __lowerCAmelCase : int = hidden_act __lowerCAmelCase : Optional[int] = use_absolute_embeddings __lowerCAmelCase : List[str] = layer_norm_eps __lowerCAmelCase : Any = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model __lowerCAmelCase : Optional[Any] = int(embed_dim * 2 ** (len(_SCREAMING_SNAKE_CASE) - 1)) __lowerCAmelCase : Any = ["stem"] + [F"""stage{idx}""" for idx in range(1 , len(_SCREAMING_SNAKE_CASE) + 1)] __lowerCAmelCase , __lowerCAmelCase : List[str] = get_aligned_output_features_output_indices( out_features=_SCREAMING_SNAKE_CASE , out_indices=_SCREAMING_SNAKE_CASE , stage_names=self.stage_names)
269
1
"""simple docstring""" class A__ : '''simple docstring''' def __init__( self: Optional[int] , _SCREAMING_SNAKE_CASE: list) -> None: """simple docstring""" __lowerCAmelCase : Any = set_counts __lowerCAmelCase : Optional[Any] = max(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = len(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = [1] * num_sets __lowerCAmelCase : Dict = list(range(_SCREAMING_SNAKE_CASE)) def _SCREAMING_SNAKE_CASE ( self: Dict , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: int) -> bool: """simple docstring""" __lowerCAmelCase : Tuple = self.get_parent(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = self.get_parent(_SCREAMING_SNAKE_CASE) if src_parent == dst_parent: return False if self.ranks[dst_parent] >= self.ranks[src_parent]: self.set_counts[dst_parent] += self.set_counts[src_parent] __lowerCAmelCase : Optional[int] = 0 __lowerCAmelCase : Optional[int] = dst_parent if self.ranks[dst_parent] == self.ranks[src_parent]: self.ranks[dst_parent] += 1 __lowerCAmelCase : Tuple = self.set_counts[dst_parent] else: self.set_counts[src_parent] += self.set_counts[dst_parent] __lowerCAmelCase : Dict = 0 __lowerCAmelCase : List[str] = src_parent __lowerCAmelCase : List[str] = self.set_counts[src_parent] __lowerCAmelCase : Tuple = max(self.max_set , _SCREAMING_SNAKE_CASE) return True def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: int) -> int: """simple docstring""" if self.parents[disj_set] == disj_set: return disj_set __lowerCAmelCase : Optional[int] = self.get_parent(self.parents[disj_set]) return self.parents[disj_set]
269
"""simple docstring""" import gc import threading import time import psutil import torch class A__ : '''simple docstring''' def __init__( self: str) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Optional[Any] = psutil.Process() __lowerCAmelCase : str = False def _SCREAMING_SNAKE_CASE ( self: int) -> Optional[int]: """simple docstring""" __lowerCAmelCase : Optional[Any] = -1 while True: __lowerCAmelCase : str = max(self.process.memory_info().rss , self.cpu_memory_peak) # can't sleep or will not catch the peak right (this comment is here on purpose) if not self.peak_monitoring: break def _SCREAMING_SNAKE_CASE ( self: Tuple) -> int: """simple docstring""" __lowerCAmelCase : List[str] = True __lowerCAmelCase : str = threading.Thread(target=self.peak_monitor) __lowerCAmelCase : Tuple = True self.thread.start() def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> List[str]: """simple docstring""" __lowerCAmelCase : Tuple = False self.thread.join() return self.cpu_memory_peak __snake_case : Tuple = PeakCPUMemory() def _lowercase ( ) -> str: # Time __lowerCAmelCase : str = {"time": time.time()} gc.collect() torch.cuda.empty_cache() # CPU mem __lowerCAmelCase : Optional[Any] = psutil.Process().memory_info().rss cpu_peak_tracker.start() # GPU mem for i in range(torch.cuda.device_count() ): __lowerCAmelCase : Union[str, Any] = torch.cuda.memory_allocated(__snake_case ) torch.cuda.reset_peak_memory_stats() return measures def _lowercase ( __snake_case ) -> Optional[Any]: # Time __lowerCAmelCase : str = {"time": time.time() - start_measures["time"]} gc.collect() torch.cuda.empty_cache() # CPU mem __lowerCAmelCase : str = (psutil.Process().memory_info().rss - start_measures["cpu"]) / 2**20 __lowerCAmelCase : List[str] = (cpu_peak_tracker.stop() - start_measures["cpu"]) / 2**20 # GPU mem for i in range(torch.cuda.device_count() ): __lowerCAmelCase : Union[str, Any] = (torch.cuda.memory_allocated(__snake_case ) - start_measures[str(__snake_case )]) / 2**20 __lowerCAmelCase : Any = (torch.cuda.max_memory_allocated(__snake_case ) - start_measures[str(__snake_case )]) / 2**20 return measures def _lowercase ( __snake_case ,__snake_case ) -> Dict: print(F"""{description}:""" ) print(F"""- Time: {measures['time']:.2f}s""" ) for i in range(torch.cuda.device_count() ): print(F"""- GPU {i} allocated: {measures[str(__snake_case )]:.2f}MiB""" ) __lowerCAmelCase : Optional[Any] = measures[F"""{i}-peak"""] print(F"""- GPU {i} peak: {peak:.2f}MiB""" ) print(F"""- CPU RAM allocated: {measures['cpu']:.2f}MiB""" ) print(F"""- CPU RAM peak: {measures['cpu-peak']:.2f}MiB""" )
269
1
"""simple docstring""" from __future__ import annotations class A__ : '''simple docstring''' def __init__( self: Dict , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: str) -> int: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : int = text, pattern __lowerCAmelCase , __lowerCAmelCase : str = len(_SCREAMING_SNAKE_CASE), len(_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Dict , _SCREAMING_SNAKE_CASE: str) -> int: """simple docstring""" for i in range(self.patLen - 1 , -1 , -1): if char == self.pattern[i]: return i return -1 def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: int) -> int: """simple docstring""" for i in range(self.patLen - 1 , -1 , -1): if self.pattern[i] != self.text[current_pos + i]: return current_pos + i return -1 def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> list[int]: """simple docstring""" __lowerCAmelCase : Optional[int] = [] for i in range(self.textLen - self.patLen + 1): __lowerCAmelCase : Tuple = self.mismatch_in_text(_SCREAMING_SNAKE_CASE) if mismatch_index == -1: positions.append(_SCREAMING_SNAKE_CASE) else: __lowerCAmelCase : str = self.match_in_pattern(self.text[mismatch_index]) __lowerCAmelCase : Tuple = ( mismatch_index - match_index ) # shifting index lgtm [py/multiple-definition] return positions __snake_case : Union[str, Any] = 'ABAABA' __snake_case : Dict = 'AB' __snake_case : str = BoyerMooreSearch(text, pattern) __snake_case : List[str] = bms.bad_character_heuristic() if len(positions) == 0: print('No match found') else: print('Pattern found in following positions: ') print(positions)
269
"""simple docstring""" from ....configuration_utils import PretrainedConfig from ....utils import logging __snake_case : List[Any] = logging.get_logger(__name__) __snake_case : Any = { 'speechbrain/m-ctc-t-large': 'https://huggingface.co/speechbrain/m-ctc-t-large/resolve/main/config.json', # See all M-CTC-T models at https://huggingface.co/models?filter=mctct } class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'mctct' def __init__( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: str=8065 , _SCREAMING_SNAKE_CASE: str=1536 , _SCREAMING_SNAKE_CASE: str=36 , _SCREAMING_SNAKE_CASE: Optional[Any]=6144 , _SCREAMING_SNAKE_CASE: Optional[Any]=4 , _SCREAMING_SNAKE_CASE: Union[str, Any]=384 , _SCREAMING_SNAKE_CASE: Optional[Any]=920 , _SCREAMING_SNAKE_CASE: Union[str, Any]=1e-5 , _SCREAMING_SNAKE_CASE: List[Any]=0.3 , _SCREAMING_SNAKE_CASE: Optional[Any]="relu" , _SCREAMING_SNAKE_CASE: Optional[int]=0.02 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.3 , _SCREAMING_SNAKE_CASE: Dict=0.3 , _SCREAMING_SNAKE_CASE: List[Any]=1 , _SCREAMING_SNAKE_CASE: Optional[Any]=0 , _SCREAMING_SNAKE_CASE: List[str]=2 , _SCREAMING_SNAKE_CASE: Union[str, Any]=1 , _SCREAMING_SNAKE_CASE: Tuple=0.3 , _SCREAMING_SNAKE_CASE: Dict=1 , _SCREAMING_SNAKE_CASE: int=(7,) , _SCREAMING_SNAKE_CASE: str=(3,) , _SCREAMING_SNAKE_CASE: Union[str, Any]=80 , _SCREAMING_SNAKE_CASE: Tuple=1 , _SCREAMING_SNAKE_CASE: Dict=None , _SCREAMING_SNAKE_CASE: Tuple="sum" , _SCREAMING_SNAKE_CASE: List[str]=False , **_SCREAMING_SNAKE_CASE: Tuple , ) -> Tuple: """simple docstring""" super().__init__(**_SCREAMING_SNAKE_CASE , pad_token_id=_SCREAMING_SNAKE_CASE , bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = vocab_size __lowerCAmelCase : str = hidden_size __lowerCAmelCase : str = num_hidden_layers __lowerCAmelCase : str = intermediate_size __lowerCAmelCase : List[Any] = num_attention_heads __lowerCAmelCase : Dict = attention_head_dim __lowerCAmelCase : Optional[int] = max_position_embeddings __lowerCAmelCase : str = layer_norm_eps __lowerCAmelCase : Tuple = layerdrop __lowerCAmelCase : str = hidden_act __lowerCAmelCase : List[Any] = initializer_range __lowerCAmelCase : int = hidden_dropout_prob __lowerCAmelCase : Optional[Any] = attention_probs_dropout_prob __lowerCAmelCase : str = pad_token_id __lowerCAmelCase : Optional[int] = bos_token_id __lowerCAmelCase : Union[str, Any] = eos_token_id __lowerCAmelCase : Any = conv_glu_dim __lowerCAmelCase : Optional[int] = conv_dropout __lowerCAmelCase : Union[str, Any] = num_conv_layers __lowerCAmelCase : Optional[int] = input_feat_per_channel __lowerCAmelCase : Union[str, Any] = input_channels __lowerCAmelCase : Optional[Any] = conv_channels __lowerCAmelCase : Dict = ctc_loss_reduction __lowerCAmelCase : int = ctc_zero_infinity # prevents config testing fail with exporting to json __lowerCAmelCase : List[str] = list(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = list(_SCREAMING_SNAKE_CASE) if len(self.conv_kernel) != self.num_conv_layers: raise ValueError( "Configuration for convolutional module is incorrect. " "It is required that `len(config.conv_kernel)` == `config.num_conv_layers` " F"""but is `len(config.conv_kernel) = {len(self.conv_kernel)}`, """ F"""`config.num_conv_layers = {self.num_conv_layers}`.""")
269
1
"""simple docstring""" # This code is adapted from OpenAI's release # https://github.com/openai/human-eval/blob/master/human_eval/execution.py import contextlib import faulthandler import io import multiprocessing import os import platform import signal import tempfile def _lowercase ( __snake_case ,__snake_case ,__snake_case ,__snake_case ) -> Any: __lowerCAmelCase : Dict = multiprocessing.Manager() __lowerCAmelCase : Union[str, Any] = manager.list() __lowerCAmelCase : str = multiprocessing.Process(target=__snake_case ,args=(check_program, result, timeout) ) p.start() p.join(timeout=timeout + 1 ) if p.is_alive(): p.kill() if not result: result.append("timed out" ) return { "task_id": task_id, "passed": result[0] == "passed", "result": result[0], "completion_id": completion_id, } def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> str: with create_tempdir(): # These system calls are needed when cleaning up tempdir. import os import shutil __lowerCAmelCase : int = shutil.rmtree __lowerCAmelCase : Dict = os.rmdir __lowerCAmelCase : Dict = os.chdir # Disable functionalities that can make destructive changes to the test. reliability_guard() # Run program. try: __lowerCAmelCase : int = {} with swallow_io(): with time_limit(__snake_case ): exec(__snake_case ,__snake_case ) result.append("passed" ) except TimeoutException: result.append("timed out" ) except BaseException as e: result.append(F"""failed: {e}""" ) # Needed for cleaning up. __lowerCAmelCase : List[str] = rmtree __lowerCAmelCase : Optional[int] = rmdir __lowerCAmelCase : Optional[Any] = chdir @contextlib.contextmanager def _lowercase ( __snake_case ) -> str: def signal_handler(__snake_case ,__snake_case ): raise TimeoutException("Timed out!" ) signal.setitimer(signal.ITIMER_REAL ,__snake_case ) signal.signal(signal.SIGALRM ,__snake_case ) try: yield finally: signal.setitimer(signal.ITIMER_REAL ,0 ) @contextlib.contextmanager def _lowercase ( ) -> Any: __lowerCAmelCase : Optional[int] = WriteOnlyStringIO() with contextlib.redirect_stdout(__snake_case ): with contextlib.redirect_stderr(__snake_case ): with redirect_stdin(__snake_case ): yield @contextlib.contextmanager def _lowercase ( ) -> Tuple: with tempfile.TemporaryDirectory() as dirname: with chdir(__snake_case ): yield dirname class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' pass class A__ ( io.StringIO ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: List[str] , *_SCREAMING_SNAKE_CASE: Dict , **_SCREAMING_SNAKE_CASE: Dict) -> Dict: """simple docstring""" raise OSError def _SCREAMING_SNAKE_CASE ( self: str , *_SCREAMING_SNAKE_CASE: Dict , **_SCREAMING_SNAKE_CASE: Union[str, Any]) -> List[str]: """simple docstring""" raise OSError def _SCREAMING_SNAKE_CASE ( self: Union[str, Any] , *_SCREAMING_SNAKE_CASE: Any , **_SCREAMING_SNAKE_CASE: Union[str, Any]) -> Union[str, Any]: """simple docstring""" raise OSError def _SCREAMING_SNAKE_CASE ( self: str , *_SCREAMING_SNAKE_CASE: Tuple , **_SCREAMING_SNAKE_CASE: Dict) -> int: """simple docstring""" return False class A__ ( contextlib._RedirectStream ): # type: ignore '''simple docstring''' SCREAMING_SNAKE_CASE = 'stdin' @contextlib.contextmanager def _lowercase ( __snake_case ) -> Any: if root == ".": yield return __lowerCAmelCase : List[str] = os.getcwd() os.chdir(__snake_case ) try: yield except BaseException as exc: raise exc finally: os.chdir(__snake_case ) def _lowercase ( __snake_case=None ) -> str: if maximum_memory_bytes is not None: import resource resource.setrlimit(resource.RLIMIT_AS ,(maximum_memory_bytes, maximum_memory_bytes) ) resource.setrlimit(resource.RLIMIT_DATA ,(maximum_memory_bytes, maximum_memory_bytes) ) if not platform.uname().system == "Darwin": resource.setrlimit(resource.RLIMIT_STACK ,(maximum_memory_bytes, maximum_memory_bytes) ) faulthandler.disable() import builtins __lowerCAmelCase : Optional[int] = None __lowerCAmelCase : str = None import os __lowerCAmelCase : Union[str, Any] = "1" __lowerCAmelCase : Tuple = None __lowerCAmelCase : int = None __lowerCAmelCase : str = None __lowerCAmelCase : Dict = None __lowerCAmelCase : Any = None __lowerCAmelCase : Optional[Any] = None __lowerCAmelCase : str = None __lowerCAmelCase : List[Any] = None __lowerCAmelCase : Any = None __lowerCAmelCase : Dict = None __lowerCAmelCase : Union[str, Any] = None __lowerCAmelCase : Optional[Any] = None __lowerCAmelCase : Optional[int] = None __lowerCAmelCase : int = None __lowerCAmelCase : Optional[Any] = None __lowerCAmelCase : List[str] = None __lowerCAmelCase : Tuple = None __lowerCAmelCase : Tuple = None __lowerCAmelCase : int = None __lowerCAmelCase : Optional[int] = None __lowerCAmelCase : Tuple = None __lowerCAmelCase : List[str] = None __lowerCAmelCase : Dict = None __lowerCAmelCase : Any = None __lowerCAmelCase : List[str] = None __lowerCAmelCase : Tuple = None __lowerCAmelCase : Tuple = None import shutil __lowerCAmelCase : Union[str, Any] = None __lowerCAmelCase : str = None __lowerCAmelCase : Tuple = None import subprocess __lowerCAmelCase : Dict = None # type: ignore __lowerCAmelCase : List[str] = None import sys __lowerCAmelCase : Union[str, Any] = None __lowerCAmelCase : List[Any] = None __lowerCAmelCase : Any = None __lowerCAmelCase : List[str] = None __lowerCAmelCase : List[str] = None
269
"""simple docstring""" from __future__ import annotations import time import numpy as np __snake_case : Optional[Any] = [8, 5, 9, 7] __snake_case : List[Any] = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] __snake_case : Optional[int] = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class A__ : '''simple docstring''' def __init__( self: Any , _SCREAMING_SNAKE_CASE: list[int] , _SCREAMING_SNAKE_CASE: list[list[int]] , _SCREAMING_SNAKE_CASE: list[list[int]] , ) -> None: """simple docstring""" __lowerCAmelCase : Any = claim_vector __lowerCAmelCase : Tuple = allocated_resources_table __lowerCAmelCase : Tuple = maximum_claim_table def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> list[int]: """simple docstring""" return [ sum(p_item[i] for p_item in self.__allocated_resources_table) for i in range(len(self.__allocated_resources_table[0])) ] def _SCREAMING_SNAKE_CASE ( self: int) -> list[int]: """simple docstring""" return np.array(self.__claim_vector) - np.array( self.__processes_resource_summation()) def _SCREAMING_SNAKE_CASE ( self: int) -> list[list[int]]: """simple docstring""" return [ list(np.array(self.__maximum_claim_table[i]) - np.array(_SCREAMING_SNAKE_CASE)) for i, allocated_resource in enumerate(self.__allocated_resources_table) ] def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> dict[int, list[int]]: """simple docstring""" return {self.__need().index(_SCREAMING_SNAKE_CASE): i for i in self.__need()} def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , **_SCREAMING_SNAKE_CASE: List[Any]) -> None: """simple docstring""" __lowerCAmelCase : Optional[int] = self.__need() __lowerCAmelCase : int = self.__allocated_resources_table __lowerCAmelCase : Dict = self.__available_resources() __lowerCAmelCase : str = self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print("_" * 50 + "\n") while need_list: __lowerCAmelCase : int = False for each_need in need_list: __lowerCAmelCase : Dict = True for index, need in enumerate(_SCREAMING_SNAKE_CASE): if need > available_resources[index]: __lowerCAmelCase : Dict = False break if execution: __lowerCAmelCase : Any = True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: __lowerCAmelCase : Union[str, Any] = original_need_index print(F"""Process {process_number + 1} is executing.""") # remove the process run from stack need_list.remove(_SCREAMING_SNAKE_CASE) # update available/freed resources stack __lowerCAmelCase : Dict = np.array(_SCREAMING_SNAKE_CASE) + np.array( alloc_resources_table[process_number]) print( "Updated available resource stack for processes: " + " ".join([str(_SCREAMING_SNAKE_CASE) for x in available_resources])) break if safe: print("The process is in a safe state.\n") else: print("System in unsafe state. Aborting...\n") break def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> List[Any]: """simple docstring""" print(" " * 9 + "Allocated Resource Table") for item in self.__allocated_resources_table: print( F"""P{self.__allocated_resources_table.index(_SCREAMING_SNAKE_CASE) + 1}""" + " ".join(F"""{it:>8}""" for it in item) + "\n") print(" " * 9 + "System Resource Table") for item in self.__maximum_claim_table: print( F"""P{self.__maximum_claim_table.index(_SCREAMING_SNAKE_CASE) + 1}""" + " ".join(F"""{it:>8}""" for it in item) + "\n") print( "Current Usage by Active Processes: " + " ".join(str(_SCREAMING_SNAKE_CASE) for x in self.__claim_vector)) print( "Initial Available Resources: " + " ".join(str(_SCREAMING_SNAKE_CASE) for x in self.__available_resources())) time.sleep(1) if __name__ == "__main__": import doctest doctest.testmod()
269
1
"""simple docstring""" import inspect import os import torch from transformers import AutoModel from transformers.testing_utils import mockenv_context from transformers.trainer_utils import set_seed import accelerate from accelerate.accelerator import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils.testing import ( AccelerateTestCase, TempDirTestCase, execute_subprocess_async, require_cuda, require_fsdp, require_multi_gpu, slow, ) from accelerate.utils.constants import ( FSDP_AUTO_WRAP_POLICY, FSDP_BACKWARD_PREFETCH, FSDP_SHARDING_STRATEGY, FSDP_STATE_DICT_TYPE, ) from accelerate.utils.dataclasses import FullyShardedDataParallelPlugin from accelerate.utils.other import patch_environment set_seed(42) __snake_case : List[Any] = 'bert-base-cased' __snake_case : Tuple = 'fp16' __snake_case : str = 'bf16' __snake_case : Optional[int] = [FPaa, BFaa] @require_fsdp @require_cuda class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: List[str]) -> Union[str, Any]: """simple docstring""" super().setUp() __lowerCAmelCase : str = dict( ACCELERATE_USE_FSDP="true" , MASTER_ADDR="localhost" , MASTER_PORT="10999" , RANK="0" , LOCAL_RANK="0" , WORLD_SIZE="1" , ) def _SCREAMING_SNAKE_CASE ( self: int) -> Optional[Any]: """simple docstring""" from torch.distributed.fsdp.fully_sharded_data_parallel import ShardingStrategy for i, strategy in enumerate(_SCREAMING_SNAKE_CASE): __lowerCAmelCase : Tuple = self.dist_env.copy() __lowerCAmelCase : Dict = F"""{i + 1}""" __lowerCAmelCase : Union[str, Any] = strategy with mockenv_context(**_SCREAMING_SNAKE_CASE): __lowerCAmelCase : str = FullyShardedDataParallelPlugin() self.assertEqual(fsdp_plugin.sharding_strategy , ShardingStrategy(i + 1)) def _SCREAMING_SNAKE_CASE ( self: List[str]) -> Union[str, Any]: """simple docstring""" from torch.distributed.fsdp.fully_sharded_data_parallel import BackwardPrefetch for i, prefetch_policy in enumerate(_SCREAMING_SNAKE_CASE): __lowerCAmelCase : int = self.dist_env.copy() __lowerCAmelCase : List[str] = prefetch_policy with mockenv_context(**_SCREAMING_SNAKE_CASE): __lowerCAmelCase : Dict = FullyShardedDataParallelPlugin() if prefetch_policy == "NO_PREFETCH": self.assertIsNone(fsdp_plugin.backward_prefetch) else: self.assertEqual(fsdp_plugin.backward_prefetch , BackwardPrefetch(i + 1)) def _SCREAMING_SNAKE_CASE ( self: int) -> Union[str, Any]: """simple docstring""" from torch.distributed.fsdp.fully_sharded_data_parallel import StateDictType for i, state_dict_type in enumerate(_SCREAMING_SNAKE_CASE): __lowerCAmelCase : Union[str, Any] = self.dist_env.copy() __lowerCAmelCase : Optional[Any] = state_dict_type with mockenv_context(**_SCREAMING_SNAKE_CASE): __lowerCAmelCase : int = FullyShardedDataParallelPlugin() self.assertEqual(fsdp_plugin.state_dict_type , StateDictType(i + 1)) if state_dict_type == "FULL_STATE_DICT": self.assertTrue(fsdp_plugin.state_dict_config.offload_to_cpu) self.assertTrue(fsdp_plugin.state_dict_config.ranka_only) def _SCREAMING_SNAKE_CASE ( self: Dict) -> Dict: """simple docstring""" __lowerCAmelCase : List[str] = AutoModel.from_pretrained(_SCREAMING_SNAKE_CASE) for policy in FSDP_AUTO_WRAP_POLICY: __lowerCAmelCase : List[str] = self.dist_env.copy() __lowerCAmelCase : List[str] = policy if policy == "TRANSFORMER_BASED_WRAP": __lowerCAmelCase : List[str] = "BertLayer" elif policy == "SIZE_BASED_WRAP": __lowerCAmelCase : Any = "2000" with mockenv_context(**_SCREAMING_SNAKE_CASE): __lowerCAmelCase : Dict = FullyShardedDataParallelPlugin() fsdp_plugin.set_auto_wrap_policy(_SCREAMING_SNAKE_CASE) if policy == "NO_WRAP": self.assertIsNone(fsdp_plugin.auto_wrap_policy) else: self.assertIsNotNone(fsdp_plugin.auto_wrap_policy) __lowerCAmelCase : List[str] = self.dist_env.copy() __lowerCAmelCase : int = "TRANSFORMER_BASED_WRAP" __lowerCAmelCase : List[Any] = "T5Layer" with mockenv_context(**_SCREAMING_SNAKE_CASE): __lowerCAmelCase : Optional[Any] = FullyShardedDataParallelPlugin() with self.assertRaises(_SCREAMING_SNAKE_CASE) as cm: fsdp_plugin.set_auto_wrap_policy(_SCREAMING_SNAKE_CASE) self.assertTrue("Could not find the transformer layer class to wrap in the model." in str(cm.exception)) __lowerCAmelCase : Any = self.dist_env.copy() __lowerCAmelCase : Any = "SIZE_BASED_WRAP" __lowerCAmelCase : str = "0" with mockenv_context(**_SCREAMING_SNAKE_CASE): __lowerCAmelCase : Tuple = FullyShardedDataParallelPlugin() fsdp_plugin.set_auto_wrap_policy(_SCREAMING_SNAKE_CASE) self.assertIsNone(fsdp_plugin.auto_wrap_policy) def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> Union[str, Any]: """simple docstring""" from torch.distributed.fsdp.fully_sharded_data_parallel import MixedPrecision from torch.distributed.fsdp.sharded_grad_scaler import ShardedGradScaler for mp_dtype in dtypes: __lowerCAmelCase : Union[str, Any] = self.dist_env.copy() __lowerCAmelCase : Tuple = mp_dtype with mockenv_context(**_SCREAMING_SNAKE_CASE): __lowerCAmelCase : Any = Accelerator() if mp_dtype == "fp16": __lowerCAmelCase : Optional[Any] = torch.floataa elif mp_dtype == "bf16": __lowerCAmelCase : List[Any] = torch.bfloataa __lowerCAmelCase : int = MixedPrecision(param_dtype=_SCREAMING_SNAKE_CASE , reduce_dtype=_SCREAMING_SNAKE_CASE , buffer_dtype=_SCREAMING_SNAKE_CASE) self.assertEqual(accelerator.state.fsdp_plugin.mixed_precision_policy , _SCREAMING_SNAKE_CASE) if mp_dtype == FPaa: self.assertTrue(isinstance(accelerator.scaler , _SCREAMING_SNAKE_CASE)) elif mp_dtype == BFaa: self.assertIsNone(accelerator.scaler) AcceleratorState._reset_state(_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> int: """simple docstring""" from torch.distributed.fsdp.fully_sharded_data_parallel import CPUOffload for flag in [True, False]: __lowerCAmelCase : Tuple = self.dist_env.copy() __lowerCAmelCase : List[Any] = str(_SCREAMING_SNAKE_CASE).lower() with mockenv_context(**_SCREAMING_SNAKE_CASE): __lowerCAmelCase : int = FullyShardedDataParallelPlugin() self.assertEqual(fsdp_plugin.cpu_offload , CPUOffload(offload_params=_SCREAMING_SNAKE_CASE)) @require_fsdp @require_multi_gpu @slow class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: List[str]) -> List[str]: """simple docstring""" super().setUp() __lowerCAmelCase : Optional[Any] = 0.82 __lowerCAmelCase : Dict = [ "fsdp_shard_grad_op_transformer_based_wrap", "fsdp_full_shard_transformer_based_wrap", ] __lowerCAmelCase : List[Any] = { "multi_gpu_fp16": 3200, "fsdp_shard_grad_op_transformer_based_wrap_fp16": 2000, "fsdp_full_shard_transformer_based_wrap_fp16": 1900, # Disabling below test as it overwhelms the RAM memory usage # on CI self-hosted runner leading to tests getting killed. # "fsdp_full_shard_cpu_offload_transformer_based_wrap_fp32": 1500, # fp16 was leading to indefinite hang } __lowerCAmelCase : Tuple = 160 __lowerCAmelCase : Any = 160 __lowerCAmelCase : Optional[Any] = inspect.getfile(accelerate.test_utils) __lowerCAmelCase : str = os.path.sep.join(mod_file.split(os.path.sep)[:-1] + ["scripts", "external_deps"]) def _SCREAMING_SNAKE_CASE ( self: Any) -> List[Any]: """simple docstring""" __lowerCAmelCase : Union[str, Any] = os.path.join(self.test_scripts_folder , "test_performance.py") __lowerCAmelCase : Tuple = ["accelerate", "launch", "--num_processes=2", "--num_machines=1", "--machine_rank=0", "--use_fsdp"] for config in self.performance_configs: __lowerCAmelCase : Tuple = cmd.copy() for i, strategy in enumerate(_SCREAMING_SNAKE_CASE): if strategy.lower() in config: cmd_config.append(F"""--fsdp_sharding_strategy={i+1}""") break if "fp32" in config: cmd_config.append("--mixed_precision=no") else: cmd_config.append("--mixed_precision=fp16") if "cpu_offload" in config: cmd_config.append("--fsdp_offload_params=True") for policy in FSDP_AUTO_WRAP_POLICY: if policy.lower() in config: cmd_config.append(F"""--fsdp_auto_wrap_policy={policy}""") break if policy == "TRANSFORMER_BASED_WRAP": cmd_config.append("--fsdp_transformer_layer_cls_to_wrap=BertLayer") elif policy == "SIZE_BASED_WRAP": cmd_config.append("--fsdp_min_num_params=2000") cmd_config.extend( [ self.test_file_path, F"""--output_dir={self.tmpdir}""", F"""--performance_lower_bound={self.performance_lower_bound}""", ]) with patch_environment(omp_num_threads=1): execute_subprocess_async(_SCREAMING_SNAKE_CASE , env=os.environ.copy()) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> Dict: """simple docstring""" __lowerCAmelCase : List[str] = os.path.join(self.test_scripts_folder , "test_checkpointing.py") __lowerCAmelCase : List[str] = [ "accelerate", "launch", "--num_processes=2", "--num_machines=1", "--machine_rank=0", "--use_fsdp", "--mixed_precision=fp16", "--fsdp_transformer_layer_cls_to_wrap=BertLayer", ] for i, strategy in enumerate(_SCREAMING_SNAKE_CASE): __lowerCAmelCase : Any = cmd.copy() cmd_config.append(F"""--fsdp_sharding_strategy={i+1}""") if strategy != "FULL_SHARD": continue __lowerCAmelCase : Union[str, Any] = len(_SCREAMING_SNAKE_CASE) for state_dict_type in FSDP_STATE_DICT_TYPE: __lowerCAmelCase : str = cmd_config[:state_dict_config_index] cmd_config.append(F"""--fsdp_state_dict_type={state_dict_type}""") cmd_config.extend( [ self.test_file_path, F"""--output_dir={self.tmpdir}""", "--partial_train_epoch=1", ]) with patch_environment(omp_num_threads=1): execute_subprocess_async(_SCREAMING_SNAKE_CASE , env=os.environ.copy()) __lowerCAmelCase : Tuple = cmd_config[:-1] __lowerCAmelCase : List[str] = os.path.join(self.tmpdir , "epoch_0") cmd_config.extend( [ F"""--resume_from_checkpoint={resume_from_checkpoint}""", ]) with patch_environment(omp_num_threads=1): execute_subprocess_async(_SCREAMING_SNAKE_CASE , env=os.environ.copy()) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> Any: """simple docstring""" __lowerCAmelCase : Union[str, Any] = os.path.join(self.test_scripts_folder , "test_peak_memory_usage.py") __lowerCAmelCase : str = [ "accelerate", "launch", "--num_processes=2", "--num_machines=1", "--machine_rank=0", ] for spec, peak_mem_upper_bound in self.peak_memory_usage_upper_bound.items(): __lowerCAmelCase : Dict = cmd.copy() if "fp16" in spec: cmd_config.extend(["--mixed_precision=fp16"]) else: cmd_config.extend(["--mixed_precision=no"]) if "multi_gpu" in spec: continue else: cmd_config.extend(["--use_fsdp"]) for i, strategy in enumerate(_SCREAMING_SNAKE_CASE): if strategy.lower() in spec: cmd_config.append(F"""--fsdp_sharding_strategy={i+1}""") break if "cpu_offload" in spec: cmd_config.append("--fsdp_offload_params=True") for policy in FSDP_AUTO_WRAP_POLICY: if policy.lower() in spec: cmd_config.append(F"""--fsdp_auto_wrap_policy={policy}""") break if policy == "TRANSFORMER_BASED_WRAP": cmd_config.append("--fsdp_transformer_layer_cls_to_wrap=BertLayer") elif policy == "SIZE_BASED_WRAP": cmd_config.append("--fsdp_min_num_params=2000") cmd_config.extend( [ self.test_file_path, F"""--output_dir={self.tmpdir}""", F"""--peak_memory_upper_bound={peak_mem_upper_bound}""", F"""--n_train={self.n_train}""", F"""--n_val={self.n_val}""", ]) with patch_environment(omp_num_threads=1): execute_subprocess_async(_SCREAMING_SNAKE_CASE , env=os.environ.copy())
269
"""simple docstring""" # # This a `torch.distributed` diagnostics script that checks that all GPUs in the cluster (one or # many nodes) can talk to each other via nccl and allocate gpu memory. # # To run first adjust the number of processes and nodes: # # python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # You may need to add --master_addr $MASTER_ADDR --master_port $MASTER_PORT if using a custom addr:port # # You can also use the rdzv API: --rdzv_endpoint $MASTER_ADDR:$MASTER_PORT --rdzv_backend c10d # # use torch.distributed.launch instead of torch.distributed.run for torch < 1.9 # # If you get a hanging in `barrier` calls you have some network issues, you may try to debug this with: # # NCCL_DEBUG=INFO python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # which should tell you what's going on behind the scenes. # # # This script can be run via `srun` in the SLURM environment as well. Here is a SLURM script that # runs on 2 nodes of 4 gpus per node: # # #SBATCH --job-name=test-nodes # name # #SBATCH --nodes=2 # nodes # #SBATCH --ntasks-per-node=1 # crucial - only 1 task per dist per node! # #SBATCH --cpus-per-task=10 # number of cores per tasks # #SBATCH --gres=gpu:4 # number of gpus # #SBATCH --time 0:05:00 # maximum execution time (HH:MM:SS) # #SBATCH --output=%x-%j.out # output file name # # GPUS_PER_NODE=4 # MASTER_ADDR=$(scontrol show hostnames $SLURM_JOB_NODELIST | head -n 1) # MASTER_PORT=6000 # # srun --jobid $SLURM_JOBID bash -c 'python -m torch.distributed.run \ # --nproc_per_node $GPUS_PER_NODE --nnodes $SLURM_NNODES --node_rank $SLURM_PROCID \ # --master_addr $MASTER_ADDR --master_port $MASTER_PORT \ # torch-distributed-gpu-test.py' # import fcntl import os import socket import torch import torch.distributed as dist def _lowercase ( *__snake_case ) -> Optional[Any]: with open(__snake_case ,"r" ) as fh: fcntl.flock(__snake_case ,fcntl.LOCK_EX ) try: print(*__snake_case ) finally: fcntl.flock(__snake_case ,fcntl.LOCK_UN ) __snake_case : List[Any] = int(os.environ['LOCAL_RANK']) torch.cuda.set_device(local_rank) __snake_case : List[str] = torch.device('cuda', local_rank) __snake_case : Optional[Any] = socket.gethostname() __snake_case : str = F"""[{hostname}-{local_rank}]""" try: # test distributed dist.init_process_group('nccl') dist.all_reduce(torch.ones(1).to(device), op=dist.ReduceOp.SUM) dist.barrier() # test cuda is available and can allocate memory torch.cuda.is_available() torch.ones(1).cuda(local_rank) # global rank __snake_case : Tuple = dist.get_rank() __snake_case : Any = dist.get_world_size() printflock(F"""{gpu} is OK (global rank: {rank}/{world_size})""") dist.barrier() if rank == 0: printflock(F"""pt={torch.__version__}, cuda={torch.version.cuda}, nccl={torch.cuda.nccl.version()}""") except Exception: printflock(F"""{gpu} is broken""") raise
269
1
"""simple docstring""" import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class A__ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> List[str]: """simple docstring""" __lowerCAmelCase : str = "| <pad> <unk> <s> </s> a b c d e f g h i j k".split() __lowerCAmelCase : List[str] = dict(zip(_SCREAMING_SNAKE_CASE , range(len(_SCREAMING_SNAKE_CASE)))) __lowerCAmelCase : str = { "unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", } __lowerCAmelCase : List[Any] = { "feature_size": 1, "padding_value": 0.0, "sampling_rate": 1_6000, "return_attention_mask": False, "do_normalize": True, } __lowerCAmelCase : int = tempfile.mkdtemp() __lowerCAmelCase : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"]) __lowerCAmelCase : Tuple = os.path.join(self.tmpdirname , _SCREAMING_SNAKE_CASE) with open(self.vocab_file , "w" , encoding="utf-8") as fp: fp.write(json.dumps(_SCREAMING_SNAKE_CASE) + "\n") with open(self.feature_extraction_file , "w" , encoding="utf-8") as fp: fp.write(json.dumps(_SCREAMING_SNAKE_CASE) + "\n") # load decoder from hub __lowerCAmelCase : int = "hf-internal-testing/ngram-beam-search-decoder" def _SCREAMING_SNAKE_CASE ( self: int , **_SCREAMING_SNAKE_CASE: int) -> Any: """simple docstring""" __lowerCAmelCase : Dict = self.add_kwargs_tokens_map.copy() kwargs.update(_SCREAMING_SNAKE_CASE) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: str , **_SCREAMING_SNAKE_CASE: str) -> List[Any]: """simple docstring""" return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: int , **_SCREAMING_SNAKE_CASE: Union[str, Any]) -> List[Any]: """simple docstring""" return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> Any: """simple docstring""" shutil.rmtree(self.tmpdirname) def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : Any = self.get_tokenizer() __lowerCAmelCase : Tuple = self.get_feature_extractor() __lowerCAmelCase : Any = self.get_decoder() __lowerCAmelCase : int = WavaVecaProcessorWithLM(tokenizer=_SCREAMING_SNAKE_CASE , feature_extractor=_SCREAMING_SNAKE_CASE , decoder=_SCREAMING_SNAKE_CASE) processor.save_pretrained(self.tmpdirname) __lowerCAmelCase : List[str] = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab()) self.assertIsInstance(processor.tokenizer , _SCREAMING_SNAKE_CASE) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string()) self.assertIsInstance(processor.feature_extractor , _SCREAMING_SNAKE_CASE) # decoder self.assertEqual(processor.decoder._alphabet.labels , decoder._alphabet.labels) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set , decoder.model_container[decoder._model_key]._unigram_set , ) self.assertIsInstance(processor.decoder , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: int) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[int] = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder()) processor.save_pretrained(self.tmpdirname) # make sure that error is thrown when decoder alphabet doesn't match __lowerCAmelCase : str = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname , alpha=5.0 , beta=3.0 , score_boundary=-7.0 , unk_score_offset=3) # decoder self.assertEqual(processor.language_model.alpha , 5.0) self.assertEqual(processor.language_model.beta , 3.0) self.assertEqual(processor.language_model.score_boundary , -7.0) self.assertEqual(processor.language_model.unk_score_offset , 3) def _SCREAMING_SNAKE_CASE ( self: List[str]) -> Tuple: """simple docstring""" __lowerCAmelCase : List[str] = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(["xx"]) with self.assertRaisesRegex(_SCREAMING_SNAKE_CASE , "include"): WavaVecaProcessorWithLM( tokenizer=_SCREAMING_SNAKE_CASE , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder()) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> Optional[int]: """simple docstring""" __lowerCAmelCase : Any = self.get_feature_extractor() __lowerCAmelCase : Any = self.get_tokenizer() __lowerCAmelCase : Dict = self.get_decoder() __lowerCAmelCase : Tuple = WavaVecaProcessorWithLM(tokenizer=_SCREAMING_SNAKE_CASE , feature_extractor=_SCREAMING_SNAKE_CASE , decoder=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = floats_list((3, 1000)) __lowerCAmelCase : List[str] = feature_extractor(_SCREAMING_SNAKE_CASE , return_tensors="np") __lowerCAmelCase : Optional[Any] = processor(_SCREAMING_SNAKE_CASE , return_tensors="np") for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2) def _SCREAMING_SNAKE_CASE ( self: List[str]) -> Dict: """simple docstring""" __lowerCAmelCase : Any = self.get_feature_extractor() __lowerCAmelCase : Optional[Any] = self.get_tokenizer() __lowerCAmelCase : List[str] = self.get_decoder() __lowerCAmelCase : Union[str, Any] = WavaVecaProcessorWithLM(tokenizer=_SCREAMING_SNAKE_CASE , feature_extractor=_SCREAMING_SNAKE_CASE , decoder=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = "This is a test string" __lowerCAmelCase : List[str] = processor(text=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = tokenizer(_SCREAMING_SNAKE_CASE) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key]) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: Tuple=(2, 10, 16) , _SCREAMING_SNAKE_CASE: Optional[Any]=77) -> int: """simple docstring""" np.random.seed(_SCREAMING_SNAKE_CASE) return np.random.rand(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Tuple) -> List[Any]: """simple docstring""" __lowerCAmelCase : List[str] = self.get_feature_extractor() __lowerCAmelCase : Optional[Any] = self.get_tokenizer() __lowerCAmelCase : int = self.get_decoder() __lowerCAmelCase : List[Any] = WavaVecaProcessorWithLM(tokenizer=_SCREAMING_SNAKE_CASE , feature_extractor=_SCREAMING_SNAKE_CASE , decoder=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = self._get_dummy_logits(shape=(10, 16) , seed=13) __lowerCAmelCase : Union[str, Any] = processor.decode(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = decoder.decode_beams(_SCREAMING_SNAKE_CASE)[0] self.assertEqual(decoded_decoder[0] , decoded_processor.text) self.assertEqual("</s> <s> </s>" , decoded_processor.text) self.assertEqual(decoded_decoder[-2] , decoded_processor.logit_score) self.assertEqual(decoded_decoder[-1] , decoded_processor.lm_score) @parameterized.expand([[None], ["fork"], ["spawn"]]) def _SCREAMING_SNAKE_CASE ( self: int , _SCREAMING_SNAKE_CASE: int) -> str: """simple docstring""" __lowerCAmelCase : Optional[Any] = self.get_feature_extractor() __lowerCAmelCase : str = self.get_tokenizer() __lowerCAmelCase : Tuple = self.get_decoder() __lowerCAmelCase : str = WavaVecaProcessorWithLM(tokenizer=_SCREAMING_SNAKE_CASE , feature_extractor=_SCREAMING_SNAKE_CASE , decoder=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: __lowerCAmelCase : int = processor.batch_decode(_SCREAMING_SNAKE_CASE) else: with get_context(_SCREAMING_SNAKE_CASE).Pool() as pool: __lowerCAmelCase : int = processor.batch_decode(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = list(_SCREAMING_SNAKE_CASE) with get_context("fork").Pool() as p: __lowerCAmelCase : Tuple = decoder.decode_beams_batch(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Dict = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0]) logit_scores_decoder.append(beams[0][-2]) lm_scores_decoder.append(beams[0][-1]) self.assertListEqual(_SCREAMING_SNAKE_CASE , decoded_processor.text) self.assertListEqual(["<s> <s> </s>", "<s> <s> <s>"] , decoded_processor.text) self.assertListEqual(_SCREAMING_SNAKE_CASE , decoded_processor.logit_score) self.assertListEqual(_SCREAMING_SNAKE_CASE , decoded_processor.lm_score) def _SCREAMING_SNAKE_CASE ( self: List[str]) -> Optional[int]: """simple docstring""" __lowerCAmelCase : Any = self.get_feature_extractor() __lowerCAmelCase : Tuple = self.get_tokenizer() __lowerCAmelCase : Union[str, Any] = self.get_decoder() __lowerCAmelCase : Tuple = WavaVecaProcessorWithLM(tokenizer=_SCREAMING_SNAKE_CASE , feature_extractor=_SCREAMING_SNAKE_CASE , decoder=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = self._get_dummy_logits() __lowerCAmelCase : Tuple = 15 __lowerCAmelCase : List[Any] = -20.0 __lowerCAmelCase : Optional[int] = -4.0 __lowerCAmelCase : Tuple = processor.batch_decode( _SCREAMING_SNAKE_CASE , beam_width=_SCREAMING_SNAKE_CASE , beam_prune_logp=_SCREAMING_SNAKE_CASE , token_min_logp=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Dict = decoded_processor_out.text __lowerCAmelCase : Optional[int] = list(_SCREAMING_SNAKE_CASE) with get_context("fork").Pool() as pool: __lowerCAmelCase : Optional[int] = decoder.decode_beams_batch( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , beam_width=_SCREAMING_SNAKE_CASE , beam_prune_logp=_SCREAMING_SNAKE_CASE , token_min_logp=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Tuple = [d[0][0] for d in decoded_decoder_out] __lowerCAmelCase : int = [d[0][2] for d in decoded_decoder_out] __lowerCAmelCase : Tuple = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) self.assertListEqual(["</s> <s> <s>", "<s> <s> <s>"] , _SCREAMING_SNAKE_CASE) self.assertTrue(np.array_equal(_SCREAMING_SNAKE_CASE , decoded_processor_out.logit_score)) self.assertTrue(np.allclose([-20.054, -18.447] , _SCREAMING_SNAKE_CASE , atol=1e-3)) self.assertTrue(np.array_equal(_SCREAMING_SNAKE_CASE , decoded_processor_out.lm_score)) self.assertTrue(np.allclose([-15.554, -13.9474] , _SCREAMING_SNAKE_CASE , atol=1e-3)) def _SCREAMING_SNAKE_CASE ( self: List[str]) -> Dict: """simple docstring""" __lowerCAmelCase : Any = self.get_feature_extractor() __lowerCAmelCase : List[str] = self.get_tokenizer() __lowerCAmelCase : Any = self.get_decoder() __lowerCAmelCase : str = WavaVecaProcessorWithLM(tokenizer=_SCREAMING_SNAKE_CASE , feature_extractor=_SCREAMING_SNAKE_CASE , decoder=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = self._get_dummy_logits() __lowerCAmelCase : Union[str, Any] = 2.0 __lowerCAmelCase : str = 5.0 __lowerCAmelCase : int = -20.0 __lowerCAmelCase : Any = True __lowerCAmelCase : Optional[Any] = processor.batch_decode( _SCREAMING_SNAKE_CASE , alpha=_SCREAMING_SNAKE_CASE , beta=_SCREAMING_SNAKE_CASE , unk_score_offset=_SCREAMING_SNAKE_CASE , lm_score_boundary=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Tuple = decoded_processor_out.text __lowerCAmelCase : Union[str, Any] = list(_SCREAMING_SNAKE_CASE) decoder.reset_params( alpha=_SCREAMING_SNAKE_CASE , beta=_SCREAMING_SNAKE_CASE , unk_score_offset=_SCREAMING_SNAKE_CASE , lm_score_boundary=_SCREAMING_SNAKE_CASE , ) with get_context("fork").Pool() as pool: __lowerCAmelCase : Any = decoder.decode_beams_batch( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Optional[Any] = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) self.assertListEqual(["<s> </s> <s> </s> </s>", "</s> </s> <s> </s> </s>"] , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha , 2.0) self.assertEqual(lm_model.beta , 5.0) self.assertEqual(lm_model.unk_score_offset , -20.0) self.assertEqual(lm_model.score_boundary , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : List[Any] = WavaVecaProcessorWithLM.from_pretrained("hf-internal-testing/processor_with_lm") __lowerCAmelCase : Optional[int] = processor.decoder.model_container[processor.decoder._model_key] __lowerCAmelCase : Any = Path(language_model._kenlm_model.path.decode("utf-8")).parent.parent.absolute() __lowerCAmelCase : List[Any] = os.listdir(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = ["alphabet.json", "language_model"] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : str = snapshot_download("hf-internal-testing/processor_with_lm") __lowerCAmelCase : str = WavaVecaProcessorWithLM.from_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = processor.decoder.model_container[processor.decoder._model_key] __lowerCAmelCase : str = Path(language_model._kenlm_model.path.decode("utf-8")).parent.parent.absolute() __lowerCAmelCase : Optional[Any] = os.listdir(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = os.listdir(_SCREAMING_SNAKE_CASE) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> str: """simple docstring""" __lowerCAmelCase : Dict = WavaVecaProcessorWithLM.from_pretrained("hf-internal-testing/processor_with_lm") __lowerCAmelCase : Dict = AutoProcessor.from_pretrained("hf-internal-testing/processor_with_lm") __lowerCAmelCase : Optional[int] = floats_list((3, 1000)) __lowerCAmelCase : List[Any] = processor_wavaveca(_SCREAMING_SNAKE_CASE , return_tensors="np") __lowerCAmelCase : Tuple = processor_auto(_SCREAMING_SNAKE_CASE , return_tensors="np") for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() , input_auto[key].sum() , delta=1e-2) __lowerCAmelCase : Optional[Any] = self._get_dummy_logits() __lowerCAmelCase : List[Any] = processor_wavaveca.batch_decode(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = processor_auto.batch_decode(_SCREAMING_SNAKE_CASE) self.assertListEqual(decoded_wavaveca.text , decoded_auto.text) def _SCREAMING_SNAKE_CASE ( self: List[str]) -> List[str]: """simple docstring""" __lowerCAmelCase : Tuple = self.get_feature_extractor() __lowerCAmelCase : int = self.get_tokenizer() __lowerCAmelCase : Tuple = self.get_decoder() __lowerCAmelCase : List[str] = WavaVecaProcessorWithLM(tokenizer=_SCREAMING_SNAKE_CASE , feature_extractor=_SCREAMING_SNAKE_CASE , decoder=_SCREAMING_SNAKE_CASE) self.assertListEqual( processor.model_input_names , feature_extractor.model_input_names , msg="`processor` and `feature_extractor` model input names do not match" , ) @staticmethod def _SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: Optional[Any]) -> Any: """simple docstring""" __lowerCAmelCase : List[Any] = [d[key] for d in offsets] return retrieved_list def _SCREAMING_SNAKE_CASE ( self: Dict) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : Optional[Any] = WavaVecaProcessorWithLM.from_pretrained("hf-internal-testing/processor_with_lm") __lowerCAmelCase : Dict = self._get_dummy_logits()[0] __lowerCAmelCase : Union[str, Any] = processor.decode(_SCREAMING_SNAKE_CASE , output_word_offsets=_SCREAMING_SNAKE_CASE) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys()) , 4) self.assertTrue("text" in outputs) self.assertTrue("word_offsets" in outputs) self.assertTrue(isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE)) self.assertEqual(" ".join(self.get_from_offsets(outputs["word_offsets"] , "word")) , outputs.text) self.assertListEqual(self.get_from_offsets(outputs["word_offsets"] , "word") , ["<s>", "<s>", "</s>"]) self.assertListEqual(self.get_from_offsets(outputs["word_offsets"] , "start_offset") , [0, 2, 4]) self.assertListEqual(self.get_from_offsets(outputs["word_offsets"] , "end_offset") , [1, 3, 5]) def _SCREAMING_SNAKE_CASE ( self: List[str]) -> int: """simple docstring""" __lowerCAmelCase : Optional[int] = WavaVecaProcessorWithLM.from_pretrained("hf-internal-testing/processor_with_lm") __lowerCAmelCase : str = self._get_dummy_logits() __lowerCAmelCase : Tuple = processor.batch_decode(_SCREAMING_SNAKE_CASE , output_word_offsets=_SCREAMING_SNAKE_CASE) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys()) , 4) self.assertTrue("text" in outputs) self.assertTrue("word_offsets" in outputs) self.assertTrue(isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE)) self.assertListEqual( [" ".join(self.get_from_offsets(_SCREAMING_SNAKE_CASE , "word")) for o in outputs["word_offsets"]] , outputs.text) self.assertListEqual(self.get_from_offsets(outputs["word_offsets"][0] , "word") , ["<s>", "<s>", "</s>"]) self.assertListEqual(self.get_from_offsets(outputs["word_offsets"][0] , "start_offset") , [0, 2, 4]) self.assertListEqual(self.get_from_offsets(outputs["word_offsets"][0] , "end_offset") , [1, 3, 5]) @slow @require_torch @require_torchaudio def _SCREAMING_SNAKE_CASE ( self: int) -> Any: """simple docstring""" import torch __lowerCAmelCase : int = load_dataset("common_voice" , "en" , split="train" , streaming=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = ds.cast_column("audio" , datasets.Audio(sampling_rate=1_6000)) __lowerCAmelCase : int = iter(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = next(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = AutoProcessor.from_pretrained("patrickvonplaten/wav2vec2-base-100h-with-lm") __lowerCAmelCase : Tuple = WavaVecaForCTC.from_pretrained("patrickvonplaten/wav2vec2-base-100h-with-lm") # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train __lowerCAmelCase : Optional[Any] = processor(sample["audio"]["array"] , return_tensors="pt").input_values with torch.no_grad(): __lowerCAmelCase : int = model(_SCREAMING_SNAKE_CASE).logits.cpu().numpy() __lowerCAmelCase : Optional[int] = processor.decode(logits[0] , output_word_offsets=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate __lowerCAmelCase : List[str] = [ { "start_time": d["start_offset"] * time_offset, "end_time": d["end_offset"] * time_offset, "word": d["word"], } for d in output["word_offsets"] ] __lowerCAmelCase : List[str] = "WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL" # output words self.assertEqual(" ".join(self.get_from_offsets(_SCREAMING_SNAKE_CASE , "word")) , _SCREAMING_SNAKE_CASE) self.assertEqual(" ".join(self.get_from_offsets(_SCREAMING_SNAKE_CASE , "word")) , output.text) # output times __lowerCAmelCase : Any = torch.tensor(self.get_from_offsets(_SCREAMING_SNAKE_CASE , "start_time")) __lowerCAmelCase : Union[str, Any] = torch.tensor(self.get_from_offsets(_SCREAMING_SNAKE_CASE , "end_time")) # fmt: off __lowerCAmelCase : Union[str, Any] = torch.tensor([1.4199, 1.6599, 2.2599, 3.0, 3.24, 3.5999, 3.7999, 4.0999, 4.26, 4.94, 5.28, 5.6599, 5.78, 5.94, 6.32, 6.5399, 6.6599]) __lowerCAmelCase : Any = torch.tensor([1.5399, 1.8999, 2.9, 3.16, 3.5399, 3.72, 4.0199, 4.1799, 4.76, 5.1599, 5.5599, 5.6999, 5.86, 6.1999, 6.38, 6.6199, 6.94]) # fmt: on self.assertTrue(torch.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=0.01)) self.assertTrue(torch.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=0.01))
269
"""simple docstring""" import numpy as np from scipy.spatial.distance import cdist from sklearn.metrics import fa_score import datasets __snake_case : Optional[int] = '\\n @inproceedings{kakwani2020indicnlpsuite,\n title={{IndicNLPSuite: Monolingual Corpora, Evaluation Benchmarks and Pre-trained Multilingual Language Models for Indian Languages}},\n author={Divyanshu Kakwani and Anoop Kunchukuttan and Satish Golla and Gokul N.C. and Avik Bhattacharyya and Mitesh M. Khapra and Pratyush Kumar},\n year={2020},\n booktitle={Findings of EMNLP},\n}\n' __snake_case : str = '\\n IndicGLUE is a natural language understanding benchmark for Indian languages. It contains a wide\n variety of tasks and covers 11 major Indian languages - as, bn, gu, hi, kn, ml, mr, or, pa, ta, te.\n' __snake_case : str = '\nCompute IndicGLUE evaluation metric associated to each IndicGLUE dataset.\nArgs:\n predictions: list of predictions to score (as int64),\n except for \'cvit-mkb-clsr\' where each prediction is a vector (of float32).\n references: list of ground truth labels corresponding to the predictions (as int64),\n except for \'cvit-mkb-clsr\' where each reference is a vector (of float32).\nReturns: depending on the IndicGLUE subset, one or several of:\n "accuracy": Accuracy\n "f1": F1 score\n "precision": Precision@10\nExamples:\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wnli\') # \'wnli\' or any of ["copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md"]\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0}\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wiki-ner\')\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0, \'f1\': 1.0}\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'cvit-mkb-clsr\')\n >>> references = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]]\n >>> predictions = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'precision@10\': 1.0}\n\n' def _lowercase ( __snake_case ,__snake_case ) -> Union[str, Any]: return float((preds == labels).mean() ) def _lowercase ( __snake_case ,__snake_case ) -> str: __lowerCAmelCase : str = simple_accuracy(__snake_case ,__snake_case ) __lowerCAmelCase : Any = float(fa_score(y_true=__snake_case ,y_pred=__snake_case ) ) return { "accuracy": acc, "f1": fa, } def _lowercase ( __snake_case ,__snake_case ) -> int: __lowerCAmelCase : Union[str, Any] = np.array(__snake_case ) __lowerCAmelCase : Tuple = np.array(__snake_case ) __lowerCAmelCase : List[Any] = en_sentvecs.shape[0] # mean centering __lowerCAmelCase : Union[str, Any] = en_sentvecs - np.mean(__snake_case ,axis=0 ) __lowerCAmelCase : int = in_sentvecs - np.mean(__snake_case ,axis=0 ) __lowerCAmelCase : Optional[Any] = cdist(__snake_case ,__snake_case ,"cosine" ) __lowerCAmelCase : int = np.array(range(__snake_case ) ) __lowerCAmelCase : int = sim.argsort(axis=1 )[:, :10] __lowerCAmelCase : Optional[Any] = np.any(preds == actual[:, None] ,axis=1 ) return float(matches.mean() ) @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A__ ( datasets.Metric ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: int) -> str: """simple docstring""" if self.config_name not in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", "wiki-ner", ]: raise KeyError( "You should supply a configuration name selected in " "[\"wnli\", \"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", " "\"cvit-mkb-clsr\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\", " "\"wiki-ner\"]") return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("int64") if self.config_name != "cvit-mkb-clsr" else datasets.Sequence(datasets.Value("float32")), "references": datasets.Value("int64") if self.config_name != "cvit-mkb-clsr" else datasets.Sequence(datasets.Value("float32")), }) , codebase_urls=[] , reference_urls=[] , format="numpy" if self.config_name != "cvit-mkb-clsr" else None , ) def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Optional[Any]) -> int: """simple docstring""" if self.config_name == "cvit-mkb-clsr": return {"precision@10": precision_at_aa(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE)} elif self.config_name in ["wiki-ner"]: return acc_and_fa(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) elif self.config_name in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md", ]: return {"accuracy": simple_accuracy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE)} else: raise KeyError( "You should supply a configuration name selected in " "[\"wnli\", \"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", " "\"cvit-mkb-clsr\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\", " "\"wiki-ner\"]")
269
1
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_camembert import CamembertTokenizer else: __snake_case : Dict = None __snake_case : str = logging.get_logger(__name__) __snake_case : Optional[Any] = {'vocab_file': 'sentencepiece.bpe.model', 'tokenizer_file': 'tokenizer.json'} __snake_case : List[str] = { 'vocab_file': { 'camembert-base': 'https://huggingface.co/camembert-base/resolve/main/sentencepiece.bpe.model', }, 'tokenizer_file': { 'camembert-base': 'https://huggingface.co/camembert-base/resolve/main/tokenizer.json', }, } __snake_case : int = { 'camembert-base': 512, } __snake_case : Optional[Any] = '▁' class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE = ['input_ids', 'attention_mask'] SCREAMING_SNAKE_CASE = CamembertTokenizer def __init__( self: Dict , _SCREAMING_SNAKE_CASE: Any=None , _SCREAMING_SNAKE_CASE: Any=None , _SCREAMING_SNAKE_CASE: Optional[int]="<s>" , _SCREAMING_SNAKE_CASE: Dict="</s>" , _SCREAMING_SNAKE_CASE: Tuple="</s>" , _SCREAMING_SNAKE_CASE: str="<s>" , _SCREAMING_SNAKE_CASE: List[Any]="<unk>" , _SCREAMING_SNAKE_CASE: Dict="<pad>" , _SCREAMING_SNAKE_CASE: Dict="<mask>" , _SCREAMING_SNAKE_CASE: Optional[int]=["<s>NOTUSED", "</s>NOTUSED"] , **_SCREAMING_SNAKE_CASE: Tuple , ) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Union[str, Any] = AddedToken(_SCREAMING_SNAKE_CASE , lstrip=_SCREAMING_SNAKE_CASE , rstrip=_SCREAMING_SNAKE_CASE) if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) else mask_token super().__init__( _SCREAMING_SNAKE_CASE , tokenizer_file=_SCREAMING_SNAKE_CASE , bos_token=_SCREAMING_SNAKE_CASE , eos_token=_SCREAMING_SNAKE_CASE , sep_token=_SCREAMING_SNAKE_CASE , cls_token=_SCREAMING_SNAKE_CASE , unk_token=_SCREAMING_SNAKE_CASE , pad_token=_SCREAMING_SNAKE_CASE , mask_token=_SCREAMING_SNAKE_CASE , additional_special_tokens=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : int = vocab_file __lowerCAmelCase : Union[str, Any] = False if not self.vocab_file else True def _SCREAMING_SNAKE_CASE ( self: Optional[int] , _SCREAMING_SNAKE_CASE: List[int] , _SCREAMING_SNAKE_CASE: Optional[List[int]] = None) -> List[int]: """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] __lowerCAmelCase : Dict = [self.cls_token_id] __lowerCAmelCase : List[Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _SCREAMING_SNAKE_CASE ( self: Dict , _SCREAMING_SNAKE_CASE: List[int] , _SCREAMING_SNAKE_CASE: Optional[List[int]] = None) -> List[int]: """simple docstring""" __lowerCAmelCase : Tuple = [self.sep_token_id] __lowerCAmelCase : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0] def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: Optional[str] = None) -> Tuple[str]: """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer.") if not os.path.isdir(_SCREAMING_SNAKE_CASE): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""") return __lowerCAmelCase : Dict = os.path.join( _SCREAMING_SNAKE_CASE , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(_SCREAMING_SNAKE_CASE): copyfile(self.vocab_file , _SCREAMING_SNAKE_CASE) return (out_vocab_file,)
269
"""simple docstring""" def _lowercase ( __snake_case ,__snake_case ) -> float: if digit_amount > 0: return round(number - int(__snake_case ) ,__snake_case ) return number - int(__snake_case ) if __name__ == "__main__": print(decimal_isolate(1.53, 0)) print(decimal_isolate(35.345, 1)) print(decimal_isolate(35.345, 2)) print(decimal_isolate(35.345, 3)) print(decimal_isolate(-14.789, 3)) print(decimal_isolate(0, 2)) print(decimal_isolate(-14.123, 1)) print(decimal_isolate(-14.123, 2)) print(decimal_isolate(-14.123, 3))
269
1
"""simple docstring""" import os import re import sys import traceback import warnings from pathlib import Path from typing import Dict, Optional, Union from uuid import uuida from huggingface_hub import HfFolder, ModelCard, ModelCardData, hf_hub_download, whoami from huggingface_hub.file_download import REGEX_COMMIT_HASH from huggingface_hub.utils import ( EntryNotFoundError, RepositoryNotFoundError, RevisionNotFoundError, is_jinja_available, ) from packaging import version from requests import HTTPError from .. import __version__ from .constants import ( DEPRECATED_REVISION_ARGS, DIFFUSERS_CACHE, HUGGINGFACE_CO_RESOLVE_ENDPOINT, SAFETENSORS_WEIGHTS_NAME, WEIGHTS_NAME, ) from .import_utils import ( ENV_VARS_TRUE_VALUES, _flax_version, _jax_version, _onnxruntime_version, _torch_version, is_flax_available, is_onnx_available, is_torch_available, ) from .logging import get_logger __snake_case : List[Any] = get_logger(__name__) __snake_case : Union[str, Any] = Path(__file__).parent / 'model_card_template.md' __snake_case : Union[str, Any] = uuida().hex __snake_case : List[Any] = os.getenv('HF_HUB_OFFLINE', '').upper() in ENV_VARS_TRUE_VALUES __snake_case : str = os.getenv('DISABLE_TELEMETRY', '').upper() in ENV_VARS_TRUE_VALUES __snake_case : Dict = HUGGINGFACE_CO_RESOLVE_ENDPOINT + '/api/telemetry/' def _lowercase ( __snake_case = None ) -> str: __lowerCAmelCase : Union[str, Any] = F"""diffusers/{__version__}; python/{sys.version.split()[0]}; session_id/{SESSION_ID}""" if DISABLE_TELEMETRY or HF_HUB_OFFLINE: return ua + "; telemetry/off" if is_torch_available(): ua += F"""; torch/{_torch_version}""" if is_flax_available(): ua += F"""; jax/{_jax_version}""" ua += F"""; flax/{_flax_version}""" if is_onnx_available(): ua += F"""; onnxruntime/{_onnxruntime_version}""" # CI will set this value to True if os.environ.get("DIFFUSERS_IS_CI" ,"" ).upper() in ENV_VARS_TRUE_VALUES: ua += "; is_ci/true" if isinstance(__snake_case ,__snake_case ): ua += "; " + "; ".join(F"""{k}/{v}""" for k, v in user_agent.items() ) elif isinstance(__snake_case ,__snake_case ): ua += "; " + user_agent return ua def _lowercase ( __snake_case ,__snake_case = None ,__snake_case = None ) -> str: if token is None: __lowerCAmelCase : List[str] = HfFolder.get_token() if organization is None: __lowerCAmelCase : Any = whoami(__snake_case )["name"] return F"""{username}/{model_id}""" else: return F"""{organization}/{model_id}""" def _lowercase ( __snake_case ,__snake_case ) -> Dict: if not is_jinja_available(): raise ValueError( "Modelcard rendering is based on Jinja templates." " Please make sure to have `jinja` installed before using `create_model_card`." " To install it, please run `pip install Jinja2`." ) if hasattr(__snake_case ,"local_rank" ) and args.local_rank not in [-1, 0]: return __lowerCAmelCase : List[Any] = args.hub_token if hasattr(__snake_case ,"hub_token" ) else None __lowerCAmelCase : List[str] = get_full_repo_name(__snake_case ,token=__snake_case ) __lowerCAmelCase : Tuple = ModelCard.from_template( card_data=ModelCardData( # Card metadata object that will be converted to YAML block language="en" ,license="apache-2.0" ,library_name="diffusers" ,tags=[] ,datasets=args.dataset_name ,metrics=[] ,) ,template_path=__snake_case ,model_name=__snake_case ,repo_name=__snake_case ,dataset_name=args.dataset_name if hasattr(__snake_case ,"dataset_name" ) else None ,learning_rate=args.learning_rate ,train_batch_size=args.train_batch_size ,eval_batch_size=args.eval_batch_size ,gradient_accumulation_steps=( args.gradient_accumulation_steps if hasattr(__snake_case ,"gradient_accumulation_steps" ) else None ) ,adam_betaa=args.adam_betaa if hasattr(__snake_case ,"adam_beta1" ) else None ,adam_betaa=args.adam_betaa if hasattr(__snake_case ,"adam_beta2" ) else None ,adam_weight_decay=args.adam_weight_decay if hasattr(__snake_case ,"adam_weight_decay" ) else None ,adam_epsilon=args.adam_epsilon if hasattr(__snake_case ,"adam_epsilon" ) else None ,lr_scheduler=args.lr_scheduler if hasattr(__snake_case ,"lr_scheduler" ) else None ,lr_warmup_steps=args.lr_warmup_steps if hasattr(__snake_case ,"lr_warmup_steps" ) else None ,ema_inv_gamma=args.ema_inv_gamma if hasattr(__snake_case ,"ema_inv_gamma" ) else None ,ema_power=args.ema_power if hasattr(__snake_case ,"ema_power" ) else None ,ema_max_decay=args.ema_max_decay if hasattr(__snake_case ,"ema_max_decay" ) else None ,mixed_precision=args.mixed_precision ,) __lowerCAmelCase : List[str] = os.path.join(args.output_dir ,"README.md" ) model_card.save(__snake_case ) def _lowercase ( __snake_case ,__snake_case = None ) -> List[str]: if resolved_file is None or commit_hash is not None: return commit_hash __lowerCAmelCase : str = str(Path(__snake_case ).as_posix() ) __lowerCAmelCase : List[str] = re.search(r"snapshots/([^/]+)/" ,__snake_case ) if search is None: return None __lowerCAmelCase : int = search.groups()[0] return commit_hash if REGEX_COMMIT_HASH.match(__snake_case ) else None # Old default cache path, potentially to be migrated. # This logic was more or less taken from `transformers`, with the following differences: # - Diffusers doesn't use custom environment variables to specify the cache path. # - There is no need to migrate the cache format, just move the files to the new location. __snake_case : List[Any] = os.path.expanduser( os.getenv('HF_HOME', os.path.join(os.getenv('XDG_CACHE_HOME', '~/.cache'), 'huggingface')) ) __snake_case : List[Any] = os.path.join(hf_cache_home, 'diffusers') def _lowercase ( __snake_case = None ,__snake_case = None ) -> None: if new_cache_dir is None: __lowerCAmelCase : Tuple = DIFFUSERS_CACHE if old_cache_dir is None: __lowerCAmelCase : Optional[int] = old_diffusers_cache __lowerCAmelCase : Dict = Path(__snake_case ).expanduser() __lowerCAmelCase : Dict = Path(__snake_case ).expanduser() for old_blob_path in old_cache_dir.glob("**/blobs/*" ): if old_blob_path.is_file() and not old_blob_path.is_symlink(): __lowerCAmelCase : Tuple = new_cache_dir / old_blob_path.relative_to(__snake_case ) new_blob_path.parent.mkdir(parents=__snake_case ,exist_ok=__snake_case ) os.replace(__snake_case ,__snake_case ) try: os.symlink(__snake_case ,__snake_case ) except OSError: logger.warning( "Could not create symlink between old cache and new cache. If you use an older version of diffusers again, files will be re-downloaded." ) # At this point, old_cache_dir contains symlinks to the new cache (it can still be used). __snake_case : Any = os.path.join(DIFFUSERS_CACHE, 'version_diffusers_cache.txt') if not os.path.isfile(cache_version_file): __snake_case : Union[str, Any] = 0 else: with open(cache_version_file) as f: try: __snake_case : Optional[Any] = int(f.read()) except ValueError: __snake_case : str = 0 if cache_version < 1: __snake_case : List[str] = os.path.isdir(old_diffusers_cache) and len(os.listdir(old_diffusers_cache)) > 0 if old_cache_is_not_empty: logger.warning( 'The cache for model files in Diffusers v0.14.0 has moved to a new location. Moving your ' 'existing cached models. This is a one-time operation, you can interrupt it or run it ' 'later by calling `diffusers.utils.hub_utils.move_cache()`.' ) try: move_cache() except Exception as e: __snake_case : Dict = '\n'.join(traceback.format_tb(e.__traceback__)) logger.error( F"""There was a problem when trying to move your cache:\n\n{trace}\n{e.__class__.__name__}: {e}\n\nPlease """ 'file an issue at https://github.com/huggingface/diffusers/issues/new/choose, copy paste this whole ' 'message and we will do our best to help.' ) if cache_version < 1: try: os.makedirs(DIFFUSERS_CACHE, exist_ok=True) with open(cache_version_file, 'w') as f: f.write('1') except Exception: logger.warning( F"""There was a problem when trying to write in your cache folder ({DIFFUSERS_CACHE}). Please, ensure """ 'the directory exists and can be written to.' ) def _lowercase ( __snake_case ,__snake_case = None ) -> str: if variant is not None: __lowerCAmelCase : Tuple = weights_name.split("." ) __lowerCAmelCase : Optional[Any] = splits[:-1] + [variant] + splits[-1:] __lowerCAmelCase : Dict = ".".join(__snake_case ) return weights_name def _lowercase ( __snake_case ,*, __snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case=None ,) -> List[str]: __lowerCAmelCase : Optional[Any] = str(__snake_case ) if os.path.isfile(__snake_case ): return pretrained_model_name_or_path elif os.path.isdir(__snake_case ): if os.path.isfile(os.path.join(__snake_case ,__snake_case ) ): # Load from a PyTorch checkpoint __lowerCAmelCase : str = os.path.join(__snake_case ,__snake_case ) return model_file elif subfolder is not None and os.path.isfile( os.path.join(__snake_case ,__snake_case ,__snake_case ) ): __lowerCAmelCase : Union[str, Any] = os.path.join(__snake_case ,__snake_case ,__snake_case ) return model_file else: raise EnvironmentError( F"""Error no file named {weights_name} found in directory {pretrained_model_name_or_path}.""" ) else: # 1. First check if deprecated way of loading from branches is used if ( revision in DEPRECATED_REVISION_ARGS and (weights_name == WEIGHTS_NAME or weights_name == SAFETENSORS_WEIGHTS_NAME) and version.parse(version.parse(__snake_case ).base_version ) >= version.parse("0.20.0" ) ): try: __lowerCAmelCase : Union[str, Any] = hf_hub_download( __snake_case ,filename=_add_variant(__snake_case ,__snake_case ) ,cache_dir=__snake_case ,force_download=__snake_case ,proxies=__snake_case ,resume_download=__snake_case ,local_files_only=__snake_case ,use_auth_token=__snake_case ,user_agent=__snake_case ,subfolder=__snake_case ,revision=revision or commit_hash ,) warnings.warn( F"""Loading the variant {revision} from {pretrained_model_name_or_path} via `revision='{revision}'` is deprecated. Loading instead from `revision='main'` with `variant={revision}`. Loading model variants via `revision='{revision}'` will be removed in diffusers v1. Please use `variant='{revision}'` instead.""" ,__snake_case ,) return model_file except: # noqa: E722 warnings.warn( F"""You are loading the variant {revision} from {pretrained_model_name_or_path} via `revision='{revision}'`. This behavior is deprecated and will be removed in diffusers v1. One should use `variant='{revision}'` instead. However, it appears that {pretrained_model_name_or_path} currently does not have a {_add_variant(__snake_case ,__snake_case )} file in the 'main' branch of {pretrained_model_name_or_path}. \n The Diffusers team and community would be very grateful if you could open an issue: https://github.com/huggingface/diffusers/issues/new with the title '{pretrained_model_name_or_path} is missing {_add_variant(__snake_case ,__snake_case )}' so that the correct variant file can be added.""" ,__snake_case ,) try: # 2. Load model file as usual __lowerCAmelCase : Any = hf_hub_download( __snake_case ,filename=__snake_case ,cache_dir=__snake_case ,force_download=__snake_case ,proxies=__snake_case ,resume_download=__snake_case ,local_files_only=__snake_case ,use_auth_token=__snake_case ,user_agent=__snake_case ,subfolder=__snake_case ,revision=revision or commit_hash ,) return model_file except RepositoryNotFoundError: raise EnvironmentError( F"""{pretrained_model_name_or_path} is not a local folder and is not a valid model identifier """ "listed on 'https://huggingface.co/models'\nIf this is a private repository, make sure to pass a " "token having permission to this repo with `use_auth_token` or log in with `huggingface-cli " "login`." ) except RevisionNotFoundError: raise EnvironmentError( F"""{revision} is not a valid git identifier (branch name, tag name or commit id) that exists for """ "this model name. Check the model page at " F"""'https://huggingface.co/{pretrained_model_name_or_path}' for available revisions.""" ) except EntryNotFoundError: raise EnvironmentError( F"""{pretrained_model_name_or_path} does not appear to have a file named {weights_name}.""" ) except HTTPError as err: raise EnvironmentError( F"""There was a specific connection error when trying to load {pretrained_model_name_or_path}:\n{err}""" ) except ValueError: raise EnvironmentError( F"""We couldn't connect to '{HUGGINGFACE_CO_RESOLVE_ENDPOINT}' to load this model, couldn't find it""" F""" in the cached files and it looks like {pretrained_model_name_or_path} is not the path to a""" F""" directory containing a file named {weights_name} or""" " \nCheckout your internet connection or see how to run the library in" " offline mode at 'https://huggingface.co/docs/diffusers/installation#offline-mode'." ) except EnvironmentError: raise EnvironmentError( F"""Can't load the model for '{pretrained_model_name_or_path}'. If you were trying to load it from """ "'https://huggingface.co/models', make sure you don't have a local directory with the same name. " F"""Otherwise, make sure '{pretrained_model_name_or_path}' is the correct path to a directory """ F"""containing a file named {weights_name}""" )
269
"""simple docstring""" from ..utils import DummyObject, requires_backends class A__ ( metaclass=__SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = ['torch', 'torchsde'] def __init__( self: int , *_SCREAMING_SNAKE_CASE: Optional[Any] , **_SCREAMING_SNAKE_CASE: str) -> Dict: """simple docstring""" requires_backends(self , ["torch", "torchsde"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls: Optional[Any] , *_SCREAMING_SNAKE_CASE: Optional[int] , **_SCREAMING_SNAKE_CASE: Optional[int]) -> Optional[Any]: """simple docstring""" requires_backends(cls , ["torch", "torchsde"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls: Dict , *_SCREAMING_SNAKE_CASE: List[Any] , **_SCREAMING_SNAKE_CASE: Any) -> Optional[int]: """simple docstring""" requires_backends(cls , ["torch", "torchsde"])
269
1
"""simple docstring""" def _lowercase ( ) -> int: return 1 def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else two_pence(x - 2 ) + one_pence() def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else five_pence(x - 5 ) + two_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else ten_pence(x - 10 ) + five_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else twenty_pence(x - 20 ) + ten_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else fifty_pence(x - 50 ) + twenty_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else one_pound(x - 100 ) + fifty_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else two_pound(x - 200 ) + one_pound(__snake_case ) def _lowercase ( __snake_case = 200 ) -> int: return two_pound(__snake_case ) if __name__ == "__main__": print(solution(int(input().strip())))
269
"""simple docstring""" def _lowercase ( ) -> int: return 1 def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else two_pence(x - 2 ) + one_pence() def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else five_pence(x - 5 ) + two_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else ten_pence(x - 10 ) + five_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else twenty_pence(x - 20 ) + ten_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else fifty_pence(x - 50 ) + twenty_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else one_pound(x - 100 ) + fifty_pence(__snake_case ) def _lowercase ( __snake_case ) -> int: return 0 if x < 0 else two_pound(x - 200 ) + one_pound(__snake_case ) def _lowercase ( __snake_case = 200 ) -> int: return two_pound(__snake_case ) if __name__ == "__main__": print(solution(int(input().strip())))
269
1
"""simple docstring""" import random import unittest import numpy as np from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionImgaImgPipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline' def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: Optional[int]=0) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : Dict = floats_tensor((1, 3, 128, 128) , rng=random.Random(_SCREAMING_SNAKE_CASE)) __lowerCAmelCase : Optional[int] = np.random.RandomState(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = { "prompt": "A painting of a squirrel eating a burger", "image": image, "generator": generator, "num_inference_steps": 3, "strength": 0.75, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def _SCREAMING_SNAKE_CASE ( self: Dict) -> Optional[int]: """simple docstring""" __lowerCAmelCase : Optional[Any] = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider") pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = self.get_dummy_inputs() __lowerCAmelCase : List[str] = pipe(**_SCREAMING_SNAKE_CASE).images __lowerCAmelCase : List[str] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 128, 128, 3) __lowerCAmelCase : Optional[Any] = np.array([0.6_9643, 0.5_8484, 0.5_0314, 0.5_8760, 0.5_5368, 0.5_9643, 0.5_1529, 0.4_1217, 0.4_9087]) assert np.abs(image_slice - expected_slice).max() < 1e-1 def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> Any: """simple docstring""" __lowerCAmelCase : List[str] = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider") __lowerCAmelCase : Optional[Any] = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=_SCREAMING_SNAKE_CASE) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = self.get_dummy_inputs() __lowerCAmelCase : Any = pipe(**_SCREAMING_SNAKE_CASE).images __lowerCAmelCase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __lowerCAmelCase : str = np.array([0.6_1737, 0.5_4642, 0.5_3183, 0.5_4465, 0.5_2742, 0.6_0525, 0.4_9969, 0.4_0655, 0.4_8154]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-1 def _SCREAMING_SNAKE_CASE ( self: Any) -> int: """simple docstring""" __lowerCAmelCase : Dict = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider") __lowerCAmelCase : Union[str, Any] = LMSDiscreteScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) # warmup pass to apply optimizations __lowerCAmelCase : Union[str, Any] = pipe(**self.get_dummy_inputs()) __lowerCAmelCase : List[Any] = self.get_dummy_inputs() __lowerCAmelCase : List[str] = pipe(**_SCREAMING_SNAKE_CASE).images __lowerCAmelCase : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __lowerCAmelCase : str = np.array([0.5_2761, 0.5_9977, 0.4_9033, 0.4_9619, 0.5_4282, 0.5_0311, 0.4_7600, 0.4_0918, 0.4_5203]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-1 def _SCREAMING_SNAKE_CASE ( self: Dict) -> Any: """simple docstring""" __lowerCAmelCase : int = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider") __lowerCAmelCase : List[str] = EulerDiscreteScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = self.get_dummy_inputs() __lowerCAmelCase : Union[str, Any] = pipe(**_SCREAMING_SNAKE_CASE).images __lowerCAmelCase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __lowerCAmelCase : int = np.array([0.5_2911, 0.6_0004, 0.4_9229, 0.4_9805, 0.5_4502, 0.5_0680, 0.4_7777, 0.4_1028, 0.4_5304]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-1 def _SCREAMING_SNAKE_CASE ( self: str) -> str: """simple docstring""" __lowerCAmelCase : Dict = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider") __lowerCAmelCase : List[str] = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = self.get_dummy_inputs() __lowerCAmelCase : int = pipe(**_SCREAMING_SNAKE_CASE).images __lowerCAmelCase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __lowerCAmelCase : Any = np.array([0.5_2911, 0.6_0004, 0.4_9229, 0.4_9805, 0.5_4502, 0.5_0680, 0.4_7777, 0.4_1028, 0.4_5304]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-1 def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> Tuple: """simple docstring""" __lowerCAmelCase : Optional[int] = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider") __lowerCAmelCase : Optional[int] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = self.get_dummy_inputs() __lowerCAmelCase : Optional[Any] = pipe(**_SCREAMING_SNAKE_CASE).images __lowerCAmelCase : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __lowerCAmelCase : Tuple = np.array([0.6_5331, 0.5_8277, 0.4_8204, 0.5_6059, 0.5_3665, 0.5_6235, 0.5_0969, 0.4_0009, 0.4_6552]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-1 @nightly @require_onnxruntime @require_torch_gpu class A__ ( unittest.TestCase ): '''simple docstring''' @property def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Optional[Any]: """simple docstring""" return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> List[str]: """simple docstring""" __lowerCAmelCase : List[Any] = ort.SessionOptions() __lowerCAmelCase : int = False return options def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : str = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg") __lowerCAmelCase : Union[str, Any] = init_image.resize((768, 512)) # using the PNDM scheduler by default __lowerCAmelCase : Optional[int] = OnnxStableDiffusionImgaImgPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="onnx" , safety_checker=_SCREAMING_SNAKE_CASE , feature_extractor=_SCREAMING_SNAKE_CASE , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = "A fantasy landscape, trending on artstation" __lowerCAmelCase : Dict = np.random.RandomState(0) __lowerCAmelCase : Any = pipe( prompt=_SCREAMING_SNAKE_CASE , image=_SCREAMING_SNAKE_CASE , strength=0.75 , guidance_scale=7.5 , num_inference_steps=10 , generator=_SCREAMING_SNAKE_CASE , output_type="np" , ) __lowerCAmelCase : Optional[int] = output.images __lowerCAmelCase : Optional[int] = images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 768, 3) __lowerCAmelCase : int = np.array([0.4909, 0.5059, 0.5372, 0.4623, 0.4876, 0.5049, 0.4820, 0.4956, 0.5019]) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice).max() < 2e-2 def _SCREAMING_SNAKE_CASE ( self: Any) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[int] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg") __lowerCAmelCase : Dict = init_image.resize((768, 512)) __lowerCAmelCase : List[str] = LMSDiscreteScheduler.from_pretrained( "runwayml/stable-diffusion-v1-5" , subfolder="scheduler" , revision="onnx") __lowerCAmelCase : str = OnnxStableDiffusionImgaImgPipeline.from_pretrained( "runwayml/stable-diffusion-v1-5" , revision="onnx" , scheduler=_SCREAMING_SNAKE_CASE , safety_checker=_SCREAMING_SNAKE_CASE , feature_extractor=_SCREAMING_SNAKE_CASE , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = "A fantasy landscape, trending on artstation" __lowerCAmelCase : int = np.random.RandomState(0) __lowerCAmelCase : Dict = pipe( prompt=_SCREAMING_SNAKE_CASE , image=_SCREAMING_SNAKE_CASE , strength=0.75 , guidance_scale=7.5 , num_inference_steps=20 , generator=_SCREAMING_SNAKE_CASE , output_type="np" , ) __lowerCAmelCase : Union[str, Any] = output.images __lowerCAmelCase : Tuple = images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 768, 3) __lowerCAmelCase : Optional[Any] = np.array([0.8043, 0.926, 0.9581, 0.8119, 0.8954, 0.913, 0.7209, 0.7463, 0.7431]) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice).max() < 2e-2
269
"""simple docstring""" import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Optional[int] , _SCREAMING_SNAKE_CASE: str) -> Dict: """simple docstring""" with open(_SCREAMING_SNAKE_CASE , encoding="utf-8") as input_file: __lowerCAmelCase : List[str] = re.compile(r"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)") __lowerCAmelCase : List[Any] = input_file.read() __lowerCAmelCase : Any = regexp.search(_SCREAMING_SNAKE_CASE) return match def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: str) -> Optional[Any]: """simple docstring""" with open(_SCREAMING_SNAKE_CASE , encoding="utf-8") as input_file: __lowerCAmelCase : Any = re.compile(r"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL) __lowerCAmelCase : Optional[int] = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` __lowerCAmelCase : int = regexp.finditer(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = [match for match in matches if match is not None and match.group(1) is not None] return matches[0] if matches else None def _SCREAMING_SNAKE_CASE ( self: str) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[Any] = Path("./datasets") __lowerCAmelCase : Optional[int] = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_encoding_on_file_open(str(_SCREAMING_SNAKE_CASE)): raise AssertionError(F"""open(...) must use utf-8 encoding in {dataset}""") def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> Optional[int]: """simple docstring""" __lowerCAmelCase : Dict = Path("./datasets") __lowerCAmelCase : Union[str, Any] = list(dataset_paths.absolute().glob("**/*.py")) for dataset in dataset_files: if self._no_print_statements(str(_SCREAMING_SNAKE_CASE)): raise AssertionError(F"""print statement found in {dataset}. Use datasets.logger/logging instead.""")
269
1
"""simple docstring""" import argparse from torch import nn # transformers_old should correspond to branch `save_old_prophetnet_model_structure` here # original prophetnet_checkpoints are saved under `patrickvonplaten/..._old` respectively from transformers_old.modeling_prophetnet import ( ProphetNetForConditionalGeneration as ProphetNetForConditionalGenerationOld, ) from transformers_old.modeling_xlm_prophetnet import ( XLMProphetNetForConditionalGeneration as XLMProphetNetForConditionalGenerationOld, ) from transformers import ProphetNetForConditionalGeneration, XLMProphetNetForConditionalGeneration, logging __snake_case : Union[str, Any] = logging.get_logger(__name__) logging.set_verbosity_info() def _lowercase ( __snake_case ,__snake_case ) -> Tuple: if "xprophetnet" in prophetnet_checkpoint_path: __lowerCAmelCase : Dict = XLMProphetNetForConditionalGenerationOld.from_pretrained(__snake_case ) __lowerCAmelCase , __lowerCAmelCase : Optional[int] = XLMProphetNetForConditionalGeneration.from_pretrained( __snake_case ,output_loading_info=__snake_case ) else: __lowerCAmelCase : Optional[int] = ProphetNetForConditionalGenerationOld.from_pretrained(__snake_case ) __lowerCAmelCase , __lowerCAmelCase : Dict = ProphetNetForConditionalGeneration.from_pretrained( __snake_case ,output_loading_info=__snake_case ) __lowerCAmelCase : List[str] = ["key_proj", "value_proj", "query_proj"] __lowerCAmelCase : Dict = { "self_attn": "ngram_self_attn", "cross_attn": "encoder_attn", "cross_attn_layer_norm": "encoder_attn_layer_norm", "feed_forward_layer_norm": "final_layer_norm", "feed_forward": "", "intermediate": "fc1", "output": "fc2", "key_proj": "k_proj", "query_proj": "q_proj", "value_proj": "v_proj", "word_embeddings": "embed_tokens", "embeddings_layer_norm": "emb_layer_norm", "relative_pos_embeddings": "relative_linear", "ngram_embeddings": "ngram_input_embed", "position_embeddings": "embed_positions", } for key in loading_info["missing_keys"]: __lowerCAmelCase : int = key.split("." ) if attributes[0] == "lm_head": __lowerCAmelCase : Union[str, Any] = prophet __lowerCAmelCase : Optional[int] = prophet_old else: __lowerCAmelCase : Optional[Any] = prophet.prophetnet __lowerCAmelCase : int = prophet_old.model __lowerCAmelCase : Tuple = False for attribute in attributes: if attribute in mapping: __lowerCAmelCase : Dict = mapping[attribute] if not hasattr(__snake_case ,__snake_case ) and len(__snake_case ) > 0: __lowerCAmelCase : int = attribute elif hasattr(__snake_case ,__snake_case ): __lowerCAmelCase : List[str] = attribute if attribute == "weight": assert old_model.weight.shape == model.weight.shape, "Shapes have to match!" __lowerCAmelCase : List[Any] = old_model.weight logger.info(F"""{attribute} is initialized.""" ) __lowerCAmelCase : Tuple = True break elif attribute == "bias": assert old_model.bias.shape == model.bias.shape, "Shapes have to match!" __lowerCAmelCase : str = old_model.bias logger.info(F"""{attribute} is initialized""" ) __lowerCAmelCase : int = True break elif attribute in special_keys and hasattr(__snake_case ,"in_proj_weight" ): __lowerCAmelCase : Tuple = old_model.in_proj_weight.shape[0] // 3 __lowerCAmelCase : Dict = getattr(__snake_case ,__snake_case ) param.weight.shape == old_model.in_proj_weight[:embed_dim, :].shape, "Shapes have to match" param.bias.shape == old_model.in_proj_bias[:embed_dim].shape, "Shapes have to match" if attribute == "query_proj": __lowerCAmelCase : Dict = nn.Parameter(old_model.in_proj_weight[:embed_dim, :] ) __lowerCAmelCase : List[str] = nn.Parameter(old_model.in_proj_bias[:embed_dim] ) elif attribute == "key_proj": __lowerCAmelCase : Dict = nn.Parameter(old_model.in_proj_weight[embed_dim : 2 * embed_dim, :] ) __lowerCAmelCase : List[Any] = nn.Parameter(old_model.in_proj_bias[embed_dim : 2 * embed_dim] ) elif attribute == "value_proj": __lowerCAmelCase : Any = nn.Parameter(old_model.in_proj_weight[2 * embed_dim :, :] ) __lowerCAmelCase : Optional[int] = nn.Parameter(old_model.in_proj_bias[2 * embed_dim :] ) __lowerCAmelCase : Optional[Any] = True break elif attribute == "position_embeddings": assert ( model.position_embeddings.weight.shape[-1] == old_model.embed_positions.weight.shape[-1] ), "Hidden size has to match" assert model.position_embeddings.weight.shape[0] == 512, "We want 512 position_embeddings." __lowerCAmelCase : str = nn.Parameter(old_model.embed_positions.weight[:512, :] ) __lowerCAmelCase : str = True break if attribute.isdigit(): __lowerCAmelCase : Optional[int] = model[int(__snake_case )] __lowerCAmelCase : Optional[Any] = old_model[int(__snake_case )] else: __lowerCAmelCase : str = getattr(__snake_case ,__snake_case ) if old_attribute == "": __lowerCAmelCase : Any = old_model else: if not hasattr(__snake_case ,__snake_case ): raise ValueError(F"""{old_model} does not have {old_attribute}""" ) __lowerCAmelCase : Dict = getattr(__snake_case ,__snake_case ) if not is_key_init: raise ValueError(F"""{key} was not correctly initialized!""" ) print(F"""Saving model to {pytorch_dump_folder_path}""" ) prophet.save_pretrained(__snake_case ) if __name__ == "__main__": __snake_case : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--prophetnet_checkpoint_path', default=None, type=str, required=True, help='Path the official PyTorch dump.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) __snake_case : Union[str, Any] = parser.parse_args() convert_prophetnet_checkpoint_to_pytorch(args.prophetnet_checkpoint_path, args.pytorch_dump_folder_path)
269
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxSeqaSeqConfigWithPast from ...utils import logging if TYPE_CHECKING: from ...feature_extraction_utils import FeatureExtractionMixin from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType __snake_case : Optional[Any] = logging.get_logger(__name__) __snake_case : Optional[Any] = { 'openai/whisper-base': 'https://huggingface.co/openai/whisper-base/resolve/main/config.json', } # fmt: off __snake_case : Any = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 357, 366, 438, 532, 685, 705, 796, 930, 1_058, 1_220, 1_267, 1_279, 1_303, 1_343, 1_377, 1_391, 1_635, 1_782, 1_875, 2_162, 2_361, 2_488, 3_467, 4_008, 4_211, 4_600, 4_808, 5_299, 5_855, 6_329, 7_203, 9_609, 9_959, 10_563, 10_786, 11_420, 11_709, 11_907, 13_163, 13_697, 13_700, 14_808, 15_306, 16_410, 16_791, 17_992, 19_203, 19_510, 20_724, 22_305, 22_935, 27_007, 30_109, 30_420, 33_409, 34_949, 40_283, 40_493, 40_549, 47_282, 49_146, 50_257, 50_359, 50_360, 50_361 ] __snake_case : str = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 359, 503, 522, 542, 873, 893, 902, 918, 922, 931, 1_350, 1_853, 1_982, 2_460, 2_627, 3_246, 3_253, 3_268, 3_536, 3_846, 3_961, 4_183, 4_667, 6_585, 6_647, 7_273, 9_061, 9_383, 10_428, 10_929, 11_938, 12_033, 12_331, 12_562, 13_793, 14_157, 14_635, 15_265, 15_618, 16_553, 16_604, 18_362, 18_956, 20_075, 21_675, 22_520, 26_130, 26_161, 26_435, 28_279, 29_464, 31_650, 32_302, 32_470, 36_865, 42_863, 47_425, 49_870, 50_254, 50_258, 50_360, 50_361, 50_362 ] class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'whisper' SCREAMING_SNAKE_CASE = ['past_key_values'] SCREAMING_SNAKE_CASE = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self: Optional[int] , _SCREAMING_SNAKE_CASE: Any=5_1865 , _SCREAMING_SNAKE_CASE: Optional[Any]=80 , _SCREAMING_SNAKE_CASE: Optional[int]=6 , _SCREAMING_SNAKE_CASE: Any=4 , _SCREAMING_SNAKE_CASE: Dict=6 , _SCREAMING_SNAKE_CASE: Dict=4 , _SCREAMING_SNAKE_CASE: Optional[Any]=1536 , _SCREAMING_SNAKE_CASE: List[str]=1536 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.0 , _SCREAMING_SNAKE_CASE: str=0.0 , _SCREAMING_SNAKE_CASE: List[str]=5_0257 , _SCREAMING_SNAKE_CASE: Optional[Any]=True , _SCREAMING_SNAKE_CASE: List[str]=True , _SCREAMING_SNAKE_CASE: Optional[int]="gelu" , _SCREAMING_SNAKE_CASE: Tuple=256 , _SCREAMING_SNAKE_CASE: str=0.0 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.0 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.0 , _SCREAMING_SNAKE_CASE: List[Any]=0.02 , _SCREAMING_SNAKE_CASE: Tuple=False , _SCREAMING_SNAKE_CASE: Union[str, Any]=1500 , _SCREAMING_SNAKE_CASE: str=448 , _SCREAMING_SNAKE_CASE: Any=5_0256 , _SCREAMING_SNAKE_CASE: Any=5_0256 , _SCREAMING_SNAKE_CASE: List[str]=5_0256 , _SCREAMING_SNAKE_CASE: Dict=None , _SCREAMING_SNAKE_CASE: List[Any]=[220, 5_0256] , _SCREAMING_SNAKE_CASE: Dict=False , _SCREAMING_SNAKE_CASE: str=256 , _SCREAMING_SNAKE_CASE: List[str]=False , _SCREAMING_SNAKE_CASE: Tuple=0.05 , _SCREAMING_SNAKE_CASE: List[str]=10 , _SCREAMING_SNAKE_CASE: str=2 , _SCREAMING_SNAKE_CASE: Any=0.0 , _SCREAMING_SNAKE_CASE: Optional[int]=10 , _SCREAMING_SNAKE_CASE: int=0 , _SCREAMING_SNAKE_CASE: Any=7 , **_SCREAMING_SNAKE_CASE: List[str] , ) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[int] = vocab_size __lowerCAmelCase : Optional[Any] = num_mel_bins __lowerCAmelCase : int = d_model __lowerCAmelCase : List[Any] = encoder_layers __lowerCAmelCase : List[Any] = encoder_attention_heads __lowerCAmelCase : List[str] = decoder_layers __lowerCAmelCase : Tuple = decoder_attention_heads __lowerCAmelCase : Any = decoder_ffn_dim __lowerCAmelCase : Tuple = encoder_ffn_dim __lowerCAmelCase : List[str] = dropout __lowerCAmelCase : Union[str, Any] = attention_dropout __lowerCAmelCase : Union[str, Any] = activation_dropout __lowerCAmelCase : Dict = activation_function __lowerCAmelCase : Tuple = init_std __lowerCAmelCase : str = encoder_layerdrop __lowerCAmelCase : int = decoder_layerdrop __lowerCAmelCase : Optional[int] = use_cache __lowerCAmelCase : Union[str, Any] = encoder_layers __lowerCAmelCase : List[str] = scale_embedding # scale factor will be sqrt(d_model) if True __lowerCAmelCase : int = max_source_positions __lowerCAmelCase : Any = max_target_positions # Audio Classification-specific parameters. Feel free to ignore for other classes. __lowerCAmelCase : Dict = classifier_proj_size __lowerCAmelCase : Dict = use_weighted_layer_sum # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __lowerCAmelCase : int = apply_spec_augment __lowerCAmelCase : Union[str, Any] = mask_time_prob __lowerCAmelCase : str = mask_time_length __lowerCAmelCase : int = mask_time_min_masks __lowerCAmelCase : List[Any] = mask_feature_prob __lowerCAmelCase : Tuple = mask_feature_length __lowerCAmelCase : Any = mask_feature_min_masks __lowerCAmelCase : Union[str, Any] = median_filter_width super().__init__( pad_token_id=_SCREAMING_SNAKE_CASE , bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE , is_encoder_decoder=_SCREAMING_SNAKE_CASE , decoder_start_token_id=_SCREAMING_SNAKE_CASE , suppress_tokens=_SCREAMING_SNAKE_CASE , begin_suppress_tokens=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' @property def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> Mapping[str, Mapping[int, str]]: """simple docstring""" __lowerCAmelCase : List[str] = OrderedDict( [ ("input_features", {0: "batch", 1: "feature_size", 2: "encoder_sequence"}), ]) if self.use_past: __lowerCAmelCase : Tuple = {0: "batch"} else: __lowerCAmelCase : Union[str, Any] = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(_SCREAMING_SNAKE_CASE , direction="inputs") return common_inputs def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: Union["PreTrainedTokenizerBase", "FeatureExtractionMixin"] , _SCREAMING_SNAKE_CASE: int = -1 , _SCREAMING_SNAKE_CASE: int = -1 , _SCREAMING_SNAKE_CASE: bool = False , _SCREAMING_SNAKE_CASE: Optional["TensorType"] = None , _SCREAMING_SNAKE_CASE: int = 2_2050 , _SCREAMING_SNAKE_CASE: float = 5.0 , _SCREAMING_SNAKE_CASE: int = 220 , ) -> Mapping[str, Any]: """simple docstring""" __lowerCAmelCase : int = OrderedDict() __lowerCAmelCase : Optional[Any] = OnnxConfig.generate_dummy_inputs( self , preprocessor=preprocessor.feature_extractor , batch_size=_SCREAMING_SNAKE_CASE , framework=_SCREAMING_SNAKE_CASE , sampling_rate=_SCREAMING_SNAKE_CASE , time_duration=_SCREAMING_SNAKE_CASE , frequency=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : List[str] = encoder_inputs["input_features"].shape[2] __lowerCAmelCase : List[str] = encoder_sequence_length // 2 if self.use_past else seq_length __lowerCAmelCase : List[Any] = super().generate_dummy_inputs( preprocessor.tokenizer , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = encoder_inputs.pop("input_features") __lowerCAmelCase : List[Any] = decoder_inputs.pop("decoder_input_ids") if "past_key_values" in decoder_inputs: __lowerCAmelCase : int = decoder_inputs.pop("past_key_values") return dummy_inputs @property def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> float: """simple docstring""" return 1e-3
269
1
"""simple docstring""" import numpy as np from cva import destroyAllWindows, imread, imshow, waitKey class A__ : '''simple docstring''' def __init__( self: Dict , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: int) -> Optional[int]: """simple docstring""" if dst_width < 0 or dst_height < 0: raise ValueError("Destination width/height should be > 0") __lowerCAmelCase : str = img __lowerCAmelCase : Dict = img.shape[1] __lowerCAmelCase : Tuple = img.shape[0] __lowerCAmelCase : Union[str, Any] = dst_width __lowerCAmelCase : List[Any] = dst_height __lowerCAmelCase : List[Any] = self.src_w / self.dst_w __lowerCAmelCase : Dict = self.src_h / self.dst_h __lowerCAmelCase : int = ( np.ones((self.dst_h, self.dst_w, 3) , np.uinta) * 255 ) def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> List[Any]: """simple docstring""" for i in range(self.dst_h): for j in range(self.dst_w): __lowerCAmelCase : Tuple = self.img[self.get_y(_SCREAMING_SNAKE_CASE)][self.get_x(_SCREAMING_SNAKE_CASE)] def _SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: int) -> int: """simple docstring""" return int(self.ratio_x * x) def _SCREAMING_SNAKE_CASE ( self: str , _SCREAMING_SNAKE_CASE: int) -> int: """simple docstring""" return int(self.ratio_y * y) if __name__ == "__main__": __snake_case , __snake_case : int = 800, 600 __snake_case : Dict = imread('image_data/lena.jpg', 1) __snake_case : Union[str, Any] = NearestNeighbour(im, dst_w, dst_h) n.process() imshow( F"""Image resized from: {im.shape[1]}x{im.shape[0]} to {dst_w}x{dst_h}""", n.output ) waitKey(0) destroyAllWindows()
269
"""simple docstring""" import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration __snake_case : Optional[int] = 50_000 __snake_case : Dict = 5_000 __snake_case , __snake_case : Union[str, Any] = os.path.split(__file__) __snake_case : Any = os.path.join(RESULTS_BASEPATH, 'results', RESULTS_FILENAME.replace('.py', '.json')) @get_duration def _lowercase ( __snake_case ,__snake_case ) -> Dict: for i in range(__snake_case ): __lowerCAmelCase : Union[str, Any] = dataset[i] @get_duration def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> Dict: for i in range(0 ,len(__snake_case ) ,__snake_case ): __lowerCAmelCase : List[str] = dataset[i : i + batch_size] @get_duration def _lowercase ( __snake_case ,__snake_case ,__snake_case ) -> Dict: with dataset.formatted_as(type=__snake_case ): for i in range(__snake_case ): __lowerCAmelCase : Union[str, Any] = dataset[i] @get_duration def _lowercase ( __snake_case ,__snake_case ,__snake_case ,__snake_case ) -> str: with dataset.formatted_as(type=__snake_case ): for i in range(0 ,__snake_case ,__snake_case ): __lowerCAmelCase : Optional[int] = dataset[i : i + batch_size] def _lowercase ( ) -> Union[str, Any]: __lowerCAmelCase : Optional[int] = {"num examples": SPEED_TEST_N_EXAMPLES} __lowerCAmelCase : Optional[int] = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted, {"type": "pandas", "length": SMALL_TEST}), (read_formatted, {"type": "torch", "length": SMALL_TEST}), (read_formatted, {"type": "tensorflow", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] __lowerCAmelCase : Any = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1_000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1_000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print("generating dataset" ) __lowerCAmelCase : int = datasets.Features( {"list": datasets.Sequence(datasets.Value("float32" ) ), "numbers": datasets.Value("float32" )} ) __lowerCAmelCase : str = generate_example_dataset( os.path.join(__snake_case ,"dataset.arrow" ) ,__snake_case ,num_examples=__snake_case ,seq_shapes={"list": (100,)} ,) print("first set of iterations" ) for func, kwargs in functions: print(func.__name__ ,str(__snake_case ) ) __lowerCAmelCase : str = func(__snake_case ,**__snake_case ) print("shuffling dataset" ) __lowerCAmelCase : Optional[int] = dataset.shuffle() print("Second set of iterations (after shuffling" ) for func, kwargs in functions_shuffled: print("shuffled " ,func.__name__ ,str(__snake_case ) ) __lowerCAmelCase : List[str] = func( __snake_case ,**__snake_case ) with open(__snake_case ,"wb" ) as f: f.write(json.dumps(__snake_case ).encode("utf-8" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
269
1
"""simple docstring""" import unittest import numpy as np from diffusers import LMSDiscreteScheduler, OnnxStableDiffusionInpaintPipeline from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' pass @nightly @require_onnxruntime @require_torch_gpu class A__ ( unittest.TestCase ): '''simple docstring''' @property def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> Dict: """simple docstring""" return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def _SCREAMING_SNAKE_CASE ( self: List[str]) -> Optional[int]: """simple docstring""" __lowerCAmelCase : Tuple = ort.SessionOptions() __lowerCAmelCase : List[str] = False return options def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> Tuple: """simple docstring""" __lowerCAmelCase : int = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/in_paint/overture-creations-5sI6fQgYIuo.png") __lowerCAmelCase : List[str] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/in_paint/overture-creations-5sI6fQgYIuo_mask.png") __lowerCAmelCase : Optional[Any] = OnnxStableDiffusionInpaintPipeline.from_pretrained( "runwayml/stable-diffusion-inpainting" , revision="onnx" , safety_checker=_SCREAMING_SNAKE_CASE , feature_extractor=_SCREAMING_SNAKE_CASE , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = "A red cat sitting on a park bench" __lowerCAmelCase : List[Any] = np.random.RandomState(0) __lowerCAmelCase : int = pipe( prompt=_SCREAMING_SNAKE_CASE , image=_SCREAMING_SNAKE_CASE , mask_image=_SCREAMING_SNAKE_CASE , guidance_scale=7.5 , num_inference_steps=10 , generator=_SCREAMING_SNAKE_CASE , output_type="np" , ) __lowerCAmelCase : Optional[Any] = output.images __lowerCAmelCase : Optional[int] = images[0, 255:258, 255:258, -1] assert images.shape == (1, 512, 512, 3) __lowerCAmelCase : Union[str, Any] = np.array([0.2514, 0.3007, 0.3517, 0.1790, 0.2382, 0.3167, 0.1944, 0.2273, 0.2464]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-3 def _SCREAMING_SNAKE_CASE ( self: Dict) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Any = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/in_paint/overture-creations-5sI6fQgYIuo.png") __lowerCAmelCase : int = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/in_paint/overture-creations-5sI6fQgYIuo_mask.png") __lowerCAmelCase : Dict = LMSDiscreteScheduler.from_pretrained( "runwayml/stable-diffusion-inpainting" , subfolder="scheduler" , revision="onnx") __lowerCAmelCase : List[str] = OnnxStableDiffusionInpaintPipeline.from_pretrained( "runwayml/stable-diffusion-inpainting" , revision="onnx" , scheduler=_SCREAMING_SNAKE_CASE , safety_checker=_SCREAMING_SNAKE_CASE , feature_extractor=_SCREAMING_SNAKE_CASE , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = "A red cat sitting on a park bench" __lowerCAmelCase : List[Any] = np.random.RandomState(0) __lowerCAmelCase : Dict = pipe( prompt=_SCREAMING_SNAKE_CASE , image=_SCREAMING_SNAKE_CASE , mask_image=_SCREAMING_SNAKE_CASE , guidance_scale=7.5 , num_inference_steps=20 , generator=_SCREAMING_SNAKE_CASE , output_type="np" , ) __lowerCAmelCase : Optional[int] = output.images __lowerCAmelCase : int = images[0, 255:258, 255:258, -1] assert images.shape == (1, 512, 512, 3) __lowerCAmelCase : int = np.array([0.0086, 0.0077, 0.0083, 0.0093, 0.0107, 0.0139, 0.0094, 0.0097, 0.0125]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-3
269
"""simple docstring""" import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self: Tuple , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: List[str]=13 , _SCREAMING_SNAKE_CASE: Tuple=7 , _SCREAMING_SNAKE_CASE: int=True , _SCREAMING_SNAKE_CASE: Optional[Any]=True , _SCREAMING_SNAKE_CASE: str=False , _SCREAMING_SNAKE_CASE: Optional[Any]=True , _SCREAMING_SNAKE_CASE: int=99 , _SCREAMING_SNAKE_CASE: int=32 , _SCREAMING_SNAKE_CASE: List[str]=5 , _SCREAMING_SNAKE_CASE: Union[str, Any]=4 , _SCREAMING_SNAKE_CASE: int=64 , _SCREAMING_SNAKE_CASE: List[str]="gelu" , _SCREAMING_SNAKE_CASE: str=0.1 , _SCREAMING_SNAKE_CASE: Any=0.1 , _SCREAMING_SNAKE_CASE: Optional[int]=512 , _SCREAMING_SNAKE_CASE: Tuple=16 , _SCREAMING_SNAKE_CASE: Any=2 , _SCREAMING_SNAKE_CASE: List[str]=0.02 , _SCREAMING_SNAKE_CASE: Tuple=3 , _SCREAMING_SNAKE_CASE: Optional[Any]=4 , _SCREAMING_SNAKE_CASE: int=None , _SCREAMING_SNAKE_CASE: int=2 , _SCREAMING_SNAKE_CASE: str=2 , _SCREAMING_SNAKE_CASE: Union[str, Any]=2 , _SCREAMING_SNAKE_CASE: List[Any]=2 , _SCREAMING_SNAKE_CASE: int=4 , _SCREAMING_SNAKE_CASE: List[str]=1 , ) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : List[str] = parent __lowerCAmelCase : Optional[Any] = batch_size __lowerCAmelCase : Union[str, Any] = seq_length __lowerCAmelCase : Optional[Any] = is_training __lowerCAmelCase : Optional[int] = use_input_mask __lowerCAmelCase : Dict = use_token_type_ids __lowerCAmelCase : Dict = use_labels __lowerCAmelCase : Dict = vocab_size __lowerCAmelCase : Tuple = hidden_size __lowerCAmelCase : List[Any] = num_hidden_layers __lowerCAmelCase : Union[str, Any] = num_attention_heads __lowerCAmelCase : Tuple = intermediate_size __lowerCAmelCase : List[Any] = hidden_act __lowerCAmelCase : Optional[Any] = hidden_dropout_prob __lowerCAmelCase : Optional[Any] = attention_probs_dropout_prob __lowerCAmelCase : Optional[int] = max_position_embeddings __lowerCAmelCase : Union[str, Any] = type_vocab_size __lowerCAmelCase : Optional[int] = type_sequence_label_size __lowerCAmelCase : Dict = initializer_range __lowerCAmelCase : Tuple = num_labels __lowerCAmelCase : Optional[Any] = num_choices __lowerCAmelCase : Union[str, Any] = scope __lowerCAmelCase : Optional[Any] = q_groups __lowerCAmelCase : Optional[int] = k_groups __lowerCAmelCase : Any = v_groups __lowerCAmelCase : int = post_attention_groups __lowerCAmelCase : List[str] = intermediate_groups __lowerCAmelCase : Optional[Any] = output_groups def _SCREAMING_SNAKE_CASE ( self: Dict) -> List[str]: """simple docstring""" __lowerCAmelCase : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) __lowerCAmelCase : Union[str, Any] = None if self.use_input_mask: __lowerCAmelCase : Tuple = random_attention_mask([self.batch_size, self.seq_length]) __lowerCAmelCase : Optional[int] = None __lowerCAmelCase : List[Any] = None __lowerCAmelCase : str = None if self.use_labels: __lowerCAmelCase : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size) __lowerCAmelCase : str = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) __lowerCAmelCase : str = ids_tensor([self.batch_size] , self.num_choices) __lowerCAmelCase : Any = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _SCREAMING_SNAKE_CASE ( self: Tuple) -> int: """simple docstring""" return SqueezeBertConfig( embedding_size=self.hidden_size , vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , attention_probs_dropout_prob=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , q_groups=self.q_groups , k_groups=self.k_groups , v_groups=self.v_groups , post_attention_groups=self.post_attention_groups , intermediate_groups=self.intermediate_groups , output_groups=self.output_groups , ) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Tuple , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Tuple) -> Tuple: """simple docstring""" __lowerCAmelCase : List[Any] = SqueezeBertModel(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : List[Any] = model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = model(_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Tuple) -> Dict: """simple docstring""" __lowerCAmelCase : int = SqueezeBertForMaskedLM(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : Dict = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def _SCREAMING_SNAKE_CASE ( self: Optional[int] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Union[str, Any]) -> int: """simple docstring""" __lowerCAmelCase : str = SqueezeBertForQuestionAnswering(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : Union[str, Any] = model( _SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , start_positions=_SCREAMING_SNAKE_CASE , end_positions=_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: Tuple , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: Tuple) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : List[Any] = self.num_labels __lowerCAmelCase : Union[str, Any] = SqueezeBertForSequenceClassification(_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : int = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[int]) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : Dict = self.num_labels __lowerCAmelCase : Optional[int] = SqueezeBertForTokenClassification(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : List[str] = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: int) -> Tuple: """simple docstring""" __lowerCAmelCase : List[str] = self.num_choices __lowerCAmelCase : str = SqueezeBertForMultipleChoice(config=_SCREAMING_SNAKE_CASE) model.to(_SCREAMING_SNAKE_CASE) model.eval() __lowerCAmelCase : int = input_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() __lowerCAmelCase : Union[str, Any] = input_mask.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() __lowerCAmelCase : str = model( _SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices)) def _SCREAMING_SNAKE_CASE ( self: str) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[int] = self.prepare_config_and_inputs() ((__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase)) : Union[str, Any] = config_and_inputs __lowerCAmelCase : int = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) SCREAMING_SNAKE_CASE = ( { 'feature-extraction': SqueezeBertModel, 'fill-mask': SqueezeBertForMaskedLM, 'question-answering': SqueezeBertForQuestionAnswering, 'text-classification': SqueezeBertForSequenceClassification, 'token-classification': SqueezeBertForTokenClassification, 'zero-shot': SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = False def _SCREAMING_SNAKE_CASE ( self: Dict) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Any = SqueezeBertModelTester(self) __lowerCAmelCase : Optional[int] = ConfigTester(self , config_class=_SCREAMING_SNAKE_CASE , dim=37) def _SCREAMING_SNAKE_CASE ( self: Any) -> Tuple: """simple docstring""" self.config_tester.run_common_tests() def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> Any: """simple docstring""" __lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> List[Any]: """simple docstring""" __lowerCAmelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> int: """simple docstring""" __lowerCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Any) -> int: """simple docstring""" __lowerCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> str: """simple docstring""" __lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*_SCREAMING_SNAKE_CASE) @slow def _SCREAMING_SNAKE_CASE ( self: Any) -> Dict: """simple docstring""" for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase : Optional[Any] = SqueezeBertModel.from_pretrained(_SCREAMING_SNAKE_CASE) self.assertIsNotNone(_SCREAMING_SNAKE_CASE) @require_sentencepiece @require_tokenizers @require_torch class A__ ( unittest.TestCase ): '''simple docstring''' @slow def _SCREAMING_SNAKE_CASE ( self: int) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[Any] = SqueezeBertForSequenceClassification.from_pretrained("squeezebert/squeezebert-mnli") __lowerCAmelCase : List[Any] = torch.tensor([[1, 2_9414, 232, 328, 740, 1140, 1_2695, 69, 13, 1588, 2]]) __lowerCAmelCase : List[Any] = model(_SCREAMING_SNAKE_CASE)[0] __lowerCAmelCase : Any = torch.Size((1, 3)) self.assertEqual(output.shape , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = torch.tensor([[0.6401, -0.0349, -0.6041]]) self.assertTrue(torch.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=1e-4))
269
1
"""simple docstring""" from __future__ import annotations class A__ : '''simple docstring''' def __init__( self: List[Any] , _SCREAMING_SNAKE_CASE: int) -> None: """simple docstring""" __lowerCAmelCase : List[str] = order # a_{0} ... a_{k} __lowerCAmelCase : Tuple = [1.0] + [0.0] * order # b_{0} ... b_{k} __lowerCAmelCase : Tuple = [1.0] + [0.0] * order # x[n-1] ... x[n-k] __lowerCAmelCase : Any = [0.0] * self.order # y[n-1] ... y[n-k] __lowerCAmelCase : List[Any] = [0.0] * self.order def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: list[float] , _SCREAMING_SNAKE_CASE: list[float]) -> None: """simple docstring""" if len(_SCREAMING_SNAKE_CASE) < self.order: __lowerCAmelCase : Tuple = [1.0, *a_coeffs] if len(_SCREAMING_SNAKE_CASE) != self.order + 1: __lowerCAmelCase : Dict = ( F"""Expected a_coeffs to have {self.order + 1} elements """ F"""for {self.order}-order filter, got {len(_SCREAMING_SNAKE_CASE)}""" ) raise ValueError(_SCREAMING_SNAKE_CASE) if len(_SCREAMING_SNAKE_CASE) != self.order + 1: __lowerCAmelCase : Any = ( F"""Expected b_coeffs to have {self.order + 1} elements """ F"""for {self.order}-order filter, got {len(_SCREAMING_SNAKE_CASE)}""" ) raise ValueError(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = a_coeffs __lowerCAmelCase : List[Any] = b_coeffs def _SCREAMING_SNAKE_CASE ( self: int , _SCREAMING_SNAKE_CASE: float) -> float: """simple docstring""" __lowerCAmelCase : Union[str, Any] = 0.0 # Start at index 1 and do index 0 at the end. for i in range(1 , self.order + 1): result += ( self.b_coeffs[i] * self.input_history[i - 1] - self.a_coeffs[i] * self.output_history[i - 1] ) __lowerCAmelCase : List[str] = (result + self.b_coeffs[0] * sample) / self.a_coeffs[0] __lowerCAmelCase : str = self.input_history[:-1] __lowerCAmelCase : Optional[int] = self.output_history[:-1] __lowerCAmelCase : Union[str, Any] = sample __lowerCAmelCase : List[Any] = result return result
269
"""simple docstring""" import os from math import logaa def _lowercase ( __snake_case = "base_exp.txt" ) -> int: __lowerCAmelCase : float = 0 __lowerCAmelCase : Any = 0 for i, line in enumerate(open(os.path.join(os.path.dirname(__snake_case ) ,__snake_case ) ) ): __lowerCAmelCase , __lowerCAmelCase : List[str] = list(map(__snake_case ,line.split("," ) ) ) if x * logaa(__snake_case ) > largest: __lowerCAmelCase : Tuple = x * logaa(__snake_case ) __lowerCAmelCase : Optional[Any] = i + 1 return result if __name__ == "__main__": print(solution())
269
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __snake_case : int = { 'configuration_encodec': [ 'ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP', 'EncodecConfig', ], 'feature_extraction_encodec': ['EncodecFeatureExtractor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : Tuple = [ 'ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST', 'EncodecModel', 'EncodecPreTrainedModel', ] if TYPE_CHECKING: from .configuration_encodec import ( ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP, EncodecConfig, ) from .feature_extraction_encodec import EncodecFeatureExtractor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encodec import ( ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST, EncodecModel, EncodecPreTrainedModel, ) else: import sys __snake_case : str = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
269
"""simple docstring""" import collections import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_flax_cross_test, require_flax, require_torch, require_vision, slow, torch_device, ) from transformers.utils import is_flax_available, is_torch_available, is_vision_available from ...test_modeling_flax_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_flax_bert import FlaxBertModelTester from ..clip.test_modeling_flax_clip import FlaxCLIPVisionModelTester from ..vit.test_modeling_flax_vit import FlaxViTModelTester if is_flax_available(): from transformers import ( FlaxBertModel, FlaxCLIPVisionModel, FlaxVisionTextDualEncoderModel, FlaxViTModel, VisionTextDualEncoderConfig, VisionTextDualEncoderProcessor, ) from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) if is_torch_available(): import torch from transformers import VisionTextDualEncoderModel if is_vision_available(): from PIL import Image def _lowercase ( __snake_case ) -> List[str]: if isinstance(__snake_case ,collections.abc.Iterable ): return x return (x, x) @require_flax class A__ : '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Optional[Any]) -> str: """simple docstring""" pass def _SCREAMING_SNAKE_CASE ( self: str) -> int: """simple docstring""" pass def _SCREAMING_SNAKE_CASE ( self: List[str]) -> Tuple: """simple docstring""" pass def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: np.ndarray , _SCREAMING_SNAKE_CASE: np.ndarray , _SCREAMING_SNAKE_CASE: float) -> List[Any]: """simple docstring""" __lowerCAmelCase : str = np.abs((a - b)).max() self.assertLessEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , F"""Difference between torch and flax is {diff} (>= {tol}).""") def _SCREAMING_SNAKE_CASE ( self: str , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Any=None , **_SCREAMING_SNAKE_CASE: Tuple) -> List[str]: """simple docstring""" __lowerCAmelCase : Tuple = VisionTextDualEncoderConfig.from_vision_text_configs(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = FlaxVisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = model(input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], config.projection_dim)) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], config.projection_dim)) def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Optional[Any]=None , **_SCREAMING_SNAKE_CASE: List[str]) -> Dict: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : List[Any] = self.get_vision_text_model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = {"vision_model": vision_model, "text_model": text_model} __lowerCAmelCase : Optional[Any] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = model(input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], model.config.projection_dim)) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], model.config.projection_dim)) def _SCREAMING_SNAKE_CASE ( self: Dict , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: List[str] , _SCREAMING_SNAKE_CASE: List[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: List[Any]=None , **_SCREAMING_SNAKE_CASE: Tuple) -> str: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : Optional[Any] = self.get_vision_text_model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = {"vision_model": vision_model, "text_model": text_model} __lowerCAmelCase : Any = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = model(input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = output[0] with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = FlaxVisionTextDualEncoderModel.from_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = model(input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = after_output[0] __lowerCAmelCase : Any = np.amax(np.abs(out_a - out_a)) self.assertLessEqual(_SCREAMING_SNAKE_CASE , 1e-3) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Any=None , **_SCREAMING_SNAKE_CASE: List[str]) -> Optional[int]: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : Any = self.get_vision_text_model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = {"vision_model": vision_model, "text_model": text_model} __lowerCAmelCase : Optional[Any] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = model( input_ids=_SCREAMING_SNAKE_CASE , pixel_values=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , output_attentions=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = output.vision_model_output.attentions self.assertEqual(len(_SCREAMING_SNAKE_CASE) , vision_config.num_hidden_layers) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) __lowerCAmelCase : List[str] = to_atuple(vision_model.config.image_size) __lowerCAmelCase : Any = to_atuple(vision_model.config.patch_size) __lowerCAmelCase : Union[str, Any] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) __lowerCAmelCase : Tuple = num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len)) __lowerCAmelCase : Union[str, Any] = output.text_model_output.attentions self.assertEqual(len(_SCREAMING_SNAKE_CASE) , text_config.num_hidden_layers) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: int) -> str: """simple docstring""" pt_model.to(_SCREAMING_SNAKE_CASE) pt_model.eval() # prepare inputs __lowerCAmelCase : Union[str, Any] = inputs_dict __lowerCAmelCase : Union[str, Any] = {k: torch.tensor(v.tolist()) for k, v in flax_inputs.items()} with torch.no_grad(): __lowerCAmelCase : Any = pt_model(**_SCREAMING_SNAKE_CASE).to_tuple() __lowerCAmelCase : List[Any] = fx_model(**_SCREAMING_SNAKE_CASE).to_tuple() self.assertEqual(len(_SCREAMING_SNAKE_CASE) , len(_SCREAMING_SNAKE_CASE) , "Output lengths differ between Flax and PyTorch") for fx_output, pt_output in zip(fx_outputs[:4] , pt_outputs[:4]): self.assert_almost_equals(_SCREAMING_SNAKE_CASE , pt_output.numpy() , 4e-2) # PT -> Flax with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = FlaxVisionTextDualEncoderModel.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = fx_model_loaded(**_SCREAMING_SNAKE_CASE).to_tuple() self.assertEqual(len(_SCREAMING_SNAKE_CASE) , len(_SCREAMING_SNAKE_CASE) , "Output lengths differ between Flax and PyTorch") for fx_output_loaded, pt_output in zip(fx_outputs_loaded[:4] , pt_outputs[:4]): self.assert_almost_equals(_SCREAMING_SNAKE_CASE , pt_output.numpy() , 4e-2) # Flax -> PT with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = VisionTextDualEncoderModel.from_pretrained(_SCREAMING_SNAKE_CASE , from_flax=_SCREAMING_SNAKE_CASE) pt_model_loaded.to(_SCREAMING_SNAKE_CASE) pt_model_loaded.eval() with torch.no_grad(): __lowerCAmelCase : Optional[Any] = pt_model_loaded(**_SCREAMING_SNAKE_CASE).to_tuple() self.assertEqual(len(_SCREAMING_SNAKE_CASE) , len(_SCREAMING_SNAKE_CASE) , "Output lengths differ between Flax and PyTorch") for fx_output, pt_output_loaded in zip(fx_outputs[:4] , pt_outputs_loaded[:4]): self.assert_almost_equals(_SCREAMING_SNAKE_CASE , pt_output_loaded.numpy() , 4e-2) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Dict) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Union[str, Any] = VisionTextDualEncoderConfig.from_vision_text_configs(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = VisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = FlaxVisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = fx_state self.check_pt_flax_equivalence(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Tuple , _SCREAMING_SNAKE_CASE: Dict) -> str: """simple docstring""" __lowerCAmelCase : Optional[int] = VisionTextDualEncoderConfig.from_vision_text_configs(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = VisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = FlaxVisionTextDualEncoderModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = load_flax_weights_in_pytorch_model(_SCREAMING_SNAKE_CASE , fx_model.params) self.check_pt_flax_equivalence(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> str: """simple docstring""" __lowerCAmelCase : List[str] = self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Dict) -> int: """simple docstring""" __lowerCAmelCase : List[Any] = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> List[Any]: """simple docstring""" __lowerCAmelCase : Dict = self.prepare_config_and_inputs() self.check_save_load(**_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: int) -> Dict: """simple docstring""" __lowerCAmelCase : List[str] = self.prepare_config_and_inputs() self.check_vision_text_output_attention(**_SCREAMING_SNAKE_CASE) @is_pt_flax_cross_test def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> Any: """simple docstring""" __lowerCAmelCase : Dict = self.prepare_config_and_inputs() __lowerCAmelCase : List[Any] = config_inputs_dict.pop("vision_config") __lowerCAmelCase : str = config_inputs_dict.pop("text_config") __lowerCAmelCase : Union[str, Any] = config_inputs_dict self.check_equivalence_pt_to_flax(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) self.check_equivalence_flax_to_pt(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) @slow def _SCREAMING_SNAKE_CASE ( self: str) -> Dict: """simple docstring""" __lowerCAmelCase , __lowerCAmelCase : Dict = self.get_pretrained_model_and_inputs() __lowerCAmelCase : Union[str, Any] = model_a(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = outputs[0] with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = FlaxVisionTextDualEncoderModel.from_pretrained(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = model_a(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = after_outputs[0] __lowerCAmelCase : List[Any] = np.amax(np.abs(out_a - out_a)) self.assertLessEqual(_SCREAMING_SNAKE_CASE , 1e-5) @require_flax class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: List[str]) -> List[Any]: """simple docstring""" __lowerCAmelCase : Any = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( "hf-internal-testing/tiny-random-vit" , "hf-internal-testing/tiny-bert" , vision_from_pt=_SCREAMING_SNAKE_CASE , text_from_pt=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Union[str, Any] = 13 __lowerCAmelCase : Optional[int] = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ]) __lowerCAmelCase : List[Any] = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size) __lowerCAmelCase : List[Any] = random_attention_mask([batch_size, 4]) __lowerCAmelCase : str = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def _SCREAMING_SNAKE_CASE ( self: int , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Union[str, Any]) -> Optional[int]: """simple docstring""" __lowerCAmelCase : List[str] = FlaxViTModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = FlaxBertModel(_SCREAMING_SNAKE_CASE) return vision_model, text_model def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> int: """simple docstring""" __lowerCAmelCase : List[Any] = FlaxViTModelTester(self) __lowerCAmelCase : Optional[Any] = FlaxBertModelTester(self) __lowerCAmelCase : int = vit_model_tester.prepare_config_and_inputs() __lowerCAmelCase : List[str] = bert_model_tester.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase : Tuple = vision_config_and_inputs __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Union[str, Any] = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_torch class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Dict) -> List[str]: """simple docstring""" __lowerCAmelCase : Optional[int] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( "hf-internal-testing/tiny-random-clip" , "hf-internal-testing/tiny-bert" , vision_from_pt=_SCREAMING_SNAKE_CASE , text_from_pt=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Optional[int] = 13 __lowerCAmelCase : List[str] = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ]) __lowerCAmelCase : Any = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size) __lowerCAmelCase : str = random_attention_mask([batch_size, 4]) __lowerCAmelCase : Optional[int] = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Union[str, Any]) -> int: """simple docstring""" __lowerCAmelCase : int = FlaxCLIPVisionModel(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = FlaxBertModel(_SCREAMING_SNAKE_CASE) return vision_model, text_model def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : List[Any] = FlaxCLIPVisionModelTester(self) __lowerCAmelCase : str = FlaxBertModelTester(self) __lowerCAmelCase : Optional[Any] = clip_model_tester.prepare_config_and_inputs() __lowerCAmelCase : Dict = bert_model_tester.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase : Any = vision_config_and_inputs __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : List[Any] = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_flax @require_vision class A__ ( unittest.TestCase ): '''simple docstring''' @slow def _SCREAMING_SNAKE_CASE ( self: List[str]) -> List[str]: """simple docstring""" __lowerCAmelCase : Dict = FlaxVisionTextDualEncoderModel.from_pretrained("clip-italian/clip-italian" , logit_scale_init_value=1.0) __lowerCAmelCase : str = VisionTextDualEncoderProcessor.from_pretrained("clip-italian/clip-italian") __lowerCAmelCase : Tuple = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png") __lowerCAmelCase : Optional[int] = processor( text=["una foto di un gatto", "una foto di un cane"] , images=_SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE , return_tensors="np") __lowerCAmelCase : List[str] = model(**_SCREAMING_SNAKE_CASE) # verify the logits self.assertEqual(outputs.logits_per_image.shape , (inputs.pixel_values.shape[0], inputs.input_ids.shape[0])) self.assertEqual( outputs.logits_per_text.shape , (inputs.input_ids.shape[0], inputs.pixel_values.shape[0]) , ) __lowerCAmelCase : List[str] = np.array([[1.228_4727, 0.310_4122]]) self.assertTrue(np.allclose(outputs.logits_per_image , _SCREAMING_SNAKE_CASE , atol=1e-3))
269
1
"""simple docstring""" import argparse from collections import defaultdict def _lowercase ( __snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ) -> Dict: __lowerCAmelCase : Optional[int] = F"""{file}_{class_name}_{test_name}""" done_test[_id] += 1 with open(__snake_case ,"r" ) as f: __lowerCAmelCase : List[Any] = f.readlines() __lowerCAmelCase : Any = F"""class {class_name}(""" __lowerCAmelCase : Dict = F"""{4 * ' '}def {test_name}(""" __lowerCAmelCase : Any = F"""{8 * ' '}{correct_line.split()[0]}""" __lowerCAmelCase : Dict = F"""{16 * ' '}{correct_line.split()[0]}""" __lowerCAmelCase : List[str] = False __lowerCAmelCase : Dict = False __lowerCAmelCase : Optional[int] = False __lowerCAmelCase : Tuple = False __lowerCAmelCase : List[Any] = 0 __lowerCAmelCase : Optional[int] = 0 __lowerCAmelCase : Any = [] for line in lines: if line.startswith(__snake_case ): __lowerCAmelCase : Optional[Any] = True elif in_class and line.startswith(__snake_case ): __lowerCAmelCase : Optional[int] = True elif in_class and in_func and (line.startswith(__snake_case ) or line.startswith(__snake_case )): __lowerCAmelCase : Any = len(line.split(correct_line.split()[0] )[0] ) count += 1 if count == done_test[_id]: __lowerCAmelCase : Tuple = True if in_class and in_func and in_line: if ")" not in line: continue else: __lowerCAmelCase : List[str] = True if in_class and in_func and in_line and insert_line: new_lines.append(F"""{spaces * ' '}{correct_line}""" ) __lowerCAmelCase : str = False else: new_lines.append(__snake_case ) with open(__snake_case ,"w" ) as f: for line in new_lines: f.write(__snake_case ) def _lowercase ( __snake_case ,__snake_case=None ) -> Tuple: if fail is not None: with open(__snake_case ,"r" ) as f: __lowerCAmelCase : Union[str, Any] = {l.strip() for l in f.readlines()} else: __lowerCAmelCase : Union[str, Any] = None with open(__snake_case ,"r" ) as f: __lowerCAmelCase : List[str] = f.readlines() __lowerCAmelCase : str = defaultdict(__snake_case ) for line in correct_lines: __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Optional[Any] = line.split(";" ) if test_failures is None or "::".join([file, class_name, test_name] ) in test_failures: overwrite_file(__snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ) if __name__ == "__main__": __snake_case : Tuple = argparse.ArgumentParser() parser.add_argument('--correct_filename', help='filename of tests with expected result') parser.add_argument('--fail_filename', help='filename of test failures', type=str, default=None) __snake_case : Optional[Any] = parser.parse_args() main(args.correct_filename, args.fail_filename)
269
"""simple docstring""" from __future__ import annotations def _lowercase ( __snake_case ,__snake_case ,__snake_case ,__snake_case ) -> list: __lowerCAmelCase : Dict = [] __lowerCAmelCase , __lowerCAmelCase : Any = input_list[low:mid], input_list[mid : high + 1] while left and right: result.append((left if left[0] <= right[0] else right).pop(0 ) ) __lowerCAmelCase : int = result + left + right return input_list def _lowercase ( __snake_case ) -> list: if len(__snake_case ) <= 1: return input_list __lowerCAmelCase : int = list(__snake_case ) # iteration for two-way merging __lowerCAmelCase : Optional[int] = 2 while p <= len(__snake_case ): # getting low, high and middle value for merge-sort of single list for i in range(0 ,len(__snake_case ) ,__snake_case ): __lowerCAmelCase : Union[str, Any] = i __lowerCAmelCase : Tuple = i + p - 1 __lowerCAmelCase : Optional[Any] = (low + high + 1) // 2 __lowerCAmelCase : Any = merge(__snake_case ,__snake_case ,__snake_case ,__snake_case ) # final merge of last two parts if p * 2 >= len(__snake_case ): __lowerCAmelCase : Optional[Any] = i __lowerCAmelCase : Union[str, Any] = merge(__snake_case ,0 ,__snake_case ,len(__snake_case ) - 1 ) break p *= 2 return input_list if __name__ == "__main__": __snake_case : Union[str, Any] = input('Enter numbers separated by a comma:\n').strip() if user_input == "": __snake_case : Optional[int] = [] else: __snake_case : int = [int(item.strip()) for item in user_input.split(',')] print(iter_merge_sort(unsorted))
269
1
"""simple docstring""" import numpy as np from numpy import ndarray from scipy.optimize import Bounds, LinearConstraint, minimize def _lowercase ( __snake_case ) -> float: return np.dot(__snake_case ,__snake_case ) class A__ : '''simple docstring''' def __init__( self: Dict , *, _SCREAMING_SNAKE_CASE: float = np.inf , _SCREAMING_SNAKE_CASE: str = "linear" , _SCREAMING_SNAKE_CASE: float = 0.0 , ) -> None: """simple docstring""" __lowerCAmelCase : Any = regularization __lowerCAmelCase : str = gamma if kernel == "linear": __lowerCAmelCase : List[str] = self.__linear elif kernel == "rbf": if self.gamma == 0: raise ValueError("rbf kernel requires gamma") if not isinstance(self.gamma , (float, int)): raise ValueError("gamma must be float or int") if not self.gamma > 0: raise ValueError("gamma must be > 0") __lowerCAmelCase : Dict = self.__rbf # in the future, there could be a default value like in sklearn # sklear: def_gamma = 1/(n_features * X.var()) (wiki) # previously it was 1/(n_features) else: __lowerCAmelCase : List[Any] = F"""Unknown kernel: {kernel}""" raise ValueError(_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: ndarray , _SCREAMING_SNAKE_CASE: ndarray) -> float: """simple docstring""" return np.dot(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Optional[int] , _SCREAMING_SNAKE_CASE: ndarray , _SCREAMING_SNAKE_CASE: ndarray) -> float: """simple docstring""" return np.exp(-(self.gamma * norm_squared(vectora - vectora))) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: list[ndarray] , _SCREAMING_SNAKE_CASE: ndarray) -> None: """simple docstring""" __lowerCAmelCase : Optional[int] = observations __lowerCAmelCase : List[str] = classes # using Wolfe's Dual to calculate w. # Primal problem: minimize 1/2*norm_squared(w) # constraint: yn(w . xn + b) >= 1 # # With l a vector # Dual problem: maximize sum_n(ln) - # 1/2 * sum_n(sum_m(ln*lm*yn*ym*xn . xm)) # constraint: self.C >= ln >= 0 # and sum_n(ln*yn) = 0 # Then we get w using w = sum_n(ln*yn*xn) # At the end we can get b ~= mean(yn - w . xn) # # Since we use kernels, we only need l_star to calculate b # and to classify observations ((__lowerCAmelCase) , ) : Dict = np.shape(_SCREAMING_SNAKE_CASE) def to_minimize(_SCREAMING_SNAKE_CASE: ndarray) -> float: __lowerCAmelCase : Optional[Any] = 0 ((__lowerCAmelCase) , ) : List[str] = np.shape(_SCREAMING_SNAKE_CASE) for i in range(_SCREAMING_SNAKE_CASE): for j in range(_SCREAMING_SNAKE_CASE): s += ( candidate[i] * candidate[j] * classes[i] * classes[j] * self.kernel(observations[i] , observations[j]) ) return 1 / 2 * s - sum(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = LinearConstraint(_SCREAMING_SNAKE_CASE , 0 , 0) __lowerCAmelCase : List[Any] = Bounds(0 , self.regularization) __lowerCAmelCase : Optional[int] = minimize( _SCREAMING_SNAKE_CASE , np.ones(_SCREAMING_SNAKE_CASE) , bounds=_SCREAMING_SNAKE_CASE , constraints=[ly_contraint]).x __lowerCAmelCase : str = l_star # calculating mean offset of separation plane to points __lowerCAmelCase : str = 0 for i in range(_SCREAMING_SNAKE_CASE): for j in range(_SCREAMING_SNAKE_CASE): s += classes[i] - classes[i] * self.optimum[i] * self.kernel( observations[i] , observations[j]) __lowerCAmelCase : Dict = s / n def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: ndarray) -> int: """simple docstring""" __lowerCAmelCase : Dict = sum( self.optimum[n] * self.classes[n] * self.kernel(self.observations[n] , _SCREAMING_SNAKE_CASE) for n in range(len(self.classes))) return 1 if s + self.offset >= 0 else -1 if __name__ == "__main__": import doctest doctest.testmod()
269
"""simple docstring""" import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class A__ ( unittest.TestCase ): '''simple docstring''' @slow def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> str: """simple docstring""" __lowerCAmelCase : Optional[Any] = FlaxMTaForConditionalGeneration.from_pretrained("google/mt5-small") __lowerCAmelCase : Optional[Any] = AutoTokenizer.from_pretrained("google/mt5-small") __lowerCAmelCase : Tuple = tokenizer("Hello there" , return_tensors="np").input_ids __lowerCAmelCase : Dict = tokenizer("Hi I am" , return_tensors="np").input_ids __lowerCAmelCase : str = shift_tokens_right(_SCREAMING_SNAKE_CASE , model.config.pad_token_id , model.config.decoder_start_token_id) __lowerCAmelCase : Optional[int] = model(_SCREAMING_SNAKE_CASE , decoder_input_ids=_SCREAMING_SNAKE_CASE).logits __lowerCAmelCase : int = optax.softmax_cross_entropy(_SCREAMING_SNAKE_CASE , onehot(_SCREAMING_SNAKE_CASE , logits.shape[-1])).mean() __lowerCAmelCase : List[str] = -(labels.shape[-1] * loss.item()) __lowerCAmelCase : str = -84.9127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE) < 1e-4)
269
1
"""simple docstring""" import itertools from dataclasses import dataclass from typing import Optional import pandas as pd import pyarrow as pa import datasets from datasets.table import table_cast @dataclass class A__ ( datasets.BuilderConfig ): '''simple docstring''' SCREAMING_SNAKE_CASE = None class A__ ( datasets.ArrowBasedBuilder ): '''simple docstring''' SCREAMING_SNAKE_CASE = PandasConfig def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> str: """simple docstring""" return datasets.DatasetInfo(features=self.config.features) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: Optional[int]) -> Optional[Any]: """simple docstring""" if not self.config.data_files: raise ValueError(F"""At least one data file must be specified, but got data_files={self.config.data_files}""") __lowerCAmelCase : str = dl_manager.download_and_extract(self.config.data_files) if isinstance(_SCREAMING_SNAKE_CASE , (str, list, tuple)): __lowerCAmelCase : Optional[int] = data_files if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE): __lowerCAmelCase : Dict = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive __lowerCAmelCase : str = [dl_manager.iter_files(_SCREAMING_SNAKE_CASE) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"files": files})] __lowerCAmelCase : Optional[Any] = [] for split_name, files in data_files.items(): if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE): __lowerCAmelCase : Optional[Any] = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive __lowerCAmelCase : List[str] = [dl_manager.iter_files(_SCREAMING_SNAKE_CASE) for file in files] splits.append(datasets.SplitGenerator(name=_SCREAMING_SNAKE_CASE , gen_kwargs={"files": files})) return splits def _SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: pa.Table) -> pa.Table: """simple docstring""" if self.config.features is not None: # more expensive cast to support nested features with keys in a different order # allows str <-> int/float or str to Audio for example __lowerCAmelCase : Optional[int] = table_cast(_SCREAMING_SNAKE_CASE , self.config.features.arrow_schema) return pa_table def _SCREAMING_SNAKE_CASE ( self: str , _SCREAMING_SNAKE_CASE: Union[str, Any]) -> Optional[int]: """simple docstring""" for i, file in enumerate(itertools.chain.from_iterable(_SCREAMING_SNAKE_CASE)): with open(_SCREAMING_SNAKE_CASE , "rb") as f: __lowerCAmelCase : Any = pa.Table.from_pandas(pd.read_pickle(_SCREAMING_SNAKE_CASE)) yield i, self._cast_table(_SCREAMING_SNAKE_CASE)
269
"""simple docstring""" import re def _lowercase ( __snake_case ) -> str: if len(re.findall("[ATCG]" ,__snake_case ) ) != len(__snake_case ): raise ValueError("Invalid Strand" ) return dna.translate(dna.maketrans("ATCG" ,"TAGC" ) ) if __name__ == "__main__": import doctest doctest.testmod()
269
1
"""simple docstring""" import re def _lowercase ( __snake_case ) -> str: if len(re.findall("[ATCG]" ,__snake_case ) ) != len(__snake_case ): raise ValueError("Invalid Strand" ) return dna.translate(dna.maketrans("ATCG" ,"TAGC" ) ) if __name__ == "__main__": import doctest doctest.testmod()
269
"""simple docstring""" import os import unittest from transformers import FunnelTokenizer, FunnelTokenizerFast from transformers.models.funnel.tokenization_funnel import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = FunnelTokenizer SCREAMING_SNAKE_CASE = FunnelTokenizerFast SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = True def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Optional[int]: """simple docstring""" super().setUp() __lowerCAmelCase : str = [ "<unk>", "<cls>", "<sep>", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] __lowerCAmelCase : int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"]) with open(self.vocab_file , "w" , encoding="utf-8") as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens])) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , **_SCREAMING_SNAKE_CASE: Union[str, Any]) -> int: """simple docstring""" return FunnelTokenizer.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Any , **_SCREAMING_SNAKE_CASE: Any) -> str: """simple docstring""" return FunnelTokenizerFast.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: str) -> Any: """simple docstring""" __lowerCAmelCase : Union[str, Any] = "UNwant\u00E9d,running" __lowerCAmelCase : str = "unwanted, running" return input_text, output_text def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> List[str]: """simple docstring""" __lowerCAmelCase : Any = self.tokenizer_class(self.vocab_file) __lowerCAmelCase : Any = tokenizer.tokenize("UNwant\u00E9d,running") self.assertListEqual(_SCREAMING_SNAKE_CASE , ["un", "##want", "##ed", ",", "runn", "##ing"]) self.assertListEqual(tokenizer.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE) , [7, 4, 5, 10, 8, 9]) def _SCREAMING_SNAKE_CASE ( self: List[str]) -> List[str]: """simple docstring""" __lowerCAmelCase : Optional[Any] = self.get_tokenizers(do_lower_case=_SCREAMING_SNAKE_CASE) for tokenizer in tokenizers: __lowerCAmelCase : List[str] = tokenizer("UNwant\u00E9d,running") __lowerCAmelCase : Optional[int] = len(inputs["input_ids"]) - 1 self.assertListEqual(inputs["token_type_ids"] , [2] + [0] * sentence_len) __lowerCAmelCase : List[str] = tokenizer("UNwant\u00E9d,running" , "UNwant\u00E9d,running") self.assertListEqual(inputs["token_type_ids"] , [2] + [0] * sentence_len + [1] * sentence_len)
269
1
"""simple docstring""" import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = LongformerTokenizer SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = LongformerTokenizerFast SCREAMING_SNAKE_CASE = True def _SCREAMING_SNAKE_CASE ( self: Dict) -> str: """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __lowerCAmelCase : List[Any] = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", ] __lowerCAmelCase : Tuple = dict(zip(_SCREAMING_SNAKE_CASE , range(len(_SCREAMING_SNAKE_CASE)))) __lowerCAmelCase : Optional[int] = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] __lowerCAmelCase : List[Any] = {"unk_token": "<unk>"} __lowerCAmelCase : List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"]) __lowerCAmelCase : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"]) with open(self.vocab_file , "w" , encoding="utf-8") as fp: fp.write(json.dumps(_SCREAMING_SNAKE_CASE) + "\n") with open(self.merges_file , "w" , encoding="utf-8") as fp: fp.write("\n".join(_SCREAMING_SNAKE_CASE)) def _SCREAMING_SNAKE_CASE ( self: int , **_SCREAMING_SNAKE_CASE: int) -> List[Any]: """simple docstring""" kwargs.update(self.special_tokens_map) return self.tokenizer_class.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: int , **_SCREAMING_SNAKE_CASE: Optional[int]) -> Tuple: """simple docstring""" kwargs.update(self.special_tokens_map) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Dict , _SCREAMING_SNAKE_CASE: Union[str, Any]) -> List[Any]: """simple docstring""" __lowerCAmelCase : Union[str, Any] = "lower newer" __lowerCAmelCase : Dict = "lower newer" return input_text, output_text def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> Optional[int]: """simple docstring""" __lowerCAmelCase : Optional[Any] = self.tokenizer_class(self.vocab_file , self.merges_file , **self.special_tokens_map) __lowerCAmelCase : List[str] = "lower newer" __lowerCAmelCase : Dict = ["l", "o", "w", "er", "\u0120", "n", "e", "w", "er"] __lowerCAmelCase : Any = tokenizer.tokenize(_SCREAMING_SNAKE_CASE) # , add_prefix_space=True) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = tokens + [tokenizer.unk_token] __lowerCAmelCase : Any = [0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE) , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: int) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : Optional[int] = self.get_tokenizer() self.assertListEqual(tokenizer.encode("Hello world!" , add_special_tokens=_SCREAMING_SNAKE_CASE) , [0, 3_1414, 232, 328, 2]) self.assertListEqual( tokenizer.encode("Hello world! cécé herlolip 418" , add_special_tokens=_SCREAMING_SNAKE_CASE) , [0, 3_1414, 232, 328, 740, 1140, 1_2695, 69, 4_6078, 1588, 2] , ) @slow def _SCREAMING_SNAKE_CASE ( self: List[str]) -> Any: """simple docstring""" __lowerCAmelCase : Any = self.tokenizer_class.from_pretrained("allenai/longformer-base-4096") __lowerCAmelCase : Any = tokenizer.encode("sequence builders" , add_special_tokens=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = tokenizer.encode("multi-sequence build" , add_special_tokens=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = tokenizer.encode( "sequence builders" , add_special_tokens=_SCREAMING_SNAKE_CASE , add_prefix_space=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = tokenizer.encode( "sequence builders" , "multi-sequence build" , add_special_tokens=_SCREAMING_SNAKE_CASE , add_prefix_space=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = tokenizer.build_inputs_with_special_tokens(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = tokenizer.build_inputs_with_special_tokens(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode def _SCREAMING_SNAKE_CASE ( self: List[str]) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Optional[Any] = self.get_tokenizer() __lowerCAmelCase : str = "Encode this sequence." __lowerCAmelCase : Dict = tokenizer.byte_encoder[" ".encode("utf-8")[0]] # Testing encoder arguments __lowerCAmelCase : Dict = tokenizer.encode(_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE , add_prefix_space=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = tokenizer.convert_ids_to_tokens(encoded[0])[0] self.assertNotEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = tokenizer.encode(_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE , add_prefix_space=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = tokenizer.convert_ids_to_tokens(encoded[0])[0] self.assertEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) tokenizer.add_special_tokens({"bos_token": "<s>"}) __lowerCAmelCase : List[Any] = tokenizer.encode(_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = tokenizer.convert_ids_to_tokens(encoded[1])[0] self.assertNotEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) # Testing spaces after special tokens __lowerCAmelCase : str = "<mask>" tokenizer.add_special_tokens( {"mask_token": AddedToken(_SCREAMING_SNAKE_CASE , lstrip=_SCREAMING_SNAKE_CASE , rstrip=_SCREAMING_SNAKE_CASE)}) # mask token has a left space __lowerCAmelCase : str = tokenizer.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = "Encode <mask> sequence" __lowerCAmelCase : Optional[int] = "Encode <mask>sequence" __lowerCAmelCase : Tuple = tokenizer.encode(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = encoded.index(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1])[0] self.assertEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = tokenizer.encode(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = encoded.index(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1])[0] self.assertNotEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> Optional[int]: """simple docstring""" pass def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> List[str]: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})"""): __lowerCAmelCase : Any = self.rust_tokenizer_class.from_pretrained(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = self.tokenizer_class.from_pretrained(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = "A, <mask> AllenNLP sentence." __lowerCAmelCase : str = tokenizer_r.encode_plus(_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE , return_token_type_ids=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = tokenizer_p.encode_plus(_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE , return_token_type_ids=_SCREAMING_SNAKE_CASE) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r["token_type_ids"]) , sum(tokens_p["token_type_ids"])) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r["attention_mask"]) / len(tokens_r["attention_mask"]) , sum(tokens_p["attention_mask"]) / len(tokens_p["attention_mask"]) , ) __lowerCAmelCase : str = tokenizer_r.convert_ids_to_tokens(tokens_r["input_ids"]) __lowerCAmelCase : Optional[int] = tokenizer_p.convert_ids_to_tokens(tokens_p["input_ids"]) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p["input_ids"] , [0, 250, 6, 5_0264, 3823, 487, 2_1992, 3645, 4, 2]) self.assertSequenceEqual(tokens_r["input_ids"] , [0, 250, 6, 5_0264, 3823, 487, 2_1992, 3645, 4, 2]) self.assertSequenceEqual( _SCREAMING_SNAKE_CASE , ["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"]) self.assertSequenceEqual( _SCREAMING_SNAKE_CASE , ["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"]) def _SCREAMING_SNAKE_CASE ( self: Tuple) -> List[Any]: """simple docstring""" for trim_offsets, add_prefix_space in itertools.product([True, False] , repeat=2): __lowerCAmelCase : Optional[Any] = self.rust_tokenizer_class.from_pretrained( self.tmpdirname , use_fast=_SCREAMING_SNAKE_CASE , add_prefix_space=_SCREAMING_SNAKE_CASE , trim_offsets=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = json.loads(tokenizer_r.backend_tokenizer.pre_tokenizer.__getstate__()) __lowerCAmelCase : Optional[int] = json.loads(tokenizer_r.backend_tokenizer.post_processor.__getstate__()) self.assertEqual(pre_tokenizer_state["add_prefix_space"] , _SCREAMING_SNAKE_CASE) self.assertEqual(post_processor_state["add_prefix_space"] , _SCREAMING_SNAKE_CASE) self.assertEqual(post_processor_state["trim_offsets"] , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> int: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})"""): __lowerCAmelCase : Any = "hello" # `hello` is a token in the vocabulary of `pretrained_name` __lowerCAmelCase : Any = F"""{text_of_1_token} {text_of_1_token}""" __lowerCAmelCase : List[Any] = self.rust_tokenizer_class.from_pretrained( _SCREAMING_SNAKE_CASE , use_fast=_SCREAMING_SNAKE_CASE , add_prefix_space=_SCREAMING_SNAKE_CASE , trim_offsets=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = tokenizer_r(_SCREAMING_SNAKE_CASE , return_offsets_mapping=_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE) self.assertEqual(encoding.offset_mapping[0] , (0, len(_SCREAMING_SNAKE_CASE))) self.assertEqual( encoding.offset_mapping[1] , (len(_SCREAMING_SNAKE_CASE) + 1, len(_SCREAMING_SNAKE_CASE) + 1 + len(_SCREAMING_SNAKE_CASE)) , ) __lowerCAmelCase : Any = self.rust_tokenizer_class.from_pretrained( _SCREAMING_SNAKE_CASE , use_fast=_SCREAMING_SNAKE_CASE , add_prefix_space=_SCREAMING_SNAKE_CASE , trim_offsets=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[int] = tokenizer_r(_SCREAMING_SNAKE_CASE , return_offsets_mapping=_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE) self.assertEqual(encoding.offset_mapping[0] , (0, len(_SCREAMING_SNAKE_CASE))) self.assertEqual( encoding.offset_mapping[1] , (len(_SCREAMING_SNAKE_CASE) + 1, len(_SCREAMING_SNAKE_CASE) + 1 + len(_SCREAMING_SNAKE_CASE)) , ) __lowerCAmelCase : int = self.rust_tokenizer_class.from_pretrained( _SCREAMING_SNAKE_CASE , use_fast=_SCREAMING_SNAKE_CASE , add_prefix_space=_SCREAMING_SNAKE_CASE , trim_offsets=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = tokenizer_r(_SCREAMING_SNAKE_CASE , return_offsets_mapping=_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE) self.assertEqual(encoding.offset_mapping[0] , (0, len(_SCREAMING_SNAKE_CASE))) self.assertEqual( encoding.offset_mapping[1] , (len(_SCREAMING_SNAKE_CASE), len(_SCREAMING_SNAKE_CASE) + 1 + len(_SCREAMING_SNAKE_CASE)) , ) __lowerCAmelCase : Optional[int] = self.rust_tokenizer_class.from_pretrained( _SCREAMING_SNAKE_CASE , use_fast=_SCREAMING_SNAKE_CASE , add_prefix_space=_SCREAMING_SNAKE_CASE , trim_offsets=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = tokenizer_r(_SCREAMING_SNAKE_CASE , return_offsets_mapping=_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE) self.assertEqual(encoding.offset_mapping[0] , (0, len(_SCREAMING_SNAKE_CASE))) self.assertEqual( encoding.offset_mapping[1] , (len(_SCREAMING_SNAKE_CASE), len(_SCREAMING_SNAKE_CASE) + 1 + len(_SCREAMING_SNAKE_CASE)) , ) __lowerCAmelCase : Tuple = F""" {text}""" # tokenizer_r = self.rust_tokenizer_class.from_pretrained( # pretrained_name, use_fast=True, add_prefix_space=True, trim_offsets=True # ) # encoding = tokenizer_r(text, return_offsets_mapping=True, add_special_tokens=False) # self.assertEqual(encoding.offset_mapping[0], (1, 1 + len(text_of_1_token))) # self.assertEqual( # encoding.offset_mapping[1], # (1 + len(text_of_1_token) + 1, 1 + len(text_of_1_token) + 1 + len(text_of_1_token)), # ) __lowerCAmelCase : str = self.rust_tokenizer_class.from_pretrained( _SCREAMING_SNAKE_CASE , use_fast=_SCREAMING_SNAKE_CASE , add_prefix_space=_SCREAMING_SNAKE_CASE , trim_offsets=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = tokenizer_r(_SCREAMING_SNAKE_CASE , return_offsets_mapping=_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(_SCREAMING_SNAKE_CASE))) self.assertEqual( encoding.offset_mapping[1] , (1 + len(_SCREAMING_SNAKE_CASE) + 1, 1 + len(_SCREAMING_SNAKE_CASE) + 1 + len(_SCREAMING_SNAKE_CASE)) , ) __lowerCAmelCase : List[str] = self.rust_tokenizer_class.from_pretrained( _SCREAMING_SNAKE_CASE , use_fast=_SCREAMING_SNAKE_CASE , add_prefix_space=_SCREAMING_SNAKE_CASE , trim_offsets=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Optional[Any] = tokenizer_r(_SCREAMING_SNAKE_CASE , return_offsets_mapping=_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(_SCREAMING_SNAKE_CASE))) self.assertEqual( encoding.offset_mapping[1] , (1 + len(_SCREAMING_SNAKE_CASE), 1 + len(_SCREAMING_SNAKE_CASE) + 1 + len(_SCREAMING_SNAKE_CASE)) , ) __lowerCAmelCase : List[Any] = self.rust_tokenizer_class.from_pretrained( _SCREAMING_SNAKE_CASE , use_fast=_SCREAMING_SNAKE_CASE , add_prefix_space=_SCREAMING_SNAKE_CASE , trim_offsets=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = tokenizer_r(_SCREAMING_SNAKE_CASE , return_offsets_mapping=_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(_SCREAMING_SNAKE_CASE))) self.assertEqual( encoding.offset_mapping[1] , (1 + len(_SCREAMING_SNAKE_CASE), 1 + len(_SCREAMING_SNAKE_CASE) + 1 + len(_SCREAMING_SNAKE_CASE)) , )
269
"""simple docstring""" from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def _lowercase ( __snake_case = "laptop" ) -> DataFrame: __lowerCAmelCase : str = F"""https://www.amazon.in/laptop/s?k={product}""" __lowerCAmelCase : Union[str, Any] = { "User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36\n (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36", "Accept-Language": "en-US, en;q=0.5", } __lowerCAmelCase : List[str] = BeautifulSoup(requests.get(__snake_case ,headers=__snake_case ).text ) # Initialize a Pandas dataframe with the column titles __lowerCAmelCase : Dict = DataFrame( columns=[ "Product Title", "Product Link", "Current Price of the product", "Product Rating", "MRP of the product", "Discount", ] ) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( "div" ,attrs={"class": "s-result-item", "data-component-type": "s-search-result"} ,) ,soup.find_all("div" ,attrs={"class": "a-row a-size-base a-color-base"} ) ,): try: __lowerCAmelCase : Any = item.ha.text __lowerCAmelCase : Union[str, Any] = "https://www.amazon.in/" + item.ha.a["href"] __lowerCAmelCase : Any = item.find("span" ,attrs={"class": "a-offscreen"} ).text try: __lowerCAmelCase : Union[str, Any] = item.find("span" ,attrs={"class": "a-icon-alt"} ).text except AttributeError: __lowerCAmelCase : Optional[Any] = "Not available" try: __lowerCAmelCase : Union[str, Any] = ( "₹" + item.find( "span" ,attrs={"class": "a-price a-text-price"} ).text.split("₹" )[1] ) except AttributeError: __lowerCAmelCase : Dict = "" try: __lowerCAmelCase : str = float( ( ( float(product_mrp.strip("₹" ).replace("," ,"" ) ) - float(product_price.strip("₹" ).replace("," ,"" ) ) ) / float(product_mrp.strip("₹" ).replace("," ,"" ) ) ) * 100 ) except ValueError: __lowerCAmelCase : List[str] = float("nan" ) except AttributeError: pass __lowerCAmelCase : int = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] __lowerCAmelCase : Union[str, Any] = " " __lowerCAmelCase : Union[str, Any] = " " data_frame.index += 1 return data_frame if __name__ == "__main__": __snake_case : Any = 'headphones' get_amazon_product_data(product).to_csv(F"""Amazon Product Data for {product}.csv""")
269
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case : str = logging.get_logger(__name__) __snake_case : int = { 'microsoft/markuplm-base': 'https://huggingface.co/microsoft/markuplm-base/resolve/main/config.json', 'microsoft/markuplm-large': 'https://huggingface.co/microsoft/markuplm-large/resolve/main/config.json', } class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'markuplm' def __init__( self: List[str] , _SCREAMING_SNAKE_CASE: Dict=3_0522 , _SCREAMING_SNAKE_CASE: Tuple=768 , _SCREAMING_SNAKE_CASE: Union[str, Any]=12 , _SCREAMING_SNAKE_CASE: Union[str, Any]=12 , _SCREAMING_SNAKE_CASE: Optional[int]=3072 , _SCREAMING_SNAKE_CASE: Tuple="gelu" , _SCREAMING_SNAKE_CASE: Dict=0.1 , _SCREAMING_SNAKE_CASE: Dict=0.1 , _SCREAMING_SNAKE_CASE: Optional[Any]=512 , _SCREAMING_SNAKE_CASE: Dict=2 , _SCREAMING_SNAKE_CASE: List[Any]=0.02 , _SCREAMING_SNAKE_CASE: Union[str, Any]=1e-12 , _SCREAMING_SNAKE_CASE: List[Any]=0 , _SCREAMING_SNAKE_CASE: Optional[int]=0 , _SCREAMING_SNAKE_CASE: str=2 , _SCREAMING_SNAKE_CASE: Optional[int]=256 , _SCREAMING_SNAKE_CASE: str=1024 , _SCREAMING_SNAKE_CASE: Union[str, Any]=216 , _SCREAMING_SNAKE_CASE: Any=1001 , _SCREAMING_SNAKE_CASE: Dict=32 , _SCREAMING_SNAKE_CASE: List[Any]=50 , _SCREAMING_SNAKE_CASE: Optional[int]="absolute" , _SCREAMING_SNAKE_CASE: Optional[Any]=True , _SCREAMING_SNAKE_CASE: str=None , **_SCREAMING_SNAKE_CASE: Optional[Any] , ) -> Optional[int]: """simple docstring""" super().__init__( pad_token_id=_SCREAMING_SNAKE_CASE , bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Dict = vocab_size __lowerCAmelCase : Optional[Any] = hidden_size __lowerCAmelCase : List[str] = num_hidden_layers __lowerCAmelCase : Tuple = num_attention_heads __lowerCAmelCase : Any = hidden_act __lowerCAmelCase : Optional[int] = intermediate_size __lowerCAmelCase : Tuple = hidden_dropout_prob __lowerCAmelCase : Optional[Any] = attention_probs_dropout_prob __lowerCAmelCase : Any = max_position_embeddings __lowerCAmelCase : Optional[int] = type_vocab_size __lowerCAmelCase : int = initializer_range __lowerCAmelCase : Union[str, Any] = layer_norm_eps __lowerCAmelCase : List[str] = position_embedding_type __lowerCAmelCase : int = use_cache __lowerCAmelCase : int = classifier_dropout # additional properties __lowerCAmelCase : Tuple = max_depth __lowerCAmelCase : Optional[int] = max_xpath_tag_unit_embeddings __lowerCAmelCase : Tuple = max_xpath_subs_unit_embeddings __lowerCAmelCase : Optional[Any] = tag_pad_id __lowerCAmelCase : int = subs_pad_id __lowerCAmelCase : Optional[int] = xpath_unit_hidden_size
269
"""simple docstring""" import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('TEST_SAGEMAKER' , 'False' ) ) is not True , reason='Skipping test because should only be run when releasing minor transformers version' , ) @pytest.mark.usefixtures('sm_env' ) @parameterized_class( [ { 'framework': 'pytorch', 'script': 'run_glue.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_5_0, 'eval_accuracy': 0.6, 'eval_loss': 0.9}, }, { 'framework': 'tensorflow', 'script': 'run_tf.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_0_0, 'eval_accuracy': 0.3, 'eval_loss': 0.9}, }, ] ) class A__ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: int) -> Tuple: """simple docstring""" if self.framework == "pytorch": subprocess.run( F"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() , encoding="utf-8" , check=_SCREAMING_SNAKE_CASE , ) assert hasattr(self , "env") def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Any=1) -> Dict: """simple docstring""" return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=F"""{self.env.base_job_name}-single""" , instance_count=_SCREAMING_SNAKE_CASE , instance_type=self.instance_type , debugger_hook_config=_SCREAMING_SNAKE_CASE , hyperparameters={**self.env.hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version="py36" , ) def _SCREAMING_SNAKE_CASE ( self: str , _SCREAMING_SNAKE_CASE: List[Any]) -> Optional[Any]: """simple docstring""" TrainingJobAnalytics(_SCREAMING_SNAKE_CASE).export_csv(F"""{self.env.test_path}/{job_name}_metrics.csv""") def _SCREAMING_SNAKE_CASE ( self: str) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Tuple = self.create_estimator() # run training estimator.fit() # result dataframe __lowerCAmelCase : Tuple = TrainingJobAnalytics(estimator.latest_training_job.name).dataframe() # extract kpis __lowerCAmelCase : Optional[Any] = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"]) __lowerCAmelCase : Union[str, Any] = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"]) # get train time from SageMaker job, this includes starting, preprocessing, stopping __lowerCAmelCase : Tuple = ( Session().describe_training_job(estimator.latest_training_job.name).get("TrainingTimeInSeconds" , 99_9999) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy) assert all(t <= self.results["eval_loss"] for t in eval_loss) # dump tests result into json file to share in PR with open(F"""{estimator.latest_training_job.name}.json""" , "w") as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , _SCREAMING_SNAKE_CASE)
269
1
"""simple docstring""" import unittest from pathlib import Path from shutil import copyfile from transformers import SPIECE_UNDERLINE, is_sentencepiece_available from transformers.models.speech_to_text import SpeechaTextTokenizer from transformers.models.speech_to_text.tokenization_speech_to_text import VOCAB_FILES_NAMES, save_json from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin __snake_case : Any = get_tests_dir('fixtures/test_sentencepiece.model') if is_sentencepiece_available(): import sentencepiece as sp __snake_case : str = 5 __snake_case : Optional[int] = 10 @require_sentencepiece @require_tokenizers class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = SpeechaTextTokenizer SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = True def _SCREAMING_SNAKE_CASE ( self: Any) -> str: """simple docstring""" super().setUp() __lowerCAmelCase : Any = sp.SentencePieceProcessor() spm_model.Load(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = ["<s>", "<pad>", "</s>", "<unk>"] vocab += [spm_model.IdToPiece(id_) for id_ in range(len(_SCREAMING_SNAKE_CASE))] __lowerCAmelCase : Dict = dict(zip(_SCREAMING_SNAKE_CASE , range(len(_SCREAMING_SNAKE_CASE)))) __lowerCAmelCase : Any = Path(self.tmpdirname) save_json(_SCREAMING_SNAKE_CASE , save_dir / VOCAB_FILES_NAMES["vocab_file"]) if not (save_dir / VOCAB_FILES_NAMES["spm_file"]).exists(): copyfile(_SCREAMING_SNAKE_CASE , save_dir / VOCAB_FILES_NAMES["spm_file"]) __lowerCAmelCase : Optional[int] = SpeechaTextTokenizer.from_pretrained(self.tmpdirname) tokenizer.save_pretrained(self.tmpdirname) def _SCREAMING_SNAKE_CASE ( self: str) -> List[Any]: """simple docstring""" __lowerCAmelCase : List[Any] = "<pad>" __lowerCAmelCase : str = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_SCREAMING_SNAKE_CASE) , _SCREAMING_SNAKE_CASE) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_SCREAMING_SNAKE_CASE) , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> Any: """simple docstring""" __lowerCAmelCase : Union[str, Any] = list(self.get_tokenizer().get_vocab().keys()) self.assertEqual(vocab_keys[0] , "<s>") self.assertEqual(vocab_keys[1] , "<pad>") self.assertEqual(vocab_keys[-1] , "j") self.assertEqual(len(_SCREAMING_SNAKE_CASE) , 1001) def _SCREAMING_SNAKE_CASE ( self: int) -> str: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1001) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> Any: """simple docstring""" __lowerCAmelCase : Dict = SpeechaTextTokenizer.from_pretrained(self.tmpdirname) __lowerCAmelCase : List[str] = tokenizer.tokenize("This is a test") self.assertListEqual(_SCREAMING_SNAKE_CASE , ["▁This", "▁is", "▁a", "▁t", "est"]) self.assertListEqual( tokenizer.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE) , [289, 50, 14, 174, 386] , ) __lowerCAmelCase : Any = tokenizer.tokenize("I was born in 92000, and this is falsé.") self.assertListEqual( _SCREAMING_SNAKE_CASE , [SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", "."] , ) __lowerCAmelCase : Tuple = tokenizer.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE) self.assertListEqual(_SCREAMING_SNAKE_CASE , [12, 25, 88, 59, 28, 23, 11, 4, 606, 351, 351, 351, 7, 16, 70, 50, 76, 84, 10, 4, 8]) __lowerCAmelCase : List[str] = tokenizer.convert_ids_to_tokens(_SCREAMING_SNAKE_CASE) self.assertListEqual( _SCREAMING_SNAKE_CASE , [SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", "."] , ) @slow def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> List[Any]: """simple docstring""" __lowerCAmelCase : Tuple = {"input_ids": [[3791, 797, 31, 11, 64, 797, 31, 2429, 433, 12, 1176, 12, 20, 786, 915, 142, 2413, 240, 37, 3238, 797, 31, 11, 35, 93, 915, 142, 2413, 240, 37, 5540, 567, 1276, 93, 37, 610, 40, 62, 455, 657, 1042, 123, 780, 177, 37, 309, 241, 1298, 514, 20, 292, 2737, 114, 2469, 241, 85, 64, 302, 548, 528, 423, 4, 509, 406, 423, 37, 601, 4, 777, 302, 548, 528, 423, 284, 4, 3388, 511, 459, 4, 3555, 40, 321, 302, 705, 4, 3388, 511, 583, 326, 5, 5, 5, 62, 3310, 560, 177, 2680, 217, 1508, 32, 31, 853, 418, 64, 583, 511, 1605, 62, 35, 93, 560, 177, 2680, 217, 1508, 1521, 64, 583, 511, 519, 62, 20, 1515, 764, 20, 149, 261, 5625, 7972, 20, 5540, 567, 1276, 93, 3925, 1675, 11, 15, 802, 7972, 576, 217, 1508, 11, 35, 93, 1253, 2441, 15, 289, 652, 31, 416, 321, 3842, 115, 40, 911, 8, 476, 619, 4, 380, 142, 423, 335, 240, 35, 93, 264, 8, 11, 335, 569, 420, 163, 5, 2], [260, 548, 528, 423, 20, 451, 20, 2681, 1153, 3434, 20, 5540, 37, 567, 126, 1253, 2441, 3376, 449, 210, 431, 1563, 177, 767, 5540, 11, 1203, 472, 11, 2953, 685, 285, 364, 706, 1153, 20, 6799, 20, 2869, 20, 4464, 126, 40, 2429, 20, 1040, 866, 2664, 418, 20, 318, 20, 1726, 186, 20, 265, 522, 35, 93, 2191, 4634, 20, 1040, 12, 6799, 15, 228, 2356, 142, 31, 11, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [2575, 2666, 684, 1582, 1176, 12, 627, 149, 619, 20, 4902, 563, 11, 20, 149, 261, 3420, 2356, 174, 142, 4714, 131, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_SCREAMING_SNAKE_CASE , model_name="facebook/s2t-small-mustc-en-de-st" , revision="a14f04cf0776c02f62a8cb800cf7909e15ea23ad" , ) @require_sentencepiece class A__ ( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'valhalla/s2t_mustc_multilinguial_medium' SCREAMING_SNAKE_CASE = 'C\'est trop cool' SCREAMING_SNAKE_CASE = 'Esto es genial' @classmethod def _SCREAMING_SNAKE_CASE ( cls: str) -> Any: """simple docstring""" __lowerCAmelCase : SpeechaTextTokenizer = SpeechaTextTokenizer.from_pretrained(cls.checkpoint_name) return cls def _SCREAMING_SNAKE_CASE ( self: Dict) -> str: """simple docstring""" self.assertEqual(self.tokenizer.lang_code_to_id["pt"] , 4) self.assertEqual(self.tokenizer.lang_code_to_id["ru"] , 6) self.assertEqual(self.tokenizer.lang_code_to_id["it"] , 9) self.assertEqual(self.tokenizer.lang_code_to_id["de"] , 11) def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> Optional[Any]: """simple docstring""" self.assertEqual(self.tokenizer.vocab_size , 1_0000) def _SCREAMING_SNAKE_CASE ( self: Dict) -> int: """simple docstring""" self.assertIn(_SCREAMING_SNAKE_CASE , self.tokenizer.all_special_ids) __lowerCAmelCase : Any = [ES_CODE, 4, 1601, 47, 7647, 2] __lowerCAmelCase : str = self.tokenizer.decode(_SCREAMING_SNAKE_CASE , skip_special_tokens=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=_SCREAMING_SNAKE_CASE) self.assertEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) self.assertNotIn(self.tokenizer.eos_token , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> int: """simple docstring""" __lowerCAmelCase : Tuple = "fr" __lowerCAmelCase : Optional[int] = self.tokenizer(self.french_text).input_ids self.assertEqual(encoded[0] , _SCREAMING_SNAKE_CASE) self.assertEqual(encoded[-1] , self.tokenizer.eos_token_id) def _SCREAMING_SNAKE_CASE ( self: List[str]) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase : Union[str, Any] = "fr" self.assertListEqual(self.tokenizer.prefix_tokens , [FR_CODE]) __lowerCAmelCase : Tuple = "es" self.assertListEqual(self.tokenizer.prefix_tokens , [ES_CODE])
269
"""simple docstring""" from typing import TYPE_CHECKING from ..utils import _LazyModule __snake_case : Optional[int] = { 'config': [ 'EXTERNAL_DATA_FORMAT_SIZE_LIMIT', 'OnnxConfig', 'OnnxConfigWithPast', 'OnnxSeq2SeqConfigWithPast', 'PatchingSpec', ], 'convert': ['export', 'validate_model_outputs'], 'features': ['FeaturesManager'], 'utils': ['ParameterFormat', 'compute_serialized_parameters_size'], } if TYPE_CHECKING: from .config import ( EXTERNAL_DATA_FORMAT_SIZE_LIMIT, OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast, PatchingSpec, ) from .convert import export, validate_model_outputs from .features import FeaturesManager from .utils import ParameterFormat, compute_serialized_parameters_size else: import sys __snake_case : Any = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
269
1